2023-03-31T05:23:34.0584058Z Requested labels: linux.2xlarge 2023-03-31T05:23:34.0584147Z Job defined at: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/pull/97655/merge 2023-03-31T05:23:34.0584247Z Reusable workflow chain: 2023-03-31T05:23:34.0584298Z pytorch/pytorch/.github/workflows/pull.yml@refs/pull/97655/merge (48284fffe3d9196187bcf2fdf764bf15956e81e8) 2023-03-31T05:23:34.0584335Z -> pytorch/pytorch/.github/workflows/_linux-test.yml@refs/pull/97655/merge (48284fffe3d9196187bcf2fdf764bf15956e81e8) 2023-03-31T05:23:34.0584363Z Waiting for a runner to pick up this job... 2023-03-31T05:23:34.3809983Z Job is about to start running on the runner: i-0da0c45d38bcc5929 (organization) 2023-03-31T05:23:38.8030788Z Current runner version: '2.303.0' 2023-03-31T05:23:38.8036400Z Runner name: 'i-0da0c45d38bcc5929' 2023-03-31T05:23:38.8036817Z Runner group name: 'Default' 2023-03-31T05:23:38.8037384Z Machine name: 'ip-10-0-2-162' 2023-03-31T05:23:38.8039173Z ##[group]GITHUB_TOKEN Permissions 2023-03-31T05:23:38.8039754Z Actions: write 2023-03-31T05:23:38.8039979Z Checks: write 2023-03-31T05:23:38.8040214Z Contents: write 2023-03-31T05:23:38.8040464Z Deployments: write 2023-03-31T05:23:38.8040701Z Discussions: write 2023-03-31T05:23:38.8040937Z Issues: write 2023-03-31T05:23:38.8041167Z Metadata: read 2023-03-31T05:23:38.8041396Z Packages: write 2023-03-31T05:23:38.8041632Z Pages: write 2023-03-31T05:23:38.8041880Z PullRequests: write 2023-03-31T05:23:38.8042155Z RepositoryProjects: write 2023-03-31T05:23:38.8042432Z SecurityEvents: write 2023-03-31T05:23:38.8042685Z Statuses: write 2023-03-31T05:23:38.8042908Z ##[endgroup] 2023-03-31T05:23:38.8046746Z Secret source: Actions 2023-03-31T05:23:38.8047383Z Prepare workflow directory 2023-03-31T05:23:39.0726316Z Prepare all required actions 2023-03-31T05:23:39.0914714Z Getting action download info 2023-03-31T05:23:39.2653772Z Download action repository 'pytorch/test-infra@main' (SHA:0347d1b47238d6276893ecf1aa2097103753aade) 2023-03-31T05:23:39.5809845Z Download action repository 'pytorch/pytorch@master' (SHA:c091aa9a2c5ba7d2b0a39f22b590807e04c936a6) 2023-03-31T05:23:42.2788828Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2023-03-31T05:23:42.5257908Z Getting action download info 2023-03-31T05:23:42.6399018Z Download action repository 'malfet/checkout@silent-checkout' (SHA:c7b8fef48edfe1bca0044a44b1f7f7c4318a3076) 2023-03-31T05:23:42.8114582Z Getting action download info 2023-03-31T05:23:42.9391780Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2023-03-31T05:23:43.0603551Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/pull/97655/merge (48284fffe3d9196187bcf2fdf764bf15956e81e8) 2023-03-31T05:23:43.0605350Z ##[group] Inputs 2023-03-31T05:23:43.0605752Z build-environment: linux-bionic-py3.8-clang9 2023-03-31T05:23:43.0606515Z test-matrix: { include: [ { config: "default", shard: 1, num_shards: 3, runner: "linux.2xlarge" }, { config: "default", shard: 2, num_shards: 3, runner: "linux.2xlarge" }, { config: "default", shard: 3, num_shards: 3, runner: "linux.2xlarge" }, { config: "crossref", shard: 1, num_shards: 2, runner: "linux.2xlarge" }, { config: "crossref", shard: 2, num_shards: 2, runner: "linux.2xlarge" }, { config: "dynamo", shard: 1, num_shards: 2, runner: "linux.2xlarge" }, { config: "dynamo", shard: 2, num_shards: 2, runner: "linux.2xlarge" }, ]} 2023-03-31T05:23:43.0607342Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:23:43.0607722Z sync-tag: 2023-03-31T05:23:43.0608481Z timeout-minutes: 240 2023-03-31T05:23:43.0608720Z use-gha: 2023-03-31T05:23:43.0608954Z ##[endgroup] 2023-03-31T05:23:43.0609452Z Complete job name: linux-bionic-py3.8-clang9 / test (default, 1, 3, linux.2xlarge) 2023-03-31T05:23:43.1360724Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2023-03-31T05:23:43.1361058Z with: 2023-03-31T05:23:43.1361670Z github-secret: *** 2023-03-31T05:23:43.1362063Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2023-03-31T05:23:43.1362438Z activate-with-label: false 2023-03-31T05:23:43.1362675Z label: with-ssh 2023-03-31T05:23:43.1362914Z remove-existing-keys: true 2023-03-31T05:23:43.1363354Z env: 2023-03-31T05:23:43.1363571Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:23:43.1363804Z ##[endgroup] 2023-03-31T05:23:43.8628167Z Grabbing public ssh keys from https://github.com/chunyuan-w.keys 2023-03-31T05:23:43.9390605Z ~/.ssh/authorized_keys file found on node, removing ~/.ssh and starting fresh 2023-03-31T05:23:43.9404099Z Public keys pulled and installed to /home/ec2-user/.ssh/authorized_keys 2023-03-31T05:23:43.9430539Z Login using: ssh ec2-user@ec2-3-95-186-70.compute-1.amazonaws.com 2023-03-31T05:23:43.9431243Z All testing is done inside the container, to start an interactive session run: 2023-03-31T05:23:43.9431710Z docker exec -it $(docker container ps --format '{{.ID}}') bash 2023-03-31T05:23:43.9658248Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@master 2023-03-31T05:23:43.9658516Z with: 2023-03-31T05:23:43.9658791Z submodules: recursive 2023-03-31T05:23:43.9659025Z fetch-depth: 0 2023-03-31T05:23:43.9659184Z env: 2023-03-31T05:23:43.9659367Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:23:43.9659562Z ##[endgroup] 2023-03-31T05:23:43.9882136Z ##[group]Run retry () { 2023-03-31T05:23:43.9882426Z retry () { 2023-03-31T05:23:43.9882669Z  $* || (sleep 1 && $*) || (sleep 2 && $*) || (sleep 4 && $*) || (sleep 8 && $*) 2023-03-31T05:23:43.9882912Z } 2023-03-31T05:23:43.9883257Z echo "${GITHUB_WORKSPACE}" 2023-03-31T05:23:43.9883493Z if [ -z "${NO_SUDO}" ]; then 2023-03-31T05:23:43.9883731Z  retry sudo rm -rf "${GITHUB_WORKSPACE}" 2023-03-31T05:23:43.9883933Z else 2023-03-31T05:23:43.9884140Z  retry rm -rf "${GITHUB_WORKSPACE}" 2023-03-31T05:23:43.9884343Z fi 2023-03-31T05:23:43.9884584Z mkdir "${GITHUB_WORKSPACE}" 2023-03-31T05:23:43.9900619Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:23:43.9900863Z env: 2023-03-31T05:23:43.9901055Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:23:43.9901250Z NO_SUDO: 2023-03-31T05:23:43.9901419Z ##[endgroup] 2023-03-31T05:23:44.0002990Z /home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T05:23:47.3938767Z ##[group]Run malfet/checkout@silent-checkout 2023-03-31T05:23:47.3939035Z with: 2023-03-31T05:23:47.3939245Z ref: 4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:23:47.3939471Z fetch-depth: 0 2023-03-31T05:23:47.3939660Z submodules: recursive 2023-03-31T05:23:47.3939858Z quiet-checkout: true 2023-03-31T05:23:47.3940052Z repository: pytorch/pytorch 2023-03-31T05:23:47.3940390Z token: *** 2023-03-31T05:23:47.3940568Z ssh-strict: true 2023-03-31T05:23:47.3940772Z persist-credentials: true 2023-03-31T05:23:47.3940957Z clean: true 2023-03-31T05:23:47.3941145Z lfs: false 2023-03-31T05:23:47.3941335Z set-safe-directory: true 2023-03-31T05:23:47.3941510Z env: 2023-03-31T05:23:47.3941688Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:23:47.3941879Z ##[endgroup] 2023-03-31T05:23:47.5130134Z Syncing repository: pytorch/pytorch 2023-03-31T05:23:47.5131807Z ##[group]Getting Git version info 2023-03-31T05:23:47.5132302Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2023-03-31T05:23:47.5132763Z [command]/usr/bin/git version 2023-03-31T05:23:47.5132971Z git version 2.39.2 2023-03-31T05:23:47.5134795Z ##[endgroup] 2023-03-31T05:23:47.5156635Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/2702acd2-1cac-40a6-b65a-2a768c3b79ed' before making global git config changes 2023-03-31T05:23:47.5157395Z Adding repository directory to the temporary git global config as a safe directory 2023-03-31T05:23:47.5161062Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T05:23:47.5207817Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2023-03-31T05:23:47.5212754Z ##[group]Initializing the repository 2023-03-31T05:23:47.5216244Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T05:23:47.5332309Z hint: Using 'master' as the name for the initial branch. This default branch name 2023-03-31T05:23:47.5332947Z hint: is subject to change. To configure the initial branch name to use in all 2023-03-31T05:23:47.5333344Z hint: of your new repositories, which will suppress this warning, call: 2023-03-31T05:23:47.5333835Z hint: 2023-03-31T05:23:47.5334147Z hint: git config --global init.defaultBranch 2023-03-31T05:23:47.5334613Z hint: 2023-03-31T05:23:47.5334923Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2023-03-31T05:23:47.5335301Z hint: 'development'. The just-created branch can be renamed via this command: 2023-03-31T05:23:47.5335548Z hint: 2023-03-31T05:23:47.5335881Z hint: git branch -m 2023-03-31T05:23:47.5336267Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2023-03-31T05:23:47.5343357Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2023-03-31T05:23:47.5378621Z ##[endgroup] 2023-03-31T05:23:47.5379025Z ##[group]Disabling automatic garbage collection 2023-03-31T05:23:47.5382069Z [command]/usr/bin/git config --local gc.auto 0 2023-03-31T05:23:47.5411744Z ##[endgroup] 2023-03-31T05:23:47.5412105Z ##[group]Setting up auth 2023-03-31T05:23:47.5418192Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2023-03-31T05:23:47.5449329Z [command]/usr/bin/git submodule foreach --recursive git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || : 2023-03-31T05:23:47.5704790Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2023-03-31T05:23:47.5736343Z [command]/usr/bin/git submodule foreach --recursive git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || : 2023-03-31T05:23:47.5995545Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2023-03-31T05:23:47.6042748Z ##[endgroup] 2023-03-31T05:23:47.6043311Z ##[group]Fetching the repository 2023-03-31T05:23:47.6050032Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --quiet --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2023-03-31T05:24:55.8340610Z [command]/usr/bin/git rev-parse --verify --quiet 4af0d9f59b3f9dcba6d4663724824d1665313b5d^{object} 2023-03-31T05:24:55.8370548Z 4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:24:55.8375199Z ##[endgroup] 2023-03-31T05:24:55.8375724Z ##[group]Determining the checkout info 2023-03-31T05:24:55.8376519Z ##[endgroup] 2023-03-31T05:24:55.8377090Z ##[group]Checking out the ref 2023-03-31T05:24:55.8380942Z [command]/usr/bin/git checkout --quiet --force 4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:24:57.2288568Z ##[endgroup] 2023-03-31T05:24:57.2289095Z ##[group]Setting up auth for fetching submodules 2023-03-31T05:24:57.2294852Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2023-03-31T05:24:57.2342571Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2023-03-31T05:24:57.2375979Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2023-03-31T05:24:57.2407449Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2023-03-31T05:24:57.2435832Z ##[endgroup] 2023-03-31T05:24:57.2436185Z ##[group]Fetching submodules 2023-03-31T05:24:57.2440289Z [command]/usr/bin/git submodule sync --recursive 2023-03-31T05:24:57.2714851Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2023-03-31T05:24:57.2984588Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2023-03-31T05:24:57.2986133Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2023-03-31T05:24:57.2986780Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2023-03-31T05:24:57.2988326Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2023-03-31T05:24:57.2990524Z Submodule 'third_party/QNNPACK' (https://github.com/pytorch/QNNPACK) registered for path 'third_party/QNNPACK' 2023-03-31T05:24:57.2992924Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2023-03-31T05:24:57.2995367Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2023-03-31T05:24:57.2997799Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2023-03-31T05:24:57.3000397Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2023-03-31T05:24:57.3003959Z Submodule 'third_party/cub' (https://github.com/NVlabs/cub.git) registered for path 'third_party/cub' 2023-03-31T05:24:57.3007916Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2023-03-31T05:24:57.3010580Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2023-03-31T05:24:57.3013598Z Submodule 'third_party/eigen' (https://gitlab.com/libeigen/eigen.git) registered for path 'third_party/eigen' 2023-03-31T05:24:57.3016831Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2023-03-31T05:24:57.3020070Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2023-03-31T05:24:57.3023386Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2023-03-31T05:24:57.3027059Z Submodule 'third_party/foxi' (https://github.com/houseroad/foxi.git) registered for path 'third_party/foxi' 2023-03-31T05:24:57.3030778Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:24:57.3034479Z Submodule 'third_party/gloo' (https://github.com/facebookincubator/gloo) registered for path 'third_party/gloo' 2023-03-31T05:24:57.3038235Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2023-03-31T05:24:57.3042207Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2023-03-31T05:24:57.3046545Z Submodule 'third_party/ios-cmake' (https://github.com/Yangqing/ios-cmake.git) registered for path 'third_party/ios-cmake' 2023-03-31T05:24:57.3050730Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2023-03-31T05:24:57.3054793Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2023-03-31T05:24:57.3058979Z Submodule 'third_party/nccl/nccl' (https://github.com/NVIDIA/nccl) registered for path 'third_party/nccl/nccl' 2023-03-31T05:24:57.3063317Z Submodule 'third_party/neon2sse' (https://github.com/intel/ARM_NEON_2_x86_SSE.git) registered for path 'third_party/neon2sse' 2023-03-31T05:24:57.3067882Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2023-03-31T05:24:57.3072365Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2023-03-31T05:24:57.3077088Z Submodule 'third_party/onnx-tensorrt' (https://github.com/onnx/onnx-tensorrt) registered for path 'third_party/onnx-tensorrt' 2023-03-31T05:24:57.3081718Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2023-03-31T05:24:57.3087193Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2023-03-31T05:24:57.3092163Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2023-03-31T05:24:57.3097365Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2023-03-31T05:24:57.3102569Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2023-03-31T05:24:57.3108081Z Submodule 'third_party/python-enum' (https://github.com/PeachPy/enum34.git) registered for path 'third_party/python-enum' 2023-03-31T05:24:57.3113284Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2023-03-31T05:24:57.3118790Z Submodule 'third_party/python-six' (https://github.com/benjaminp/six.git) registered for path 'third_party/python-six' 2023-03-31T05:24:57.3125222Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2023-03-31T05:24:57.3130918Z Submodule 'third_party/tbb' (https://github.com/01org/tbb) registered for path 'third_party/tbb' 2023-03-31T05:24:57.3136707Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2023-03-31T05:24:57.3142532Z Submodule 'third_party/zstd' (https://github.com/facebook/zstd.git) registered for path 'third_party/zstd' 2023-03-31T05:24:57.3170686Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2023-03-31T05:24:57.5947843Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2023-03-31T05:24:57.7968300Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2023-03-31T05:24:58.0157189Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2023-03-31T05:24:58.3140217Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/QNNPACK'... 2023-03-31T05:24:58.5669808Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2023-03-31T05:25:00.7546676Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2023-03-31T05:25:08.4411890Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2023-03-31T05:25:08.8618120Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2023-03-31T05:25:09.3831806Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cub'... 2023-03-31T05:25:10.9352415Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2023-03-31T05:25:12.2782754Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2023-03-31T05:25:14.0585923Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/eigen'... 2023-03-31T05:25:19.4682840Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2023-03-31T05:25:20.1286739Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2023-03-31T05:25:21.7088471Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2023-03-31T05:25:22.8924074Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/foxi'... 2023-03-31T05:25:23.0939313Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2023-03-31T05:25:23.5248331Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2023-03-31T05:25:23.8182656Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2023-03-31T05:25:25.0583789Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2023-03-31T05:25:25.5105650Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ios-cmake'... 2023-03-31T05:25:25.7216791Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2023-03-31T05:25:25.9904095Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2023-03-31T05:25:27.3666984Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nccl/nccl'... 2023-03-31T05:25:27.7374658Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/neon2sse'... 2023-03-31T05:25:28.0572710Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2023-03-31T05:25:34.3566247Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2023-03-31T05:25:36.3154392Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx-tensorrt'... 2023-03-31T05:25:36.9097809Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2023-03-31T05:25:37.1880487Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2023-03-31T05:25:43.2798053Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2023-03-31T05:25:43.4910002Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2023-03-31T05:25:43.7027969Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2023-03-31T05:25:44.5294845Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-enum'... 2023-03-31T05:25:44.7327139Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2023-03-31T05:25:45.0447832Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-six'... 2023-03-31T05:25:45.3291269Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2023-03-31T05:25:45.8717515Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tbb'... 2023-03-31T05:25:47.6968359Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2023-03-31T05:25:48.1638209Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/zstd'... 2023-03-31T05:25:50.7227506Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2023-03-31T05:25:50.7328621Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2023-03-31T05:25:50.7408412Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2023-03-31T05:25:50.7626161Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2023-03-31T05:25:50.7845287Z Submodule path 'third_party/QNNPACK': checked out '7d2a4e9931a82adc3814275b6219a03e24e36b4c' 2023-03-31T05:25:50.8214863Z Submodule path 'third_party/VulkanMemoryAllocator': checked out 'a6bfc237255a6bac1513f7c1ebde6d8aed6b5191' 2023-03-31T05:25:51.4448918Z Submodule path 'third_party/XNNPACK': checked out '51a987591a6fc9f0fc0707077f53d763ac132cbf' 2023-03-31T05:25:51.4654590Z Submodule path 'third_party/benchmark': checked out '0d98dba29d66e93259db7daa53a9327df767a415' 2023-03-31T05:25:51.5615055Z Submodule path 'third_party/cpuinfo': checked out '8ec7bd91ad0470e61cf38f618cc1f270dede599c' 2023-03-31T05:25:51.5948057Z Submodule path 'third_party/cub': checked out 'd106ddb991a56c3df1b6d51b2409e36ba8181ce4' 2023-03-31T05:25:51.8746479Z Submodule path 'third_party/cudnn_frontend': checked out '81a041a68245cd8f871c43bbbbd5b6b627979a30' 2023-03-31T05:25:52.2647721Z Submodule path 'third_party/cutlass': checked out 'b72cbf957df8cf84a6d0ff91c190ad51a9c1d24a' 2023-03-31T05:25:52.4986886Z Submodule path 'third_party/eigen': checked out '3147391d946bb4b6c68edd901f2add6ac1f31f8c' 2023-03-31T05:25:52.5456545Z Submodule path 'third_party/fbgemm': checked out '03b2046676707da64504e898490ab46104d4682a' 2023-03-31T05:25:52.5471259Z Submodule 'third_party/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:25:52.5473156Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:25:52.5475241Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:25:52.5477550Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:25:52.5479812Z Submodule 'third_party/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:25:52.5504230Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/third_party/asmjit'... 2023-03-31T05:25:53.1965548Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/third_party/cpuinfo'... 2023-03-31T05:25:53.6964586Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/third_party/cutlass'... 2023-03-31T05:25:55.8021663Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/third_party/googletest'... 2023-03-31T05:25:57.0343246Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/third_party/hipify_torch'... 2023-03-31T05:25:57.3460754Z Submodule path 'third_party/fbgemm/third_party/asmjit': checked out 'd3fbf7c9bc7c1d1365a94a45614b91c5a3706b81' 2023-03-31T05:25:57.4430518Z Submodule path 'third_party/fbgemm/third_party/cpuinfo': checked out 'ed8b86a253800bafdb7b25c5c399f91bff9cb1f3' 2023-03-31T05:25:57.8316887Z Submodule path 'third_party/fbgemm/third_party/cutlass': checked out 'fc9ebc645b63f3a6bc80aaefde5c063fb72110d6' 2023-03-31T05:25:57.8915359Z Submodule path 'third_party/fbgemm/third_party/googletest': checked out 'cbf019de22c8dd37b2108da35b2748fd702d1796' 2023-03-31T05:25:57.9019983Z Submodule path 'third_party/fbgemm/third_party/hipify_torch': checked out '1840658c184f3eeba787dae0f06c45756c1daaf5' 2023-03-31T05:25:57.9869957Z Submodule path 'third_party/flatbuffers': checked out 'd0cede9c90c5257537c293517a21376408b549fa' 2023-03-31T05:25:58.0187849Z Submodule path 'third_party/fmt': checked out 'a33701196adfad74917046096bf5a2aa0ab0bb50' 2023-03-31T05:25:58.0277047Z Submodule path 'third_party/foxi': checked out 'c278588e34e535f0bb8f00df3880d26928038cad' 2023-03-31T05:25:58.0660388Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2023-03-31T05:25:58.0893168Z Submodule path 'third_party/gloo': checked out '10909297fedab0a680799211a299203e53515032' 2023-03-31T05:25:58.1333032Z Submodule path 'third_party/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2023-03-31T05:25:58.1449801Z Submodule path 'third_party/ideep': checked out '7bc3e12f7c0cad7fb24f8d4ab63dcd467ffa60c7' 2023-03-31T05:25:58.1464444Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2023-03-31T05:25:58.1487099Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2023-03-31T05:26:06.8064883Z Submodule path 'third_party/ideep/mkl-dnn': checked out '6dbeffbae1f23cbbeae17adb7b5b13f1f37c080e' 2023-03-31T05:26:06.8166680Z Submodule path 'third_party/ios-cmake': checked out '8abaed637d56f1337d6e1d2c4026e25c1eade724' 2023-03-31T05:26:06.8308132Z Submodule path 'third_party/ittapi': checked out '5b8a7d7422611c3a0d799fb5fc5dd4abfae35b42' 2023-03-31T05:26:06.9209594Z Submodule path 'third_party/kineto': checked out 'e64df4dc31285a6129a74d26d67365cedf7aa6d1' 2023-03-31T05:26:06.9225324Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:06.9227450Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:06.9229982Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:06.9254800Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2023-03-31T05:26:07.3350165Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2023-03-31T05:26:08.6539685Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2023-03-31T05:26:09.9649087Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out '7d04a0053a845370ae06ce317a22a48e9edcc74e' 2023-03-31T05:26:09.9663610Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:09.9665641Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:09.9667992Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:09.9670369Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:09.9672888Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:09.9675479Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:09.9678163Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:09.9680973Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:09.9706337Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2023-03-31T05:26:10.6803328Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2023-03-31T05:26:11.0108121Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2023-03-31T05:26:12.2078240Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2023-03-31T05:26:12.5484546Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2023-03-31T05:26:13.0178957Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2023-03-31T05:26:14.2288171Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2023-03-31T05:26:20.7266788Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2023-03-31T05:26:21.1199240Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2023-03-31T05:26:21.1355256Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2023-03-31T05:26:21.1704030Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2023-03-31T05:26:21.1825867Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2023-03-31T05:26:21.1839416Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:21.1863016Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2023-03-31T05:26:21.5281713Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2023-03-31T05:26:21.5449898Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2023-03-31T05:26:21.5839235Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '58d77fa8070e8cec2dc1ed015d66b454c8d78850' 2023-03-31T05:26:21.6830297Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2023-03-31T05:26:21.6983097Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2023-03-31T05:26:21.7300553Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out 'a33701196adfad74917046096bf5a2aa0ab0bb50' 2023-03-31T05:26:21.7842364Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '7aca84427f224eeed3144123d5230d5871e93347' 2023-03-31T05:26:21.8049033Z Submodule path 'third_party/nccl/nccl': checked out 'f89fd4777d2ef9229c039ff750ae21da01626f52' 2023-03-31T05:26:21.8189587Z Submodule path 'third_party/neon2sse': checked out '97a126f08ce318023be604d03f88bf0820a9464a' 2023-03-31T05:26:21.9224209Z Submodule path 'third_party/nlohmann': checked out '87cda1d6646592ac5866dc703c8e1839046a6806' 2023-03-31T05:26:22.2023110Z Submodule path 'third_party/onnx': checked out '389b6bcb05b9479d149d29b2461fbffe8472ed14' 2023-03-31T05:26:22.2053169Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:22.2054814Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:22.2080099Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/benchmark'... 2023-03-31T05:26:22.7093467Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2023-03-31T05:26:23.5563997Z Submodule path 'third_party/onnx/third_party/benchmark': checked out '0d98dba29d66e93259db7daa53a9327df767a415' 2023-03-31T05:26:23.5886187Z Submodule path 'third_party/onnx/third_party/pybind11': checked out '914c06fb252b6cc3727d0eedab6736e88a3fcb01' 2023-03-31T05:26:23.6057025Z Submodule path 'third_party/onnx-tensorrt': checked out 'c153211418a7c57ce071d9ce2a41f8d1c85a878f' 2023-03-31T05:26:23.6071787Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:23.6094713Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx-tensorrt/third_party/onnx'... 2023-03-31T05:26:25.9633328Z Submodule path 'third_party/onnx-tensorrt/third_party/onnx': checked out '765f5ee823a67a866f4bd28a9860e81f3c811ce8' 2023-03-31T05:26:25.9653443Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:25.9655317Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:25.9679966Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark'... 2023-03-31T05:26:26.4128096Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11'... 2023-03-31T05:26:27.2796401Z Submodule path 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark': checked out 'e776aa0275e293707b6a0901e0e8d8a8a3679508' 2023-03-31T05:26:27.3466003Z Submodule path 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11': checked out 'a1041190c8b8ff0cd9e2f0752248ad5e3789ea0c' 2023-03-31T05:26:27.3479899Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:27.3504045Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang'... 2023-03-31T05:26:27.5520289Z Submodule path 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2023-03-31T05:26:27.5609094Z Submodule path 'third_party/pocketfft': checked out 'ea778e37710c07723435b1be58235996d1d43a5a' 2023-03-31T05:26:27.8053510Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2023-03-31T05:26:27.8072302Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:27.8073978Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:27.8098224Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2023-03-31T05:26:28.3987911Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2023-03-31T05:26:29.6312170Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2023-03-31T05:26:29.6995081Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2023-03-31T05:26:29.7079571Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2023-03-31T05:26:29.7184028Z Submodule path 'third_party/pthreadpool': checked out 'a134dd5d4cee80cce15db81a72e7f929d71dd413' 2023-03-31T05:26:29.7496828Z Submodule path 'third_party/pybind11': checked out '80dc998efced8ceb2be59756668a7e90e8bef917' 2023-03-31T05:26:29.7585074Z Submodule path 'third_party/python-enum': checked out '4cfedc426c4e2fc52e3f5c2b4297e15ed8d6b8c7' 2023-03-31T05:26:29.7862088Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2023-03-31T05:26:29.7953848Z Submodule path 'third_party/python-six': checked out '15e31431af97e5e64b80af0a3f598d382bcdd49a' 2023-03-31T05:26:29.8377118Z Submodule path 'third_party/sleef': checked out 'e0a003ee838b75d11763aa9c3ef17bf71a725bff' 2023-03-31T05:26:29.9405151Z Submodule path 'third_party/tbb': checked out 'a51a90bc609bb73db8ea13841b5cf7aa4344d4a9' 2023-03-31T05:26:29.9649060Z Submodule path 'third_party/tensorpipe': checked out '52791a2fd214b2a9dc5759d36725909c1daa7f2e' 2023-03-31T05:26:29.9664634Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:29.9666716Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:29.9669161Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:29.9671725Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:29.9694725Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2023-03-31T05:26:31.2057957Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2023-03-31T05:26:31.4534455Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2023-03-31T05:26:32.6734186Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2023-03-31T05:26:33.5496028Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2023-03-31T05:26:33.5636585Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2023-03-31T05:26:33.6243363Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '1dff88e5161cba5c59276d2070d2e304e4dcb242' 2023-03-31T05:26:33.6516460Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2023-03-31T05:26:33.6530661Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:33.6553995Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2023-03-31T05:26:33.8766714Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2023-03-31T05:26:34.0119806Z Submodule path 'third_party/zstd': checked out 'aec56a52fbab207fc639a1937d1e708a282edca8' 2023-03-31T05:26:34.0148932Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2023-03-31T05:26:34.0426997Z Entering 'android/libs/fbjni' 2023-03-31T05:26:34.0465968Z Entering 'third_party/FP16' 2023-03-31T05:26:34.0505011Z Entering 'third_party/FXdiv' 2023-03-31T05:26:34.0544988Z Entering 'third_party/NNPACK' 2023-03-31T05:26:34.0583946Z Entering 'third_party/QNNPACK' 2023-03-31T05:26:34.0622238Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T05:26:34.0660576Z Entering 'third_party/XNNPACK' 2023-03-31T05:26:34.0711158Z Entering 'third_party/benchmark' 2023-03-31T05:26:34.0748062Z Entering 'third_party/cpuinfo' 2023-03-31T05:26:34.0785503Z Entering 'third_party/cub' 2023-03-31T05:26:34.0823565Z Entering 'third_party/cudnn_frontend' 2023-03-31T05:26:34.0866463Z Entering 'third_party/cutlass' 2023-03-31T05:26:34.0910070Z Entering 'third_party/eigen' 2023-03-31T05:26:34.0949190Z Entering 'third_party/fbgemm' 2023-03-31T05:26:34.0987672Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:26:34.1025003Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:26:34.1062618Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:26:34.1105944Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:26:34.1143360Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:26:34.1183038Z Entering 'third_party/flatbuffers' 2023-03-31T05:26:34.1222241Z Entering 'third_party/fmt' 2023-03-31T05:26:34.1259522Z Entering 'third_party/foxi' 2023-03-31T05:26:34.1297359Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:26:34.1334952Z Entering 'third_party/gloo' 2023-03-31T05:26:34.1372364Z Entering 'third_party/googletest' 2023-03-31T05:26:34.1410183Z Entering 'third_party/ideep' 2023-03-31T05:26:34.1446531Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T05:26:34.1491689Z Entering 'third_party/ios-cmake' 2023-03-31T05:26:34.1528940Z Entering 'third_party/ittapi' 2023-03-31T05:26:34.1565792Z Entering 'third_party/kineto' 2023-03-31T05:26:34.1603446Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:34.1641744Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:34.1683581Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:34.1721409Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:34.1758524Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:34.1794712Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:34.1833627Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:34.1871299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:34.1907207Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:34.1944362Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:34.1983023Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:34.2020730Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:34.2057894Z Entering 'third_party/nccl/nccl' 2023-03-31T05:26:34.2096974Z Entering 'third_party/neon2sse' 2023-03-31T05:26:34.2133461Z Entering 'third_party/nlohmann' 2023-03-31T05:26:34.2171597Z Entering 'third_party/onnx' 2023-03-31T05:26:34.2222487Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.2259407Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.2298700Z Entering 'third_party/onnx-tensorrt' 2023-03-31T05:26:34.2335674Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:34.2376321Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.2413107Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.2449670Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:34.2491549Z Entering 'third_party/pocketfft' 2023-03-31T05:26:34.2529317Z Entering 'third_party/protobuf' 2023-03-31T05:26:34.2569539Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:34.2605173Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:34.2642110Z Entering 'third_party/psimd' 2023-03-31T05:26:34.2678596Z Entering 'third_party/pthreadpool' 2023-03-31T05:26:34.2715019Z Entering 'third_party/pybind11' 2023-03-31T05:26:34.2751192Z Entering 'third_party/python-enum' 2023-03-31T05:26:34.2788562Z Entering 'third_party/python-peachpy' 2023-03-31T05:26:34.2826705Z Entering 'third_party/python-six' 2023-03-31T05:26:34.2863274Z Entering 'third_party/sleef' 2023-03-31T05:26:34.2900636Z Entering 'third_party/tbb' 2023-03-31T05:26:34.2939598Z Entering 'third_party/tensorpipe' 2023-03-31T05:26:34.2976320Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:34.3012324Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:34.3048447Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:34.3084142Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:34.3120619Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:34.3160833Z Entering 'third_party/zstd' 2023-03-31T05:26:34.3208680Z ##[endgroup] 2023-03-31T05:26:34.3210056Z ##[group]Persisting credentials for submodules 2023-03-31T05:26:34.3214902Z [command]/usr/bin/git submodule foreach --recursive git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || : 2023-03-31T05:26:34.3488296Z Entering 'android/libs/fbjni' 2023-03-31T05:26:34.3526528Z Entering 'third_party/FP16' 2023-03-31T05:26:34.3562047Z Entering 'third_party/FXdiv' 2023-03-31T05:26:34.3597950Z Entering 'third_party/NNPACK' 2023-03-31T05:26:34.3634369Z Entering 'third_party/QNNPACK' 2023-03-31T05:26:34.3670890Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T05:26:34.3708277Z Entering 'third_party/XNNPACK' 2023-03-31T05:26:34.3759004Z Entering 'third_party/benchmark' 2023-03-31T05:26:34.3797008Z Entering 'third_party/cpuinfo' 2023-03-31T05:26:34.3834648Z Entering 'third_party/cub' 2023-03-31T05:26:34.3871373Z Entering 'third_party/cudnn_frontend' 2023-03-31T05:26:34.3914554Z Entering 'third_party/cutlass' 2023-03-31T05:26:34.3958226Z Entering 'third_party/eigen' 2023-03-31T05:26:34.3996109Z Entering 'third_party/fbgemm' 2023-03-31T05:26:34.4034666Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:26:34.4070333Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:26:34.4106516Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:26:34.4150031Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:26:34.4184558Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:26:34.4223256Z Entering 'third_party/flatbuffers' 2023-03-31T05:26:34.4261702Z Entering 'third_party/fmt' 2023-03-31T05:26:34.4298121Z Entering 'third_party/foxi' 2023-03-31T05:26:34.4335108Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:26:34.4371447Z Entering 'third_party/gloo' 2023-03-31T05:26:34.4407909Z Entering 'third_party/googletest' 2023-03-31T05:26:34.4444256Z Entering 'third_party/ideep' 2023-03-31T05:26:34.4479324Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T05:26:34.4524374Z Entering 'third_party/ios-cmake' 2023-03-31T05:26:34.4560212Z Entering 'third_party/ittapi' 2023-03-31T05:26:34.4598026Z Entering 'third_party/kineto' 2023-03-31T05:26:34.4635465Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:34.4671811Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:34.4709233Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:34.4746043Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:34.4782638Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:34.4818070Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:34.4856134Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:34.4892827Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:34.4929275Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:34.4966803Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:34.5005208Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:34.5041049Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:34.5078438Z Entering 'third_party/nccl/nccl' 2023-03-31T05:26:34.5117814Z Entering 'third_party/neon2sse' 2023-03-31T05:26:34.5154253Z Entering 'third_party/nlohmann' 2023-03-31T05:26:34.5192977Z Entering 'third_party/onnx' 2023-03-31T05:26:34.5245253Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.5282598Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.5322102Z Entering 'third_party/onnx-tensorrt' 2023-03-31T05:26:34.5359576Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:34.5401242Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.5437189Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.5472898Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:34.5514506Z Entering 'third_party/pocketfft' 2023-03-31T05:26:34.5552459Z Entering 'third_party/protobuf' 2023-03-31T05:26:34.5592881Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:34.5629665Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:34.5668432Z Entering 'third_party/psimd' 2023-03-31T05:26:34.5704729Z Entering 'third_party/pthreadpool' 2023-03-31T05:26:34.5741275Z Entering 'third_party/pybind11' 2023-03-31T05:26:34.5777032Z Entering 'third_party/python-enum' 2023-03-31T05:26:34.5813811Z Entering 'third_party/python-peachpy' 2023-03-31T05:26:34.5850630Z Entering 'third_party/python-six' 2023-03-31T05:26:34.5889523Z Entering 'third_party/sleef' 2023-03-31T05:26:34.5926015Z Entering 'third_party/tbb' 2023-03-31T05:26:34.5965859Z Entering 'third_party/tensorpipe' 2023-03-31T05:26:34.6003653Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:34.6040548Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:34.6078465Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:34.6115625Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:34.6151441Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:34.6190816Z Entering 'third_party/zstd' 2023-03-31T05:26:34.6241389Z [command]/usr/bin/git submodule foreach --recursive git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url 2023-03-31T05:26:34.6516839Z Entering 'android/libs/fbjni' 2023-03-31T05:26:34.6550288Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2023-03-31T05:26:34.6565734Z Entering 'third_party/FP16' 2023-03-31T05:26:34.6599762Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2023-03-31T05:26:34.6615029Z Entering 'third_party/FXdiv' 2023-03-31T05:26:34.6649820Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2023-03-31T05:26:34.6664076Z Entering 'third_party/NNPACK' 2023-03-31T05:26:34.6700503Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2023-03-31T05:26:34.6714475Z Entering 'third_party/QNNPACK' 2023-03-31T05:26:34.6748879Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/QNNPACK/config remote.origin.url 2023-03-31T05:26:34.6763986Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T05:26:34.6799394Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2023-03-31T05:26:34.6814277Z Entering 'third_party/XNNPACK' 2023-03-31T05:26:34.6848544Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2023-03-31T05:26:34.6877608Z Entering 'third_party/benchmark' 2023-03-31T05:26:34.6912745Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2023-03-31T05:26:34.6928902Z Entering 'third_party/cpuinfo' 2023-03-31T05:26:34.6962846Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2023-03-31T05:26:34.6979434Z Entering 'third_party/cub' 2023-03-31T05:26:34.7015103Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cub/config remote.origin.url 2023-03-31T05:26:34.7030084Z Entering 'third_party/cudnn_frontend' 2023-03-31T05:26:34.7065335Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2023-03-31T05:26:34.7087044Z Entering 'third_party/cutlass' 2023-03-31T05:26:34.7122819Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2023-03-31T05:26:34.7145265Z Entering 'third_party/eigen' 2023-03-31T05:26:34.7180632Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/eigen/config remote.origin.url 2023-03-31T05:26:34.7198086Z Entering 'third_party/fbgemm' 2023-03-31T05:26:34.7234019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2023-03-31T05:26:34.7250226Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:26:34.7285925Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/third_party/asmjit/config remote.origin.url 2023-03-31T05:26:34.7299604Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:26:34.7333998Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/third_party/cpuinfo/config remote.origin.url 2023-03-31T05:26:34.7349204Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:26:34.7382564Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/third_party/cutlass/config remote.origin.url 2023-03-31T05:26:34.7405349Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:26:34.7438481Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.7452978Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:26:34.7486778Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/third_party/hipify_torch/config remote.origin.url 2023-03-31T05:26:34.7502174Z Entering 'third_party/flatbuffers' 2023-03-31T05:26:34.7537702Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2023-03-31T05:26:34.7554865Z Entering 'third_party/fmt' 2023-03-31T05:26:34.7590670Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2023-03-31T05:26:34.7606121Z Entering 'third_party/foxi' 2023-03-31T05:26:34.7640550Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/foxi/config remote.origin.url 2023-03-31T05:26:34.7654954Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:26:34.7689332Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2023-03-31T05:26:34.7703886Z Entering 'third_party/gloo' 2023-03-31T05:26:34.7737482Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2023-03-31T05:26:34.7752933Z Entering 'third_party/googletest' 2023-03-31T05:26:34.7787033Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.7802519Z Entering 'third_party/ideep' 2023-03-31T05:26:34.7837427Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2023-03-31T05:26:34.7852950Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T05:26:34.7885991Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2023-03-31T05:26:34.7910327Z Entering 'third_party/ios-cmake' 2023-03-31T05:26:34.7945385Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ios-cmake/config remote.origin.url 2023-03-31T05:26:34.7959941Z Entering 'third_party/ittapi' 2023-03-31T05:26:34.7994180Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2023-03-31T05:26:34.8010935Z Entering 'third_party/kineto' 2023-03-31T05:26:34.8046203Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2023-03-31T05:26:34.8060798Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:34.8098781Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2023-03-31T05:26:34.8114337Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:34.8150873Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2023-03-31T05:26:34.8166699Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:34.8200254Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2023-03-31T05:26:34.8215495Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:34.8249188Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2023-03-31T05:26:34.8263929Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:34.8298896Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2023-03-31T05:26:34.8313413Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:34.8348881Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2023-03-31T05:26:34.8365392Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:34.8399585Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2023-03-31T05:26:34.8416807Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:34.8450514Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.8465877Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:34.8501111Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2023-03-31T05:26:34.8516946Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:34.8550128Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2023-03-31T05:26:34.8567133Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:34.8600452Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2023-03-31T05:26:34.8615145Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:34.8649019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.8665596Z Entering 'third_party/nccl/nccl' 2023-03-31T05:26:34.8701027Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nccl/nccl/config remote.origin.url 2023-03-31T05:26:34.8716718Z Entering 'third_party/neon2sse' 2023-03-31T05:26:34.8750679Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/neon2sse/config remote.origin.url 2023-03-31T05:26:34.8765449Z Entering 'third_party/nlohmann' 2023-03-31T05:26:34.8800500Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2023-03-31T05:26:34.8816959Z Entering 'third_party/onnx' 2023-03-31T05:26:34.8852446Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2023-03-31T05:26:34.8881062Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.8916595Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/benchmark/config remote.origin.url 2023-03-31T05:26:34.8933365Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.8969901Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2023-03-31T05:26:34.8987617Z Entering 'third_party/onnx-tensorrt' 2023-03-31T05:26:34.9024229Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx-tensorrt/config remote.origin.url 2023-03-31T05:26:34.9038343Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:34.9072328Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx-tensorrt/modules/third_party/onnx/config remote.origin.url 2023-03-31T05:26:34.9092722Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:34.9126446Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx-tensorrt/modules/third_party/onnx/modules/third_party/benchmark/config remote.origin.url 2023-03-31T05:26:34.9141061Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:34.9176712Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx-tensorrt/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2023-03-31T05:26:34.9191771Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:34.9226832Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx-tensorrt/modules/third_party/onnx/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2023-03-31T05:26:34.9246890Z Entering 'third_party/pocketfft' 2023-03-31T05:26:34.9282208Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2023-03-31T05:26:34.9297031Z Entering 'third_party/protobuf' 2023-03-31T05:26:34.9332371Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2023-03-31T05:26:34.9351029Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:34.9385746Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2023-03-31T05:26:34.9399737Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:34.9435713Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.9453735Z Entering 'third_party/psimd' 2023-03-31T05:26:34.9489004Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2023-03-31T05:26:34.9503237Z Entering 'third_party/pthreadpool' 2023-03-31T05:26:34.9538815Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2023-03-31T05:26:34.9553500Z Entering 'third_party/pybind11' 2023-03-31T05:26:34.9588424Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2023-03-31T05:26:34.9604800Z Entering 'third_party/python-enum' 2023-03-31T05:26:34.9638462Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-enum/config remote.origin.url 2023-03-31T05:26:34.9653654Z Entering 'third_party/python-peachpy' 2023-03-31T05:26:34.9689191Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2023-03-31T05:26:34.9704124Z Entering 'third_party/python-six' 2023-03-31T05:26:34.9738997Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-six/config remote.origin.url 2023-03-31T05:26:34.9754186Z Entering 'third_party/sleef' 2023-03-31T05:26:34.9788055Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2023-03-31T05:26:34.9804220Z Entering 'third_party/tbb' 2023-03-31T05:26:34.9837411Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tbb/config remote.origin.url 2023-03-31T05:26:34.9856062Z Entering 'third_party/tensorpipe' 2023-03-31T05:26:34.9890951Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2023-03-31T05:26:34.9905782Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:34.9941517Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2023-03-31T05:26:34.9956405Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:34.9991071Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2023-03-31T05:26:35.0005135Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:35.0041221Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2023-03-31T05:26:35.0056571Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:35.0093172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2023-03-31T05:26:35.0108108Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:35.0143599Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2023-03-31T05:26:35.0161540Z Entering 'third_party/zstd' 2023-03-31T05:26:35.0195835Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/zstd/config remote.origin.url 2023-03-31T05:26:35.0851693Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2023-03-31T05:26:35.1130915Z Entering 'android/libs/fbjni' 2023-03-31T05:26:35.1173108Z Entering 'third_party/FP16' 2023-03-31T05:26:35.1211254Z Entering 'third_party/FXdiv' 2023-03-31T05:26:35.1249350Z Entering 'third_party/NNPACK' 2023-03-31T05:26:35.1289277Z Entering 'third_party/QNNPACK' 2023-03-31T05:26:35.1327375Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T05:26:35.1364587Z Entering 'third_party/XNNPACK' 2023-03-31T05:26:35.1414973Z Entering 'third_party/benchmark' 2023-03-31T05:26:35.1452795Z Entering 'third_party/cpuinfo' 2023-03-31T05:26:35.1491135Z Entering 'third_party/cub' 2023-03-31T05:26:35.1529114Z Entering 'third_party/cudnn_frontend' 2023-03-31T05:26:35.1571487Z Entering 'third_party/cutlass' 2023-03-31T05:26:35.1615203Z Entering 'third_party/eigen' 2023-03-31T05:26:35.1654985Z Entering 'third_party/fbgemm' 2023-03-31T05:26:35.1694215Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:26:35.1730437Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:26:35.1767093Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:26:35.1810087Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:26:35.1846460Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:26:35.1884513Z Entering 'third_party/flatbuffers' 2023-03-31T05:26:35.1925273Z Entering 'third_party/fmt' 2023-03-31T05:26:35.1964252Z Entering 'third_party/foxi' 2023-03-31T05:26:35.2002778Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:26:35.2041877Z Entering 'third_party/gloo' 2023-03-31T05:26:35.2080007Z Entering 'third_party/googletest' 2023-03-31T05:26:35.2117683Z Entering 'third_party/ideep' 2023-03-31T05:26:35.2155354Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T05:26:35.2201899Z Entering 'third_party/ios-cmake' 2023-03-31T05:26:35.2240070Z Entering 'third_party/ittapi' 2023-03-31T05:26:35.2278107Z Entering 'third_party/kineto' 2023-03-31T05:26:35.2317195Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:35.2354694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:35.2392999Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:35.2429425Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:35.2467044Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:35.2503009Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:35.2540715Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:35.2577203Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:35.2613689Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:35.2651062Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:35.2688305Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:35.2724934Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:35.2762871Z Entering 'third_party/nccl/nccl' 2023-03-31T05:26:35.2800603Z Entering 'third_party/neon2sse' 2023-03-31T05:26:35.2836329Z Entering 'third_party/nlohmann' 2023-03-31T05:26:35.2874445Z Entering 'third_party/onnx' 2023-03-31T05:26:35.2926745Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:35.2965363Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:35.3004933Z Entering 'third_party/onnx-tensorrt' 2023-03-31T05:26:35.3041187Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:35.3083523Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:35.3120617Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:35.3158350Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:35.3199988Z Entering 'third_party/pocketfft' 2023-03-31T05:26:35.3237723Z Entering 'third_party/protobuf' 2023-03-31T05:26:35.3279224Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:35.3316911Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:35.3356204Z Entering 'third_party/psimd' 2023-03-31T05:26:35.3395299Z Entering 'third_party/pthreadpool' 2023-03-31T05:26:35.3433634Z Entering 'third_party/pybind11' 2023-03-31T05:26:35.3470629Z Entering 'third_party/python-enum' 2023-03-31T05:26:35.3508097Z Entering 'third_party/python-peachpy' 2023-03-31T05:26:35.3546443Z Entering 'third_party/python-six' 2023-03-31T05:26:35.3583250Z Entering 'third_party/sleef' 2023-03-31T05:26:35.3619330Z Entering 'third_party/tbb' 2023-03-31T05:26:35.3658740Z Entering 'third_party/tensorpipe' 2023-03-31T05:26:35.3696503Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:35.3733060Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:35.3768443Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:35.3808281Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:35.3845681Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:35.3885426Z Entering 'third_party/zstd' 2023-03-31T05:26:35.3934259Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2023-03-31T05:26:35.4205670Z Entering 'android/libs/fbjni' 2023-03-31T05:26:35.4242436Z Entering 'third_party/FP16' 2023-03-31T05:26:35.4281015Z Entering 'third_party/FXdiv' 2023-03-31T05:26:35.4317319Z Entering 'third_party/NNPACK' 2023-03-31T05:26:35.4353090Z Entering 'third_party/QNNPACK' 2023-03-31T05:26:35.4389930Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T05:26:35.4427044Z Entering 'third_party/XNNPACK' 2023-03-31T05:26:35.4476494Z Entering 'third_party/benchmark' 2023-03-31T05:26:35.4513564Z Entering 'third_party/cpuinfo' 2023-03-31T05:26:35.4550254Z Entering 'third_party/cub' 2023-03-31T05:26:35.4586364Z Entering 'third_party/cudnn_frontend' 2023-03-31T05:26:35.4628146Z Entering 'third_party/cutlass' 2023-03-31T05:26:35.4670398Z Entering 'third_party/eigen' 2023-03-31T05:26:35.4707967Z Entering 'third_party/fbgemm' 2023-03-31T05:26:35.4744285Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T05:26:35.4779661Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T05:26:35.4815840Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T05:26:35.4858841Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T05:26:35.4894115Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T05:26:35.4931391Z Entering 'third_party/flatbuffers' 2023-03-31T05:26:35.4971684Z Entering 'third_party/fmt' 2023-03-31T05:26:35.5008378Z Entering 'third_party/foxi' 2023-03-31T05:26:35.5045526Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T05:26:35.5082618Z Entering 'third_party/gloo' 2023-03-31T05:26:35.5120214Z Entering 'third_party/googletest' 2023-03-31T05:26:35.5158715Z Entering 'third_party/ideep' 2023-03-31T05:26:35.5193983Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T05:26:35.5240171Z Entering 'third_party/ios-cmake' 2023-03-31T05:26:35.5277312Z Entering 'third_party/ittapi' 2023-03-31T05:26:35.5314158Z Entering 'third_party/kineto' 2023-03-31T05:26:35.5351160Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T05:26:35.5388552Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T05:26:35.5426472Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T05:26:35.5464216Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T05:26:35.5501301Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T05:26:35.5536399Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T05:26:35.5574459Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T05:26:35.5612596Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T05:26:35.5649117Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T05:26:35.5687318Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T05:26:35.5724838Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T05:26:35.5761655Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T05:26:35.5800145Z Entering 'third_party/nccl/nccl' 2023-03-31T05:26:35.5838112Z Entering 'third_party/neon2sse' 2023-03-31T05:26:35.5877297Z Entering 'third_party/nlohmann' 2023-03-31T05:26:35.5915470Z Entering 'third_party/onnx' 2023-03-31T05:26:35.5968878Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T05:26:35.6006479Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T05:26:35.6045441Z Entering 'third_party/onnx-tensorrt' 2023-03-31T05:26:35.6082533Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T05:26:35.6124310Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T05:26:35.6160369Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T05:26:35.6196245Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T05:26:35.6236718Z Entering 'third_party/pocketfft' 2023-03-31T05:26:35.6273035Z Entering 'third_party/protobuf' 2023-03-31T05:26:35.6312806Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T05:26:35.6349021Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T05:26:35.6388225Z Entering 'third_party/psimd' 2023-03-31T05:26:35.6425068Z Entering 'third_party/pthreadpool' 2023-03-31T05:26:35.6461674Z Entering 'third_party/pybind11' 2023-03-31T05:26:35.6497624Z Entering 'third_party/python-enum' 2023-03-31T05:26:35.6534363Z Entering 'third_party/python-peachpy' 2023-03-31T05:26:35.6570145Z Entering 'third_party/python-six' 2023-03-31T05:26:35.6606619Z Entering 'third_party/sleef' 2023-03-31T05:26:35.6643320Z Entering 'third_party/tbb' 2023-03-31T05:26:35.6681573Z Entering 'third_party/tensorpipe' 2023-03-31T05:26:35.6719477Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T05:26:35.6756070Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T05:26:35.6792215Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T05:26:35.6828835Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T05:26:35.6864269Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T05:26:35.6903332Z Entering 'third_party/zstd' 2023-03-31T05:26:35.6950612Z ##[endgroup] 2023-03-31T05:26:35.6991417Z [command]/usr/bin/git log -1 --format='%H' 2023-03-31T05:26:35.7019860Z '4af0d9f59b3f9dcba6d4663724824d1665313b5d' 2023-03-31T05:26:35.7163853Z Prepare all required actions 2023-03-31T05:26:35.7164152Z Getting action download info 2023-03-31T05:26:35.8703737Z ##[group]Run ./.github/actions/setup-linux 2023-03-31T05:26:35.8703950Z env: 2023-03-31T05:26:35.8704133Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:35.8704331Z ##[endgroup] 2023-03-31T05:26:35.8765148Z ##[group]Run set -euo pipefail 2023-03-31T05:26:35.8765426Z set -euo pipefail 2023-03-31T05:26:35.8765647Z function get_ec2_metadata() { 2023-03-31T05:26:35.8765908Z  # Pulled from instance metadata endpoint for EC2 2023-03-31T05:26:35.8766284Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2023-03-31T05:26:35.8766601Z  category=$1 2023-03-31T05:26:35.8766850Z  # If it is GCP runner (runner name contains gcp), do not run this 2023-03-31T05:26:35.8767110Z  runner_name_str=i-0da0c45d38bcc5929 2023-03-31T05:26:35.8767409Z  if [[ $runner_name_str != *"gcp"* ]]; then 2023-03-31T05:26:35.8767675Z  curl -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2023-03-31T05:26:35.8767913Z  else 2023-03-31T05:26:35.8768159Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2023-03-31T05:26:35.8768398Z  fi 2023-03-31T05:26:35.8768551Z } 2023-03-31T05:26:35.8768758Z echo "ami-id: $(get_ec2_metadata ami-id)" 2023-03-31T05:26:35.8769036Z echo "instance-id: $(get_ec2_metadata instance-id)" 2023-03-31T05:26:35.8769310Z echo "instance-type: $(get_ec2_metadata instance-type)" 2023-03-31T05:26:35.8769571Z echo "system info $(uname -a)" 2023-03-31T05:26:35.8781094Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:26:35.8781447Z env: 2023-03-31T05:26:35.8781636Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:35.8781838Z ##[endgroup] 2023-03-31T05:26:35.8903362Z ami-id: ami-096198a0bccc6bad4 2023-03-31T05:26:35.8960377Z instance-id: i-0da0c45d38bcc5929 2023-03-31T05:26:35.9016222Z instance-type: c5.2xlarge 2023-03-31T05:26:35.9023579Z system info Linux ip-10-0-2-162.ec2.internal 4.14.252-195.483.amzn2.x86_64 #1 SMP Mon Nov 1 20:58:46 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux 2023-03-31T05:26:35.9044086Z ##[group]Run if systemctl is-active --quiet docker; then 2023-03-31T05:26:35.9044723Z if systemctl is-active --quiet docker; then 2023-03-31T05:26:35.9045063Z  echo "Docker daemon is running..."; 2023-03-31T05:26:35.9045381Z else 2023-03-31T05:26:35.9045704Z  echo "Starting docker deamon..." && sudo systemctl start docker; 2023-03-31T05:26:35.9045984Z fi 2023-03-31T05:26:35.9057551Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:26:35.9057867Z env: 2023-03-31T05:26:35.9058096Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:35.9058378Z ##[endgroup] 2023-03-31T05:26:35.9182782Z Docker daemon is running... 2023-03-31T05:26:35.9238686Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2023-03-31T05:26:35.9238966Z with: 2023-03-31T05:26:35.9239164Z shell: bash 2023-03-31T05:26:35.9239363Z timeout_minutes: 5 2023-03-31T05:26:35.9239578Z max_attempts: 3 2023-03-31T05:26:35.9239794Z retry_wait_seconds: 30 2023-03-31T05:26:35.9240377Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login*** "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" 2023-03-31T05:26:35.9241002Z polling_interval_seconds: 1 2023-03-31T05:26:35.9241225Z warning_on_retry: true 2023-03-31T05:26:35.9241455Z continue_on_error: false 2023-03-31T05:26:35.9241665Z env: 2023-03-31T05:26:35.9241861Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:35.9242097Z AWS_RETRY_MODE: standard 2023-03-31T05:26:35.9242327Z AWS_MAX_ATTEMPTS: 5 2023-03-31T05:26:35.9242551Z AWS_DEFAULT_REGION: us-east-1 2023-03-31T05:26:35.9242880Z ##[endgroup] 2023-03-31T05:26:37.2760755Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2023-03-31T05:26:37.2761301Z Configure a credential helper to remove this warning. See 2023-03-31T05:26:37.2762010Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2023-03-31T05:26:37.2762429Z 2023-03-31T05:26:37.2762580Z Login Succeeded 2023-03-31T05:26:37.9762777Z Command completed after 1 attempt(s). 2023-03-31T05:26:37.9808228Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2023-03-31T05:26:37.9808551Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2023-03-31T05:26:37.9808837Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2023-03-31T05:26:37.9819945Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:26:37.9820163Z env: 2023-03-31T05:26:37.9820348Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:37.9820549Z ##[endgroup] 2023-03-31T05:26:37.9898452Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2023-03-31T05:26:37.9898728Z with: 2023-03-31T05:26:37.9899085Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:26:37.9899414Z env: 2023-03-31T05:26:37.9899594Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:37.9899786Z ##[endgroup] 2023-03-31T05:26:37.9913713Z ##[group]Run retry () { "$@" || (sleep 1 && "$@") || (sleep 2 && "$@") } 2023-03-31T05:26:37.9914019Z retry () { "$@" || (sleep 1 && "$@") || (sleep 2 && "$@") } 2023-03-31T05:26:37.9914300Z # ignore output since only exit code is used for conditional 2023-03-31T05:26:37.9914597Z # only pull docker image if it's not available locally 2023-03-31T05:26:37.9914894Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2023-03-31T05:26:37.9915236Z  retry docker pull "${DOCKER_IMAGE}" 2023-03-31T05:26:37.9915430Z fi 2023-03-31T05:26:37.9926833Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:26:37.9927060Z env: 2023-03-31T05:26:37.9927233Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:26:37.9927607Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:26:37.9927958Z ##[endgroup] 2023-03-31T05:26:38.2119768Z 51c670a3d430f7c7d9886677c30a0f7ce33e23ab: Pulling from pytorch/pytorch-linux-bionic-py3.8-clang9 2023-03-31T05:26:38.2128364Z cd150c608fee: Pulling fs layer 2023-03-31T05:26:38.2129190Z 07101faecbc3: Pulling fs layer 2023-03-31T05:26:38.2129545Z ed15b1bfe4d6: Pulling fs layer 2023-03-31T05:26:38.2129857Z eb360859a01d: Pulling fs layer 2023-03-31T05:26:38.2130191Z 0dd16982f8bc: Pulling fs layer 2023-03-31T05:26:38.2130509Z e434d3096857: Pulling fs layer 2023-03-31T05:26:38.2130755Z 3c7f2a1402ee: Pulling fs layer 2023-03-31T05:26:38.2133710Z 8ba13a22c056: Pulling fs layer 2023-03-31T05:26:38.2134067Z 65f277b8a623: Pulling fs layer 2023-03-31T05:26:38.2134414Z 24a1fa11e4c0: Pulling fs layer 2023-03-31T05:26:38.2134699Z 93ff0c4bc8ca: Pulling fs layer 2023-03-31T05:26:38.2134946Z be345439decb: Pulling fs layer 2023-03-31T05:26:38.2135327Z 2d5c311b996a: Pulling fs layer 2023-03-31T05:26:38.2135691Z b23af0f8ac11: Pulling fs layer 2023-03-31T05:26:38.2136012Z 2a855320a6b0: Pulling fs layer 2023-03-31T05:26:38.2136306Z 3c7f2a1402ee: Waiting 2023-03-31T05:26:38.2136960Z 6510beaf2701: Pulling fs layer 2023-03-31T05:26:38.2137201Z 8ba13a22c056: Waiting 2023-03-31T05:26:38.2137389Z 1b1438056498: Pulling fs layer 2023-03-31T05:26:38.2137643Z 35454e292f1f: Pulling fs layer 2023-03-31T05:26:38.2137897Z 65f277b8a623: Waiting 2023-03-31T05:26:38.2138136Z 63e19fb2d689: Pulling fs layer 2023-03-31T05:26:38.2138336Z 63c85171362d: Pulling fs layer 2023-03-31T05:26:38.2138584Z be345439decb: Waiting 2023-03-31T05:26:38.2138867Z 24a1fa11e4c0: Waiting 2023-03-31T05:26:38.2139155Z 2d5c311b996a: Waiting 2023-03-31T05:26:38.2139418Z eb360859a01d: Waiting 2023-03-31T05:26:38.2139762Z 11f764c4954b: Pulling fs layer 2023-03-31T05:26:38.2140105Z 93ff0c4bc8ca: Waiting 2023-03-31T05:26:38.2140424Z 0c7db931bc6b: Pulling fs layer 2023-03-31T05:26:38.2140766Z 0dd16982f8bc: Waiting 2023-03-31T05:26:38.2141093Z 8c1b2c386b49: Pulling fs layer 2023-03-31T05:26:38.2141402Z 1b1438056498: Waiting 2023-03-31T05:26:38.2141711Z 2a855320a6b0: Waiting 2023-03-31T05:26:38.2141990Z 180f4a03fb75: Pulling fs layer 2023-03-31T05:26:38.2142393Z 6510beaf2701: Waiting 2023-03-31T05:26:38.2142631Z 35454e292f1f: Waiting 2023-03-31T05:26:38.2142887Z 678572cc2f67: Pulling fs layer 2023-03-31T05:26:38.2143190Z 1aff90094438: Pulling fs layer 2023-03-31T05:26:38.2143495Z fa585aa4a8e1: Pulling fs layer 2023-03-31T05:26:38.2143789Z 63e19fb2d689: Waiting 2023-03-31T05:26:38.2144073Z 11f764c4954b: Waiting 2023-03-31T05:26:38.2144361Z 3c56e8c2b04d: Pulling fs layer 2023-03-31T05:26:38.2144896Z 63c85171362d: Waiting 2023-03-31T05:26:38.2145207Z 50b626ecc083: Pulling fs layer 2023-03-31T05:26:38.2145488Z 0c7db931bc6b: Waiting 2023-03-31T05:26:38.2145805Z 248092c7f5fe: Pulling fs layer 2023-03-31T05:26:38.2146087Z 678572cc2f67: Waiting 2023-03-31T05:26:38.2146494Z 6f73309f9d7e: Pulling fs layer 2023-03-31T05:26:38.2146780Z 8c1b2c386b49: Waiting 2023-03-31T05:26:38.2147053Z 1aff90094438: Waiting 2023-03-31T05:26:38.2147338Z 68a4afeeb912: Pulling fs layer 2023-03-31T05:26:38.2147636Z 180f4a03fb75: Waiting 2023-03-31T05:26:38.2147984Z 9ea487b01ac3: Pulling fs layer 2023-03-31T05:26:38.2148374Z fa585aa4a8e1: Waiting 2023-03-31T05:26:38.2148669Z fadc1f6932dd: Pulling fs layer 2023-03-31T05:26:38.2148964Z 3c56e8c2b04d: Waiting 2023-03-31T05:26:38.2149274Z 858bbadb10e2: Pulling fs layer 2023-03-31T05:26:38.2149571Z 248092c7f5fe: Waiting 2023-03-31T05:26:38.2149862Z 50b626ecc083: Waiting 2023-03-31T05:26:38.2150175Z 126a2feff570: Pulling fs layer 2023-03-31T05:26:38.2150500Z 29363c6e3c8e: Pulling fs layer 2023-03-31T05:26:38.2150811Z 6f73309f9d7e: Waiting 2023-03-31T05:26:38.2151104Z b56488832663: Pulling fs layer 2023-03-31T05:26:38.2151387Z fadc1f6932dd: Waiting 2023-03-31T05:26:38.2151694Z 604db63ab385: Pulling fs layer 2023-03-31T05:26:38.2151999Z 68a4afeeb912: Waiting 2023-03-31T05:26:38.2152278Z 9ea487b01ac3: Waiting 2023-03-31T05:26:38.2152607Z c364fb9f7729: Pulling fs layer 2023-03-31T05:26:38.2152931Z 858bbadb10e2: Waiting 2023-03-31T05:26:38.2153235Z d24b370e64b9: Pulling fs layer 2023-03-31T05:26:38.2153581Z 0e2c6138be96: Pulling fs layer 2023-03-31T05:26:38.2153912Z 29363c6e3c8e: Waiting 2023-03-31T05:26:38.2154198Z 126a2feff570: Waiting 2023-03-31T05:26:38.2154526Z 703a795d5a8a: Pulling fs layer 2023-03-31T05:26:38.2154853Z 28435edf4c93: Pulling fs layer 2023-03-31T05:26:38.2155167Z 09ce6e8b4cf8: Pulling fs layer 2023-03-31T05:26:38.2155480Z a96ae6c66533: Pulling fs layer 2023-03-31T05:26:38.2163938Z b56488832663: Waiting 2023-03-31T05:26:38.2164309Z 604db63ab385: Waiting 2023-03-31T05:26:38.2164612Z fd8f62cf4228: Pulling fs layer 2023-03-31T05:26:38.2164927Z 54fb6e3adb85: Pulling fs layer 2023-03-31T05:26:38.2165175Z d24b370e64b9: Waiting 2023-03-31T05:26:38.2165458Z 28435edf4c93: Waiting 2023-03-31T05:26:38.2165699Z 703a795d5a8a: Waiting 2023-03-31T05:26:38.2166003Z a0bdf07b84f7: Pulling fs layer 2023-03-31T05:26:38.2166348Z 0e2c6138be96: Waiting 2023-03-31T05:26:38.2166748Z b20c7c19a483: Pulling fs layer 2023-03-31T05:26:38.2166937Z 09ce6e8b4cf8: Waiting 2023-03-31T05:26:38.2167130Z 8f3f97fea448: Pulling fs layer 2023-03-31T05:26:38.2167547Z 49c34afc7493: Pulling fs layer 2023-03-31T05:26:38.2167727Z a96ae6c66533: Waiting 2023-03-31T05:26:38.2167961Z 71a4d7e24c83: Pulling fs layer 2023-03-31T05:26:38.2168153Z fd8f62cf4228: Waiting 2023-03-31T05:26:38.2168320Z b20c7c19a483: Waiting 2023-03-31T05:26:38.2168557Z 54fb6e3adb85: Waiting 2023-03-31T05:26:38.2168853Z 187f67d75df7: Pulling fs layer 2023-03-31T05:26:38.2169147Z 49c34afc7493: Waiting 2023-03-31T05:26:38.2169494Z 8f3f97fea448: Waiting 2023-03-31T05:26:38.2169849Z a04718b4da41: Pulling fs layer 2023-03-31T05:26:38.2170106Z 38151b60945d: Pulling fs layer 2023-03-31T05:26:38.2170295Z fcd08880d1db: Pulling fs layer 2023-03-31T05:26:38.2170494Z ac039681aba5: Pulling fs layer 2023-03-31T05:26:38.2170685Z a04718b4da41: Waiting 2023-03-31T05:26:38.2170852Z 38151b60945d: Waiting 2023-03-31T05:26:38.2171041Z 38e09001e7e1: Pulling fs layer 2023-03-31T05:26:38.2171242Z 43c424cda7b2: Pulling fs layer 2023-03-31T05:26:38.2171433Z a519848d8bae: Pulling fs layer 2023-03-31T05:26:38.2171640Z 62ac17d0a94c: Pulling fs layer 2023-03-31T05:26:38.2171849Z ea2d55a0ef58: Pulling fs layer 2023-03-31T05:26:38.2172039Z bc725dbb7a96: Pulling fs layer 2023-03-31T05:26:38.2172243Z c8e3b5a367fa: Pulling fs layer 2023-03-31T05:26:38.2172448Z 18d18bc54b95: Pulling fs layer 2023-03-31T05:26:38.2172624Z 187f67d75df7: Waiting 2023-03-31T05:26:38.2172802Z 38e09001e7e1: Waiting 2023-03-31T05:26:38.2172979Z c8e3b5a367fa: Waiting 2023-03-31T05:26:38.2173143Z 18d18bc54b95: Waiting 2023-03-31T05:26:38.2173448Z bc725dbb7a96: Waiting 2023-03-31T05:26:38.2173683Z 43c424cda7b2: Waiting 2023-03-31T05:26:38.2173847Z ac039681aba5: Waiting 2023-03-31T05:26:38.2174025Z 62ac17d0a94c: Waiting 2023-03-31T05:26:38.2999318Z 07101faecbc3: Download complete 2023-03-31T05:26:38.3797072Z eb360859a01d: Verifying Checksum 2023-03-31T05:26:38.3797505Z eb360859a01d: Download complete 2023-03-31T05:26:38.5325955Z cd150c608fee: Verifying Checksum 2023-03-31T05:26:38.5326563Z cd150c608fee: Download complete 2023-03-31T05:26:38.6120701Z e434d3096857: Verifying Checksum 2023-03-31T05:26:38.6122367Z e434d3096857: Download complete 2023-03-31T05:26:38.6821735Z 3c7f2a1402ee: Verifying Checksum 2023-03-31T05:26:38.6822175Z 3c7f2a1402ee: Download complete 2023-03-31T05:26:38.7679989Z 8ba13a22c056: Verifying Checksum 2023-03-31T05:26:38.7680572Z 8ba13a22c056: Download complete 2023-03-31T05:26:38.8415934Z 65f277b8a623: Verifying Checksum 2023-03-31T05:26:38.8416286Z 65f277b8a623: Download complete 2023-03-31T05:26:38.9357748Z 24a1fa11e4c0: Verifying Checksum 2023-03-31T05:26:39.0145550Z 24a1fa11e4c0: Download complete 2023-03-31T05:26:39.0146029Z 93ff0c4bc8ca: Verifying Checksum 2023-03-31T05:26:39.0146414Z 93ff0c4bc8ca: Download complete 2023-03-31T05:26:39.0927681Z be345439decb: Verifying Checksum 2023-03-31T05:26:39.0928136Z be345439decb: Download complete 2023-03-31T05:26:39.1614983Z 2d5c311b996a: Verifying Checksum 2023-03-31T05:26:39.1615466Z 2d5c311b996a: Download complete 2023-03-31T05:26:39.2266340Z b23af0f8ac11: Verifying Checksum 2023-03-31T05:26:39.2266842Z b23af0f8ac11: Download complete 2023-03-31T05:26:39.2538115Z cd150c608fee: Pull complete 2023-03-31T05:26:39.3728718Z 07101faecbc3: Pull complete 2023-03-31T05:26:39.4463796Z 0dd16982f8bc: Verifying Checksum 2023-03-31T05:26:39.4464163Z 0dd16982f8bc: Download complete 2023-03-31T05:26:39.5224461Z 6510beaf2701: Download complete 2023-03-31T05:26:39.5994147Z 1b1438056498: Download complete 2023-03-31T05:26:39.6735045Z 35454e292f1f: Download complete 2023-03-31T05:26:39.7814650Z 63e19fb2d689: Download complete 2023-03-31T05:26:39.8529044Z 63c85171362d: Verifying Checksum 2023-03-31T05:26:39.8529472Z 63c85171362d: Download complete 2023-03-31T05:26:39.9360034Z 11f764c4954b: Verifying Checksum 2023-03-31T05:26:39.9360645Z 11f764c4954b: Download complete 2023-03-31T05:26:40.0187611Z 0c7db931bc6b: Verifying Checksum 2023-03-31T05:26:40.0188016Z 0c7db931bc6b: Download complete 2023-03-31T05:26:40.0948223Z 8c1b2c386b49: Verifying Checksum 2023-03-31T05:26:40.0948695Z 8c1b2c386b49: Download complete 2023-03-31T05:26:40.1743035Z 180f4a03fb75: Verifying Checksum 2023-03-31T05:26:40.1743324Z 180f4a03fb75: Download complete 2023-03-31T05:26:40.9991055Z ed15b1bfe4d6: Verifying Checksum 2023-03-31T05:26:40.9991455Z ed15b1bfe4d6: Download complete 2023-03-31T05:26:41.0700959Z 1aff90094438: Verifying Checksum 2023-03-31T05:26:41.0701377Z 1aff90094438: Download complete 2023-03-31T05:26:41.1204407Z 678572cc2f67: Verifying Checksum 2023-03-31T05:26:41.1205003Z 678572cc2f67: Download complete 2023-03-31T05:26:41.1592817Z fa585aa4a8e1: Verifying Checksum 2023-03-31T05:26:41.2061870Z fa585aa4a8e1: Download complete 2023-03-31T05:26:41.2062315Z 3c56e8c2b04d: Verifying Checksum 2023-03-31T05:26:41.2062657Z 3c56e8c2b04d: Download complete 2023-03-31T05:26:41.2585598Z 50b626ecc083: Verifying Checksum 2023-03-31T05:26:41.2716313Z 248092c7f5fe: Verifying Checksum 2023-03-31T05:26:41.2716687Z 248092c7f5fe: Download complete 2023-03-31T05:26:41.3402575Z 68a4afeeb912: Download complete 2023-03-31T05:26:41.4124282Z 9ea487b01ac3: Verifying Checksum 2023-03-31T05:26:41.4124744Z 9ea487b01ac3: Download complete 2023-03-31T05:26:41.5000317Z fadc1f6932dd: Verifying Checksum 2023-03-31T05:26:41.5000616Z fadc1f6932dd: Download complete 2023-03-31T05:26:41.5746356Z 858bbadb10e2: Verifying Checksum 2023-03-31T05:26:41.5746755Z 858bbadb10e2: Download complete 2023-03-31T05:26:41.6508584Z 126a2feff570: Download complete 2023-03-31T05:26:41.7248288Z 29363c6e3c8e: Verifying Checksum 2023-03-31T05:26:41.7249613Z 29363c6e3c8e: Download complete 2023-03-31T05:26:41.8161375Z b56488832663: Download complete 2023-03-31T05:26:41.8854899Z 604db63ab385: Verifying Checksum 2023-03-31T05:26:41.8855328Z 604db63ab385: Download complete 2023-03-31T05:26:43.2495361Z 6f73309f9d7e: Verifying Checksum 2023-03-31T05:26:43.2496803Z 6f73309f9d7e: Download complete 2023-03-31T05:26:43.3369247Z d24b370e64b9: Verifying Checksum 2023-03-31T05:26:43.3369612Z d24b370e64b9: Download complete 2023-03-31T05:26:43.4265100Z 0e2c6138be96: Verifying Checksum 2023-03-31T05:26:43.4265537Z 0e2c6138be96: Download complete 2023-03-31T05:26:44.0601342Z c364fb9f7729: Verifying Checksum 2023-03-31T05:26:44.0601900Z c364fb9f7729: Download complete 2023-03-31T05:26:44.1272614Z 28435edf4c93: Verifying Checksum 2023-03-31T05:26:44.1965880Z 09ce6e8b4cf8: Verifying Checksum 2023-03-31T05:26:44.1966163Z 09ce6e8b4cf8: Download complete 2023-03-31T05:26:44.6974874Z a96ae6c66533: Verifying Checksum 2023-03-31T05:26:44.6978090Z a96ae6c66533: Download complete 2023-03-31T05:26:44.7709974Z fd8f62cf4228: Verifying Checksum 2023-03-31T05:26:44.7710460Z fd8f62cf4228: Download complete 2023-03-31T05:26:44.8424897Z 54fb6e3adb85: Verifying Checksum 2023-03-31T05:26:44.8425336Z 54fb6e3adb85: Download complete 2023-03-31T05:26:44.9086716Z a0bdf07b84f7: Verifying Checksum 2023-03-31T05:26:44.9087027Z a0bdf07b84f7: Download complete 2023-03-31T05:26:45.0088810Z b20c7c19a483: Download complete 2023-03-31T05:26:45.2587344Z 8f3f97fea448: Verifying Checksum 2023-03-31T05:26:45.2587641Z 8f3f97fea448: Download complete 2023-03-31T05:26:45.3404291Z 49c34afc7493: Download complete 2023-03-31T05:26:45.4075999Z 71a4d7e24c83: Verifying Checksum 2023-03-31T05:26:45.4076459Z 71a4d7e24c83: Download complete 2023-03-31T05:26:45.4783050Z 187f67d75df7: Verifying Checksum 2023-03-31T05:26:45.4783510Z 187f67d75df7: Download complete 2023-03-31T05:26:47.6778392Z ed15b1bfe4d6: Pull complete 2023-03-31T05:26:47.9083654Z eb360859a01d: Pull complete 2023-03-31T05:27:07.8746890Z a04718b4da41: Verifying Checksum 2023-03-31T05:27:07.8747362Z a04718b4da41: Download complete 2023-03-31T05:27:07.9503088Z 38151b60945d: Verifying Checksum 2023-03-31T05:27:07.9513546Z 38151b60945d: Download complete 2023-03-31T05:27:08.0465111Z fcd08880d1db: Verifying Checksum 2023-03-31T05:27:08.0465410Z fcd08880d1db: Download complete 2023-03-31T05:27:08.1211156Z ac039681aba5: Verifying Checksum 2023-03-31T05:27:08.1211555Z ac039681aba5: Download complete 2023-03-31T05:27:08.1941049Z 38e09001e7e1: Download complete 2023-03-31T05:27:08.4133372Z 43c424cda7b2: Verifying Checksum 2023-03-31T05:27:08.4133840Z 43c424cda7b2: Download complete 2023-03-31T05:27:08.4780776Z a519848d8bae: Verifying Checksum 2023-03-31T05:27:08.4781304Z a519848d8bae: Download complete 2023-03-31T05:27:08.5576233Z 62ac17d0a94c: Download complete 2023-03-31T05:27:08.6396064Z ea2d55a0ef58: Verifying Checksum 2023-03-31T05:27:08.6396423Z ea2d55a0ef58: Download complete 2023-03-31T05:27:08.7216627Z bc725dbb7a96: Download complete 2023-03-31T05:27:08.7972277Z c8e3b5a367fa: Verifying Checksum 2023-03-31T05:27:08.7972731Z c8e3b5a367fa: Download complete 2023-03-31T05:27:10.3072265Z 18d18bc54b95: Verifying Checksum 2023-03-31T05:27:10.3072637Z 18d18bc54b95: Download complete 2023-03-31T05:27:11.3845452Z 0dd16982f8bc: Pull complete 2023-03-31T05:27:11.9244309Z e434d3096857: Pull complete 2023-03-31T05:27:12.4346959Z 3c7f2a1402ee: Pull complete 2023-03-31T05:27:12.9667215Z 8ba13a22c056: Pull complete 2023-03-31T05:27:13.5655320Z 65f277b8a623: Pull complete 2023-03-31T05:27:13.6135251Z 2a855320a6b0: Verifying Checksum 2023-03-31T05:27:13.6135603Z 2a855320a6b0: Download complete 2023-03-31T05:27:14.0152681Z 24a1fa11e4c0: Pull complete 2023-03-31T05:27:14.5201045Z 93ff0c4bc8ca: Pull complete 2023-03-31T05:27:15.0056987Z be345439decb: Pull complete 2023-03-31T05:27:15.4598728Z 2d5c311b996a: Pull complete 2023-03-31T05:27:15.9085790Z b23af0f8ac11: Pull complete 2023-03-31T05:27:17.9321732Z 703a795d5a8a: Verifying Checksum 2023-03-31T05:27:17.9321990Z 703a795d5a8a: Download complete 2023-03-31T05:27:50.7791781Z 2a855320a6b0: Pull complete 2023-03-31T05:27:51.2242340Z 6510beaf2701: Pull complete 2023-03-31T05:27:51.3925191Z 1b1438056498: Pull complete 2023-03-31T05:27:51.5697957Z 35454e292f1f: Pull complete 2023-03-31T05:27:51.8172479Z 63e19fb2d689: Pull complete 2023-03-31T05:27:51.9728367Z 63c85171362d: Pull complete 2023-03-31T05:27:52.1203866Z 11f764c4954b: Pull complete 2023-03-31T05:27:52.3020690Z 0c7db931bc6b: Pull complete 2023-03-31T05:27:52.4830690Z 8c1b2c386b49: Pull complete 2023-03-31T05:27:52.6785237Z 180f4a03fb75: Pull complete 2023-03-31T05:27:55.2000921Z 678572cc2f67: Pull complete 2023-03-31T05:27:55.4264734Z 1aff90094438: Pull complete 2023-03-31T05:27:55.6630322Z fa585aa4a8e1: Pull complete 2023-03-31T05:27:55.9416445Z 3c56e8c2b04d: Pull complete 2023-03-31T05:27:56.1859509Z 50b626ecc083: Pull complete 2023-03-31T05:27:56.4163343Z 248092c7f5fe: Pull complete 2023-03-31T05:28:00.8342804Z 6f73309f9d7e: Pull complete 2023-03-31T05:28:01.0565867Z 68a4afeeb912: Pull complete 2023-03-31T05:28:01.2709060Z 9ea487b01ac3: Pull complete 2023-03-31T05:28:01.5113560Z fadc1f6932dd: Pull complete 2023-03-31T05:28:01.7509377Z 858bbadb10e2: Pull complete 2023-03-31T05:28:01.9932317Z 126a2feff570: Pull complete 2023-03-31T05:28:02.2156489Z 29363c6e3c8e: Pull complete 2023-03-31T05:28:02.4632448Z b56488832663: Pull complete 2023-03-31T05:28:02.6308497Z 604db63ab385: Pull complete 2023-03-31T05:28:09.7515531Z c364fb9f7729: Pull complete 2023-03-31T05:28:09.9638511Z d24b370e64b9: Pull complete 2023-03-31T05:28:10.2000983Z 0e2c6138be96: Pull complete 2023-03-31T05:28:26.4322700Z 703a795d5a8a: Pull complete 2023-03-31T05:28:26.6357484Z 28435edf4c93: Pull complete 2023-03-31T05:28:26.8116120Z 09ce6e8b4cf8: Pull complete 2023-03-31T05:28:29.0087136Z a96ae6c66533: Pull complete 2023-03-31T05:28:29.2251328Z fd8f62cf4228: Pull complete 2023-03-31T05:28:29.4473071Z 54fb6e3adb85: Pull complete 2023-03-31T05:28:29.6236455Z a0bdf07b84f7: Pull complete 2023-03-31T05:28:29.7677364Z b20c7c19a483: Pull complete 2023-03-31T05:28:30.5940815Z 8f3f97fea448: Pull complete 2023-03-31T05:28:30.7882074Z 49c34afc7493: Pull complete 2023-03-31T05:28:31.0127370Z 71a4d7e24c83: Pull complete 2023-03-31T05:28:31.4114442Z 187f67d75df7: Pull complete 2023-03-31T05:28:51.7457868Z a04718b4da41: Pull complete 2023-03-31T05:28:52.1933840Z 38151b60945d: Pull complete 2023-03-31T05:28:52.5058731Z fcd08880d1db: Pull complete 2023-03-31T05:28:52.7359381Z ac039681aba5: Pull complete 2023-03-31T05:28:52.9631660Z 38e09001e7e1: Pull complete 2023-03-31T05:28:53.3232498Z 43c424cda7b2: Pull complete 2023-03-31T05:28:53.5595333Z a519848d8bae: Pull complete 2023-03-31T05:28:53.8017791Z 62ac17d0a94c: Pull complete 2023-03-31T05:28:53.9893133Z ea2d55a0ef58: Pull complete 2023-03-31T05:28:54.2066102Z bc725dbb7a96: Pull complete 2023-03-31T05:28:54.4298232Z c8e3b5a367fa: Pull complete 2023-03-31T05:28:56.1678872Z 18d18bc54b95: Pull complete 2023-03-31T05:28:56.2741164Z Digest: sha256:33691e310ed8c7b7d7d19369f1fbba6440bed4595f5bf54deb94711e24e235ce 2023-03-31T05:28:56.3176469Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:28:56.3380751Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:28:56.3461451Z ##[group]Run python3 -m pip install psutil==5.9.1 nvidia-ml-py==11.525.84 2023-03-31T05:28:56.3461802Z python3 -m pip install psutil==5.9.1 nvidia-ml-py==11.525.84 2023-03-31T05:28:56.3462116Z python3 -m tools.stats.monitor > usage_log.txt 2>&1 & 2023-03-31T05:28:56.3462407Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2023-03-31T05:28:56.3474220Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:28:56.3474453Z env: 2023-03-31T05:28:56.3474637Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:28:56.3474821Z ##[endgroup] 2023-03-31T05:28:59.0291911Z Defaulting to user installation because normal site-packages is not writeable 2023-03-31T05:28:59.3667679Z Collecting psutil==5.9.1 2023-03-31T05:28:59.3812418Z Downloading psutil-5.9.1-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (281 kB) 2023-03-31T05:28:59.4695540Z Collecting nvidia-ml-py==11.525.84 2023-03-31T05:28:59.4728042Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2023-03-31T05:28:59.5788392Z Installing collected packages: psutil, nvidia-ml-py 2023-03-31T05:28:59.8220699Z Successfully installed nvidia-ml-py-11.525.84 psutil-5.9.1 2023-03-31T05:28:59.8893053Z Prepare all required actions 2023-03-31T05:28:59.8893312Z Getting action download info 2023-03-31T05:29:00.0250055Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:4a8bfae15cc25cc0785c1603ee87a9da8fd442ea) 2023-03-31T05:29:00.2094517Z Download action repository 'actions/download-artifact@v3' (SHA:9bc31d5ccc31df68ecc42ccf4149144866c47d8a) 2023-03-31T05:29:00.3413537Z ##[group]Run ./.github/actions/download-build-artifacts 2023-03-31T05:29:00.3413775Z with: 2023-03-31T05:29:00.3413980Z name: linux-bionic-py3.8-clang9 2023-03-31T05:29:00.3414167Z env: 2023-03-31T05:29:00.3414347Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:00.3414546Z ##[endgroup] 2023-03-31T05:29:00.3441250Z ##[group]Run seemethere/download-artifact-s3@v4 2023-03-31T05:29:00.3441483Z with: 2023-03-31T05:29:00.3441669Z name: linux-bionic-py3.8-clang9 2023-03-31T05:29:00.3441891Z s3-bucket: gha-artifacts 2023-03-31T05:29:00.3442089Z region: us-east-1 2023-03-31T05:29:00.3442259Z env: 2023-03-31T05:29:00.3442488Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:00.3442669Z ##[endgroup] 2023-03-31T05:29:00.7803655Z Found 1 objects with prefix pytorch/pytorch/4571766038/linux-bionic-py3.8-clang9/ 2023-03-31T05:29:00.7804248Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2023-03-31T05:29:04.3053647Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2023-03-31T05:29:04.3054051Z 2023-03-31T05:29:04.3073220Z ##[warning]The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/ 2023-03-31T05:29:04.3083871Z Artifact download has finished successfully 2023-03-31T05:29:04.3196082Z ##[group]Run unzip -o artifacts.zip 2023-03-31T05:29:04.3196321Z unzip -o artifacts.zip 2023-03-31T05:29:04.3208101Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:29:04.3208324Z env: 2023-03-31T05:29:04.3208506Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:04.3208706Z ##[endgroup] 2023-03-31T05:29:04.3558912Z Archive: artifacts.zip 2023-03-31T05:29:04.3559619Z creating: dist/ 2023-03-31T05:29:05.2251510Z inflating: dist/torch-2.1.0a0+git4af0d9f-cp38-cp38-linux_x86_64.whl 2023-03-31T05:29:05.2251902Z creating: build/custom_test_artifacts/ 2023-03-31T05:29:05.2252240Z creating: build/custom_test_artifacts/custom-op-build/ 2023-03-31T05:29:05.2252615Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2023-03-31T05:29:05.2253499Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeOutput.log 2023-03-31T05:29:05.2253920Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/ 2023-03-31T05:29:05.2254364Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CMakeSystem.cmake 2023-03-31T05:29:05.2254840Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdC/ 2023-03-31T05:29:05.2255266Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdC/tmp/ 2023-03-31T05:29:05.2257010Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdC/CMakeCCompilerId.c 2023-03-31T05:29:05.2257982Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdC/a.out 2023-03-31T05:29:05.2258420Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdCXX/ 2023-03-31T05:29:05.2258851Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdCXX/tmp/ 2023-03-31T05:29:05.2260625Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdCXX/CMakeCXXCompilerId.cpp 2023-03-31T05:29:05.2261821Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CompilerIdCXX/a.out 2023-03-31T05:29:05.2262889Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_C.bin 2023-03-31T05:29:05.2263458Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CMakeCCompiler.cmake 2023-03-31T05:29:05.2264405Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_CXX.bin 2023-03-31T05:29:05.2265203Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.22.1/CMakeCXXCompiler.cmake 2023-03-31T05:29:05.2265640Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2023-03-31T05:29:05.2266053Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2023-03-31T05:29:05.2266475Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2023-03-31T05:29:05.2266935Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2023-03-31T05:29:05.2267428Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2023-03-31T05:29:05.2267895Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2023-03-31T05:29:05.2268344Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2023-03-31T05:29:05.2268804Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2023-03-31T05:29:05.2269266Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2023-03-31T05:29:05.2269720Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2023-03-31T05:29:05.2270183Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2023-03-31T05:29:05.2270643Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2023-03-31T05:29:05.2289390Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2023-03-31T05:29:05.2352331Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2023-03-31T05:29:05.2352794Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2023-03-31T05:29:05.2353272Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2023-03-31T05:29:05.2353775Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2023-03-31T05:29:05.2354251Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2023-03-31T05:29:05.2354827Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2023-03-31T05:29:05.2355309Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2023-03-31T05:29:05.2355796Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2023-03-31T05:29:05.2356273Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2023-03-31T05:29:05.2356753Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2023-03-31T05:29:05.2357226Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2023-03-31T05:29:05.2376090Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2023-03-31T05:29:05.2416689Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2023-03-31T05:29:05.2417491Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2023-03-31T05:29:05.2418146Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2023-03-31T05:29:05.2418950Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2023-03-31T05:29:05.2419395Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2023-03-31T05:29:05.2419825Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2023-03-31T05:29:05.2420446Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2023-03-31T05:29:05.2421213Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2023-03-31T05:29:05.2421744Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2023-03-31T05:29:05.2477538Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2023-03-31T05:29:05.2510955Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2023-03-31T05:29:05.2511412Z creating: build/custom_test_artifacts/jit-hook-build/ 2023-03-31T05:29:05.2511754Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2023-03-31T05:29:05.2513622Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeOutput.log 2023-03-31T05:29:05.2514221Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/ 2023-03-31T05:29:05.2514748Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CMakeSystem.cmake 2023-03-31T05:29:05.2515185Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdC/ 2023-03-31T05:29:05.2515621Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdC/tmp/ 2023-03-31T05:29:05.2516638Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdC/CMakeCCompilerId.c 2023-03-31T05:29:05.2517769Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdC/a.out 2023-03-31T05:29:05.2518308Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdCXX/ 2023-03-31T05:29:05.2518752Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdCXX/tmp/ 2023-03-31T05:29:05.2520357Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdCXX/CMakeCXXCompilerId.cpp 2023-03-31T05:29:05.2521264Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CompilerIdCXX/a.out 2023-03-31T05:29:05.2522372Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_C.bin 2023-03-31T05:29:05.2523008Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CMakeCCompiler.cmake 2023-03-31T05:29:05.2524211Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_CXX.bin 2023-03-31T05:29:05.2525075Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.22.1/CMakeCXXCompiler.cmake 2023-03-31T05:29:05.2525708Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2023-03-31T05:29:05.2526354Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2023-03-31T05:29:05.2527012Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2023-03-31T05:29:05.2527668Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2023-03-31T05:29:05.2528293Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2023-03-31T05:29:05.2528986Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2023-03-31T05:29:05.2529634Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2023-03-31T05:29:05.2530085Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2023-03-31T05:29:05.2530551Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2023-03-31T05:29:05.2531144Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2023-03-31T05:29:05.2531617Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2023-03-31T05:29:05.2532068Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2023-03-31T05:29:05.2549016Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2023-03-31T05:29:05.2579610Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2023-03-31T05:29:05.2580392Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2023-03-31T05:29:05.2580943Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2023-03-31T05:29:05.2581452Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2023-03-31T05:29:05.2582028Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2023-03-31T05:29:05.2582452Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2023-03-31T05:29:05.2583460Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2023-03-31T05:29:05.2584245Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2023-03-31T05:29:05.2584783Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2023-03-31T05:29:05.2610206Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2023-03-31T05:29:05.2610641Z creating: build/custom_test_artifacts/custom-backend-build/ 2023-03-31T05:29:05.2611024Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2023-03-31T05:29:05.2612724Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeOutput.log 2023-03-31T05:29:05.2613366Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/ 2023-03-31T05:29:05.2614028Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CMakeSystem.cmake 2023-03-31T05:29:05.2614491Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdC/ 2023-03-31T05:29:05.2614945Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdC/tmp/ 2023-03-31T05:29:05.2615793Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdC/CMakeCCompilerId.c 2023-03-31T05:29:05.2616886Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdC/a.out 2023-03-31T05:29:05.2617433Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdCXX/ 2023-03-31T05:29:05.2617891Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdCXX/tmp/ 2023-03-31T05:29:05.2619449Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdCXX/CMakeCXXCompilerId.cpp 2023-03-31T05:29:05.2620454Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CompilerIdCXX/a.out 2023-03-31T05:29:05.2621452Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_C.bin 2023-03-31T05:29:05.2622175Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CMakeCCompiler.cmake 2023-03-31T05:29:05.2623300Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CMakeDetermineCompilerABI_CXX.bin 2023-03-31T05:29:05.2624251Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.22.1/CMakeCXXCompiler.cmake 2023-03-31T05:29:05.2625012Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2023-03-31T05:29:05.2625785Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2023-03-31T05:29:05.2626580Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2023-03-31T05:29:05.2627459Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2023-03-31T05:29:05.2628319Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2023-03-31T05:29:05.2628967Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2023-03-31T05:29:05.2629515Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2023-03-31T05:29:05.2630078Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2023-03-31T05:29:05.2630620Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2023-03-31T05:29:05.2631180Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2023-03-31T05:29:05.2631764Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2023-03-31T05:29:05.2632327Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2023-03-31T05:29:05.2633119Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2023-03-31T05:29:05.2717932Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2023-03-31T05:29:05.2718742Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2023-03-31T05:29:05.2719584Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2023-03-31T05:29:05.2720242Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2023-03-31T05:29:05.2721166Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2023-03-31T05:29:05.2722122Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2023-03-31T05:29:05.2722699Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2023-03-31T05:29:05.2723414Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2023-03-31T05:29:05.2724024Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2023-03-31T05:29:05.2724602Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2023-03-31T05:29:05.2725168Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2023-03-31T05:29:05.2741366Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2023-03-31T05:29:05.2763149Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2023-03-31T05:29:05.2805370Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2023-03-31T05:29:05.2805892Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2023-03-31T05:29:05.2806354Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2023-03-31T05:29:05.2806789Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2023-03-31T05:29:05.2807218Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2023-03-31T05:29:05.2807646Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2023-03-31T05:29:05.2808046Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2023-03-31T05:29:05.2808645Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2023-03-31T05:29:05.2842831Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2023-03-31T05:29:05.2861024Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2023-03-31T05:29:05.2861288Z creating: build/lib/ 2023-03-31T05:29:05.2861916Z inflating: build/lib/libclog.a 2023-03-31T05:29:05.2870254Z inflating: build/lib/libpthreadpool.a 2023-03-31T05:29:05.2929786Z inflating: build/lib/libgtest.a 2023-03-31T05:29:05.2937505Z inflating: build/lib/libittnotify.a 2023-03-31T05:29:05.3002742Z inflating: build/lib/libbenchmark.a 2023-03-31T05:29:05.3079934Z inflating: build/lib/libprotobuf-lite.a 2023-03-31T05:29:05.3105532Z inflating: build/lib/libtensorpipe_uv.a 2023-03-31T05:29:05.3507141Z inflating: build/lib/libprotobuf.a 2023-03-31T05:29:05.3571495Z inflating: build/lib/libasmjit.a 2023-03-31T05:29:05.3641630Z inflating: build/lib/libgloo.a 2023-03-31T05:29:05.3659427Z inflating: build/lib/libfmt.a 2023-03-31T05:29:05.3659977Z inflating: build/lib/libfoxi_loader.a 2023-03-31T05:29:05.3660982Z inflating: build/lib/libtorch_global_deps.so 2023-03-31T05:29:05.3721039Z inflating: build/lib/libc10.so 2023-03-31T05:29:05.3728450Z inflating: build/lib/libcpuinfo.a 2023-03-31T05:29:05.3735296Z inflating: build/lib/libcpuinfo_internals.a 2023-03-31T05:29:05.3747874Z inflating: build/lib/libqnnpack.a 2023-03-31T05:29:05.3750041Z inflating: build/lib/libnnpack_reference_layers.a 2023-03-31T05:29:05.3768202Z inflating: build/lib/libpytorch_qnnpack.a 2023-03-31T05:29:05.3780152Z inflating: build/lib/libgmock.a 2023-03-31T05:29:05.3780923Z inflating: build/lib/libgtest_main.a 2023-03-31T05:29:05.3781285Z inflating: build/lib/libbenchmark_main.a 2023-03-31T05:29:05.4230765Z inflating: build/lib/libprotoc.a 2023-03-31T05:29:06.5738981Z inflating: build/lib/libdnnl.a 2023-03-31T05:29:06.5756077Z inflating: build/lib/libnnpack.a 2023-03-31T05:29:06.6249473Z inflating: build/lib/libtensorpipe.a 2023-03-31T05:29:06.7273419Z inflating: build/lib/libfbgemm.a 2023-03-31T05:29:06.7273836Z inflating: build/lib/libgmock_main.a 2023-03-31T05:29:06.8280989Z inflating: build/lib/libdnnl_graph.a 2023-03-31T05:29:06.8521576Z inflating: build/lib/libkineto.a 2023-03-31T05:29:06.8560480Z inflating: build/lib/libcaffe2_protos.a 2023-03-31T05:29:06.8602927Z inflating: build/lib/libonnx_proto.a 2023-03-31T05:29:06.8720555Z inflating: build/lib/libXNNPACK.a 2023-03-31T05:29:06.9340291Z inflating: build/lib/libonnx.a 2023-03-31T05:29:09.1147164Z inflating: build/lib/libtorch_cpu.so 2023-03-31T05:29:09.1147643Z inflating: build/lib/libtorch.so 2023-03-31T05:29:09.1160770Z inflating: build/lib/libunbox_lib.a 2023-03-31T05:29:09.1165288Z inflating: build/lib/libshm.so 2023-03-31T05:29:09.1187722Z inflating: build/lib/libjitbackend_test.so 2023-03-31T05:29:09.1244828Z inflating: build/lib/libtorchbind_test.so 2023-03-31T05:29:09.1271344Z inflating: build/lib/libbackend_with_compiler.so 2023-03-31T05:29:09.3080114Z inflating: build/lib/libtorch_python.so 2023-03-31T05:29:09.3118126Z inflating: build/lib/libnnapi_backend.so 2023-03-31T05:29:09.3118528Z creating: build/bin/ 2023-03-31T05:29:09.3118901Z creating: build/bin/CMakeFiles/ 2023-03-31T05:29:09.3119443Z inflating: build/bin/CMakeFiles/CMakeDirectoryInformation.cmake 2023-03-31T05:29:09.3120020Z extracting: build/bin/CMakeFiles/progress.marks 2023-03-31T05:29:09.3171942Z inflating: build/bin/test_parallel 2023-03-31T05:29:09.3172600Z inflating: build/bin/Makefile 2023-03-31T05:29:09.3222781Z inflating: build/bin/StorageUtils_test 2023-03-31T05:29:09.3223312Z inflating: build/bin/cmake_install.cmake 2023-03-31T05:29:09.3274618Z inflating: build/bin/stride_properties_test 2023-03-31T05:29:09.3275373Z inflating: build/bin/CTestTestfile.cmake 2023-03-31T05:29:09.3323971Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2023-03-31T05:29:09.3373807Z inflating: build/bin/c10_DeviceGuard_test 2023-03-31T05:29:09.3422931Z inflating: build/bin/c10_Device_test 2023-03-31T05:29:09.3470903Z inflating: build/bin/c10_StreamGuard_test 2023-03-31T05:29:09.3527861Z inflating: build/bin/c10_DispatchKeySet_test 2023-03-31T05:29:09.3576740Z inflating: build/bin/c10_SymInt_test 2023-03-31T05:29:09.3629679Z inflating: build/bin/c10_InlineDeviceGuard_test 2023-03-31T05:29:09.3684016Z inflating: build/bin/c10_InlineStreamGuard_test 2023-03-31T05:29:09.3738918Z inflating: build/bin/c10_SizesAndStrides_test 2023-03-31T05:29:09.3786917Z inflating: build/bin/c10_Array_test 2023-03-31T05:29:09.3838200Z inflating: build/bin/c10_Bitset_test 2023-03-31T05:29:09.3888088Z inflating: build/bin/c10_C++17_test 2023-03-31T05:29:09.3936263Z inflating: build/bin/c10_ConstexprCrc_test 2023-03-31T05:29:09.3985164Z inflating: build/bin/c10_DeadlockDetection_test 2023-03-31T05:29:09.4039622Z inflating: build/bin/c10_LeftRight_test 2023-03-31T05:29:09.4088466Z inflating: build/bin/c10_Half_test 2023-03-31T05:29:09.4147354Z inflating: build/bin/c10_Metaprogramming_test 2023-03-31T05:29:09.4288677Z inflating: build/bin/c10_SmallVectorTest 2023-03-31T05:29:09.4342005Z inflating: build/bin/c10_ThreadLocal_test 2023-03-31T05:29:09.4391247Z inflating: build/bin/c10_Synchronized_test 2023-03-31T05:29:09.4440879Z inflating: build/bin/c10_TypeList_test 2023-03-31T05:29:09.4491871Z inflating: build/bin/c10_TypeIndex_test 2023-03-31T05:29:09.4539863Z inflating: build/bin/c10_TypeTraits_test 2023-03-31T05:29:09.4590667Z inflating: build/bin/c10_accumulate_test 2023-03-31T05:29:09.4643419Z inflating: build/bin/c10_bfloat16_test 2023-03-31T05:29:09.4700183Z inflating: build/bin/c10_complex_math_test 2023-03-31T05:29:09.4753759Z inflating: build/bin/c10_complex_test 2023-03-31T05:29:09.4850522Z inflating: build/bin/c10_either_test 2023-03-31T05:29:09.4901687Z inflating: build/bin/c10_exception_test 2023-03-31T05:29:09.4951795Z inflating: build/bin/c10_flags_test 2023-03-31T05:29:09.5091293Z inflating: build/bin/c10_intrusive_ptr_test 2023-03-31T05:29:09.5144401Z inflating: build/bin/c10_logging_test 2023-03-31T05:29:09.5193771Z inflating: build/bin/c10_irange_test 2023-03-31T05:29:09.5268313Z inflating: build/bin/c10_optional_test 2023-03-31T05:29:09.5324259Z inflating: build/bin/c10_string_view_test 2023-03-31T05:29:09.5376885Z inflating: build/bin/c10_registry_test 2023-03-31T05:29:09.5437042Z inflating: build/bin/c10_ordered_preserving_dict_test 2023-03-31T05:29:09.5487594Z inflating: build/bin/c10_ssize_test 2023-03-31T05:29:09.5537107Z inflating: build/bin/c10_tempfile_test 2023-03-31T05:29:09.5579637Z inflating: build/bin/c10_intrusive_ptr_benchmark 2023-03-31T05:29:09.5631913Z inflating: build/bin/c10_typeid_test 2023-03-31T05:29:09.6022055Z inflating: build/bin/protoc-3.13.0.0 2023-03-31T05:29:09.6411150Z inflating: build/bin/protoc 2023-03-31T05:29:09.6720614Z inflating: build/bin/vec_test_all_types_DEFAULT 2023-03-31T05:29:09.7044545Z inflating: build/bin/vec_test_all_types_AVX512 2023-03-31T05:29:09.7380819Z inflating: build/bin/vec_test_all_types_AVX2 2023-03-31T05:29:09.7438204Z inflating: build/bin/TCPStoreTest 2023-03-31T05:29:09.7489951Z inflating: build/bin/HashStoreTest 2023-03-31T05:29:09.7541794Z inflating: build/bin/FileStoreTest 2023-03-31T05:29:09.7606520Z inflating: build/bin/ProcessGroupGlooTest 2023-03-31T05:29:09.7609429Z inflating: build/bin/example_allreduce 2023-03-31T05:29:09.7663284Z inflating: build/bin/test_dist_autograd 2023-03-31T05:29:09.7665626Z inflating: build/bin/parallel_benchmark 2023-03-31T05:29:09.7732540Z inflating: build/bin/test_cpp_rpc 2023-03-31T05:29:09.7741579Z inflating: build/bin/aot_model_compiler_test 2023-03-31T05:29:09.7807637Z inflating: build/bin/test_mobile_nnc 2023-03-31T05:29:09.8145894Z inflating: build/bin/test_lazy 2023-03-31T05:29:09.8200473Z inflating: build/bin/static_runtime_bench 2023-03-31T05:29:09.8250576Z inflating: build/bin/Dimname_test 2023-03-31T05:29:09.8319648Z inflating: build/bin/Dict_test 2023-03-31T05:29:09.8374674Z inflating: build/bin/NamedTensor_test 2023-03-31T05:29:09.8688860Z inflating: build/bin/static_runtime_test 2023-03-31T05:29:09.8752672Z inflating: build/bin/MaybeOwned_test 2023-03-31T05:29:09.8809061Z inflating: build/bin/apply_utils_test 2023-03-31T05:29:09.8866189Z inflating: build/bin/atest 2023-03-31T05:29:09.8928054Z inflating: build/bin/basic 2023-03-31T05:29:09.8980187Z inflating: build/bin/broadcast_test 2023-03-31T05:29:09.9035866Z inflating: build/bin/cpu_generator_test 2023-03-31T05:29:09.9087341Z inflating: build/bin/cpu_profiling_allocator_test 2023-03-31T05:29:09.9135563Z inflating: build/bin/dispatch_key_set_test 2023-03-31T05:29:09.9224715Z inflating: build/bin/cpu_rng_test 2023-03-31T05:29:09.9273220Z inflating: build/bin/dlconvertor_test 2023-03-31T05:29:09.9330689Z inflating: build/bin/extension_backend_test 2023-03-31T05:29:09.9381607Z inflating: build/bin/half_test 2023-03-31T05:29:09.9477516Z inflating: build/bin/ivalue_test 2023-03-31T05:29:09.9525608Z inflating: build/bin/lazy_tensor_test 2023-03-31T05:29:09.9576964Z inflating: build/bin/math_kernel_test 2023-03-31T05:29:09.9628446Z inflating: build/bin/memory_format_test 2023-03-31T05:29:09.9679026Z inflating: build/bin/memory_overlapping_test 2023-03-31T05:29:09.9730157Z inflating: build/bin/mobile_memory_cleanup 2023-03-31T05:29:09.9785855Z inflating: build/bin/native_test 2023-03-31T05:29:09.9834444Z inflating: build/bin/operator_name_test 2023-03-31T05:29:09.9883686Z inflating: build/bin/operators_test 2023-03-31T05:29:09.9953930Z inflating: build/bin/pow_test 2023-03-31T05:29:10.0004129Z inflating: build/bin/packedtensoraccessor_test 2023-03-31T05:29:10.0058568Z inflating: build/bin/quantized_test 2023-03-31T05:29:10.0107043Z inflating: build/bin/reduce_ops_test 2023-03-31T05:29:10.0156042Z inflating: build/bin/reportMemoryUsage_test 2023-03-31T05:29:10.0211199Z inflating: build/bin/scalar_tensor_test 2023-03-31T05:29:10.0266494Z inflating: build/bin/scalar_test 2023-03-31T05:29:10.0347770Z inflating: build/bin/tensor_iterator_test 2023-03-31T05:29:10.0350003Z inflating: build/bin/thread_init_test 2023-03-31T05:29:10.0400631Z inflating: build/bin/type_ptr_test 2023-03-31T05:29:10.0448820Z inflating: build/bin/variant_test 2023-03-31T05:29:10.0449755Z inflating: build/bin/verify_api_visibility 2023-03-31T05:29:10.0500412Z inflating: build/bin/undefined_tensor_test 2023-03-31T05:29:10.0560070Z inflating: build/bin/type_test 2023-03-31T05:29:10.0648298Z inflating: build/bin/legacy_vmap_test 2023-03-31T05:29:10.0712534Z inflating: build/bin/weakref_test 2023-03-31T05:29:10.0760861Z inflating: build/bin/xla_tensor_test 2023-03-31T05:29:10.0818257Z inflating: build/bin/IListRef_test 2023-03-31T05:29:10.0868364Z inflating: build/bin/wrapdim_test 2023-03-31T05:29:10.0966736Z inflating: build/bin/List_test 2023-03-31T05:29:10.1096035Z inflating: build/bin/kernel_function_legacy_test 2023-03-31T05:29:10.1192421Z inflating: build/bin/kernel_function_test 2023-03-31T05:29:10.1324301Z inflating: build/bin/kernel_lambda_legacy_test 2023-03-31T05:29:10.1386485Z inflating: build/bin/KernelFunction_test 2023-03-31T05:29:10.1489842Z inflating: build/bin/kernel_lambda_test 2023-03-31T05:29:10.1539445Z inflating: build/bin/CppSignature_test 2023-03-31T05:29:10.1598060Z inflating: build/bin/kernel_stackbased_test 2023-03-31T05:29:10.1694918Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2023-03-31T05:29:10.1742123Z inflating: build/bin/op_allowlist_test 2023-03-31T05:29:10.1796009Z inflating: build/bin/backend_fallback_test 2023-03-31T05:29:10.1847187Z inflating: build/bin/inline_container_test 2023-03-31T05:29:10.2160760Z inflating: build/bin/op_registration_test 2023-03-31T05:29:10.2218626Z inflating: build/bin/test_edge_op_registration 2023-03-31T05:29:10.2237190Z inflating: build/bin/tutorial_tensorexpr 2023-03-31T05:29:10.3605009Z inflating: build/bin/test_api 2023-03-31T05:29:10.3609306Z inflating: build/bin/torch_shm_manager 2023-03-31T05:29:10.4790029Z inflating: build/bin/test_tensorexpr 2023-03-31T05:29:10.5379920Z inflating: build/bin/test_jit 2023-03-31T05:29:10.5381743Z inflating: .pytorch-test-times.json 2023-03-31T05:29:10.5406263Z ##[group]Run df -H 2023-03-31T05:29:10.5406449Z df -H 2023-03-31T05:29:10.5417371Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:29:10.5417603Z env: 2023-03-31T05:29:10.5417775Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:10.5418005Z ##[endgroup] 2023-03-31T05:29:10.5824766Z Filesystem Size Used Avail Use% Mounted on 2023-03-31T05:29:10.5825056Z devtmpfs 8.1G 0 8.1G 0% /dev 2023-03-31T05:29:10.5825270Z tmpfs 8.2G 0 8.2G 0% /dev/shm 2023-03-31T05:29:10.5825495Z tmpfs 8.2G 426k 8.2G 1% /run 2023-03-31T05:29:10.5825718Z tmpfs 8.2G 0 8.2G 0% /sys/fs/cgroup 2023-03-31T05:29:10.5825929Z /dev/nvme0n1p1 162G 25G 137G 16% / 2023-03-31T05:29:10.5863161Z ##[group]Run .github/scripts/parse_ref.py 2023-03-31T05:29:10.5863420Z .github/scripts/parse_ref.py 2023-03-31T05:29:10.5874431Z shell: /usr/bin/bash -e {0} 2023-03-31T05:29:10.5874629Z env: 2023-03-31T05:29:10.5874802Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:10.5875002Z ##[endgroup] 2023-03-31T05:29:10.6127830Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2023-03-31T05:29:10.6128153Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2023-03-31T05:29:10.6139324Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T05:29:10.6139560Z env: 2023-03-31T05:29:10.6139748Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:10.6139936Z JOB_TIMEOUT: 240 2023-03-31T05:29:10.6140119Z ##[endgroup] 2023-03-31T05:29:10.6234924Z ##[group]Run set -x 2023-03-31T05:29:10.6235213Z set -x 2023-03-31T05:29:10.6235389Z  2023-03-31T05:29:10.6235584Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2023-03-31T05:29:10.6235845Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2023-03-31T05:29:10.6236116Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2023-03-31T05:29:10.6236358Z  TEST_COMMAND=.ci/onnx/test.sh 2023-03-31T05:29:10.6236546Z else 2023-03-31T05:29:10.6236756Z  TEST_COMMAND=.ci/pytorch/test.sh 2023-03-31T05:29:10.6236959Z fi 2023-03-31T05:29:10.6237109Z  2023-03-31T05:29:10.6237354Z COMMIT_MESSAGES=$(git cherry -v "origin/${GIT_DEFAULT_BRANCH:-master}") 2023-03-31T05:29:10.6237611Z  2023-03-31T05:29:10.6237829Z # sanitize the input commit message and PR body here: 2023-03-31T05:29:10.6238050Z # 2023-03-31T05:29:10.6238361Z # trim all new lines from commit messages + PR_BODY to avoid issues with batch environment 2023-03-31T05:29:10.6238744Z # variable copying. see https://github.com/pytorch/pytorch/pull/80043#issuecomment-1167796028 2023-03-31T05:29:10.6239072Z COMMIT_MESSAGES="${COMMIT_MESSAGES//[$'\n\r']}" 2023-03-31T05:29:10.6239314Z PR_BODY="${PR_BODY//[$'\n\r']}" 2023-03-31T05:29:10.6239497Z  2023-03-31T05:29:10.6239768Z # then trim all special characters like single and double quotes to avoid unescaped inputs to 2023-03-31T05:29:10.6240058Z # wreak havoc internally 2023-03-31T05:29:10.6240302Z export COMMIT_MESSAGES="${COMMIT_MESSAGES//[\'\"]}" 2023-03-31T05:29:10.6240541Z export PR_BODY="${PR_BODY//[\'\"]}" 2023-03-31T05:29:10.6240743Z  2023-03-31T05:29:10.6240977Z # detached container should get cleaned up by teardown_ec2_linux 2023-03-31T05:29:10.6241270Z # TODO: Stop building test binaries as part of the build phase 2023-03-31T05:29:10.6241560Z # Used for GPU_FLAG since that doesn't play nice 2023-03-31T05:29:10.6241811Z # shellcheck disable=SC2086,SC2090 2023-03-31T05:29:10.6242029Z container_name=$(docker run \ 2023-03-31T05:29:10.6242238Z  ${GPU_FLAG:-} \ 2023-03-31T05:29:10.6242443Z  -e BUILD_ENVIRONMENT \ 2023-03-31T05:29:10.6242654Z  -e PR_NUMBER \ 2023-03-31T05:29:10.6242841Z  -e GITHUB_ACTIONS \ 2023-03-31T05:29:10.6243221Z  -e BASE_SHA \ 2023-03-31T05:29:10.6243414Z  -e BRANCH \ 2023-03-31T05:29:10.6243588Z  -e SHA1 \ 2023-03-31T05:29:10.6243789Z  -e AWS_DEFAULT_REGION \ 2023-03-31T05:29:10.6243997Z  -e IN_WHEEL_TEST \ 2023-03-31T05:29:10.6244192Z  -e SHARD_NUMBER \ 2023-03-31T05:29:10.6244392Z  -e TEST_CONFIG \ 2023-03-31T05:29:10.6244599Z  -e NUM_TEST_SHARDS \ 2023-03-31T05:29:10.6244783Z  -e PR_BODY \ 2023-03-31T05:29:10.6244985Z  -e COMMIT_MESSAGES \ 2023-03-31T05:29:10.6245803Z  -e CONTINUE_THROUGH_ERROR \ 2023-03-31T05:29:10.6246022Z  -e PYTORCH_RETRY_TEST_CASES \ 2023-03-31T05:29:10.6246264Z  -e PYTORCH_OVERRIDE_FLAKY_SIGNAL \ 2023-03-31T05:29:10.6246487Z  -e PR_LABELS \ 2023-03-31T05:29:10.6246707Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2023-03-31T05:29:10.6246915Z  -e SCCACHE_BUCKET \ 2023-03-31T05:29:10.6247128Z  -e SCCACHE_S3_KEY_PREFIX \ 2023-03-31T05:29:10.6247332Z  -e XLA_CUDA \ 2023-03-31T05:29:10.6247534Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2023-03-31T05:29:10.6247777Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2023-03-31T05:29:10.6248189Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2023-03-31T05:29:10.6248429Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2023-03-31T05:29:10.6248698Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2023-03-31T05:29:10.6249036Z  --ulimit stack=10485760:83886080 \ 2023-03-31T05:29:10.6249276Z  --security-opt seccomp=unconfined \ 2023-03-31T05:29:10.6249496Z  --cap-add=SYS_PTRACE \ 2023-03-31T05:29:10.6249696Z  --ipc=host \ 2023-03-31T05:29:10.6249899Z  --shm-size="${SHM_SIZE}" \ 2023-03-31T05:29:10.6250085Z  --tty \ 2023-03-31T05:29:10.6250270Z  --detach \ 2023-03-31T05:29:10.6250479Z  --name="${container_name}" \ 2023-03-31T05:29:10.6250673Z  --user jenkins \ 2023-03-31T05:29:10.6250920Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2023-03-31T05:29:10.6251181Z  -w /var/lib/jenkins/workspace \ 2023-03-31T05:29:10.6251382Z  "${DOCKER_IMAGE}" 2023-03-31T05:29:10.6251566Z ) 2023-03-31T05:29:10.6251794Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2023-03-31T05:29:10.6252145Z docker exec -t "${container_name}" sh -c "pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2023-03-31T05:29:10.6263139Z shell: /usr/bin/bash -e {0} 2023-03-31T05:29:10.6263339Z env: 2023-03-31T05:29:10.6263525Z GIT_DEFAULT_BRANCH: master 2023-03-31T05:29:10.6263761Z BUILD_ENVIRONMENT: linux-bionic-py3.8-clang9 2023-03-31T05:29:10.6263992Z PR_NUMBER: 97655 2023-03-31T05:29:10.6264181Z BRANCH: pull/97655 2023-03-31T05:29:10.6264392Z SHA1: 4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:29:10.6264658Z BASE_SHA: ea1d5ab619567884455a87f53d7eefb5b60a4c0b 2023-03-31T05:29:10.6264895Z PYTORCH_RETRY_TEST_CASES: 1 2023-03-31T05:29:10.6265103Z PYTORCH_OVERRIDE_FLAKY_SIGNAL: 1 2023-03-31T05:29:10.6265312Z TEST_CONFIG: default 2023-03-31T05:29:10.6265501Z SHARD_NUMBER: 1 2023-03-31T05:29:10.6265678Z NUM_TEST_SHARDS: 3 2023-03-31T05:29:10.6266196Z PR_BODY: Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom): * #89743 * #94977 * #94964 * __->__ #97655 * #90755 cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire 2023-03-31T05:29:10.6266689Z CONTINUE_THROUGH_ERROR: False 2023-03-31T05:29:10.6266949Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2023-03-31T05:29:10.6267203Z SCCACHE_S3_KEY_PREFIX: pull 2023-03-31T05:29:10.6267384Z SHM_SIZE: 1g 2023-03-31T05:29:10.6267742Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:29:10.6268084Z XLA_CUDA: 2023-03-31T05:29:10.6268346Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2023-03-31T05:29:10.6268643Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2023-03-31T05:29:10.6268881Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2023-03-31T05:29:10.6269074Z ##[endgroup] 2023-03-31T05:29:10.6293349Z + [[ default == \m\u\l\t\i\g\p\u ]] 2023-03-31T05:29:10.6294031Z + [[ linux-bionic-py3.8-clang9 == *onnx* ]] 2023-03-31T05:29:10.6294294Z + TEST_COMMAND=.ci/pytorch/test.sh 2023-03-31T05:29:10.6296675Z ++ git cherry -v origin/master 2023-03-31T05:29:10.7214917Z + COMMIT_MESSAGES='+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor 2023-03-31T05:29:10.7215702Z + d0ea2938333efb927419758dbe4978a63db8ee06 Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7216172Z + 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7216515Z + b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7216840Z + 33a9875107144a3f813754949eaf40670de3434b Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7217153Z + 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7217755Z + a625ea99d524c33246a1d576404d80266a4f37b4 Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7218104Z + 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7218580Z + cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7218909Z + 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7219258Z + b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7219624Z + 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on "Add a cpp wrapper for Inductor" 2023-03-31T05:29:10.7219984Z + d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output 2023-03-31T05:29:10.7220319Z + 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction 2023-03-31T05:29:10.7220692Z + 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on "Inductor cpp wrapper: support Reduction" 2023-03-31T05:29:10.7221098Z + f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on "Inductor cpp wrapper: support Reduction" 2023-03-31T05:29:10.7221467Z + 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input 2023-03-31T05:29:10.7221844Z + 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call 2023-03-31T05:29:10.7222245Z + f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on "Inductor cpp wrapper: support bmm, mm, addmm extern call" 2023-03-31T05:29:10.7222747Z + 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on "Inductor cpp wrapper: support bmm, mm, addmm extern call" 2023-03-31T05:29:10.7223127Z + 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel 2023-03-31T05:29:10.7223465Z + 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper 2023-03-31T05:29:10.7223831Z + 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7224228Z + 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7224589Z + 29da460ce369580d10e337591399928b435347cb Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7224967Z + 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7225358Z + e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7225734Z + 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7226085Z + c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7226462Z + 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7226851Z + ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7227234Z + 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7227623Z + f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7228001Z + 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7228382Z + c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7228755Z + 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7229297Z + 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7229760Z + c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7230209Z + 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7230580Z + 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7230951Z + 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7231332Z + 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7231724Z + 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on "Inductor cpp wrapper: cache the wrapper" 2023-03-31T05:29:10.7232087Z + 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace 2023-03-31T05:29:10.7232467Z + f861398f8594246d54ee7ebc38eaf2463ad25252 Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7232885Z + 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7233306Z + 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7233736Z + 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7234242Z + b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7234662Z + 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7235064Z + e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7235492Z + b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7235926Z + 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7236360Z + 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7236753Z + 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace" 2023-03-31T05:29:10.7237130Z + 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary 2023-03-31T05:29:10.7237509Z + 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on "Inductor cpp wrapper: support LinearUnary" 2023-03-31T05:29:10.7237994Z + 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on "Inductor cpp wrapper: support LinearUnary" 2023-03-31T05:29:10.7238447Z + bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on "Inductor cpp wrapper: support LinearUnary" 2023-03-31T05:29:10.7238858Z + b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on "Inductor cpp wrapper: support LinearUnary" 2023-03-31T05:29:10.7239265Z + ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on "Inductor cpp wrapper: support LinearUnary" 2023-03-31T05:29:10.7239668Z + ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on "Inductor cpp wrapper: support LinearUnary"' 2023-03-31T05:29:10.7265539Z + COMMIT_MESSAGES='+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor+ d0ea2938333efb927419758dbe4978a63db8ee06 Update on "Add a cpp wrapper for Inductor"+ 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on "Add a cpp wrapper for Inductor"+ b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on "Add a cpp wrapper for Inductor"+ 33a9875107144a3f813754949eaf40670de3434b Update on "Add a cpp wrapper for Inductor"+ 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on "Add a cpp wrapper for Inductor"+ a625ea99d524c33246a1d576404d80266a4f37b4 Update on "Add a cpp wrapper for Inductor"+ 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on "Add a cpp wrapper for Inductor"+ cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on "Add a cpp wrapper for Inductor"+ 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on "Add a cpp wrapper for Inductor"+ b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on "Add a cpp wrapper for Inductor"+ 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on "Add a cpp wrapper for Inductor"+ d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output+ 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction+ 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on "Inductor cpp wrapper: support Reduction"+ f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on "Inductor cpp wrapper: support Reduction"+ 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input+ 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call+ f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on "Inductor cpp wrapper: support bmm, mm, addmm extern call"+ 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on "Inductor cpp wrapper: support bmm, mm, addmm extern call"+ 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel+ 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper+ 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 29da460ce369580d10e337591399928b435347cb Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on "Inductor cpp wrapper: cache the wrapper"+ 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace+ f861398f8594246d54ee7ebc38eaf2463ad25252 Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on "Inductor cpp wrapper: support ConvBinaryInplace"+ 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary+ 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on "Inductor cpp wrapper: support LinearUnary"+ 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on "Inductor cpp wrapper: support LinearUnary"+ bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on "Inductor cpp wrapper: support LinearUnary"+ b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on "Inductor cpp wrapper: support LinearUnary"+ ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on "Inductor cpp wrapper: support LinearUnary"+ ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on "Inductor cpp wrapper: support LinearUnary"' 2023-03-31T05:29:10.7273153Z + PR_BODY='Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom):* #89743* #94977* #94964* __->__ #97655* #90755cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire' 2023-03-31T05:29:10.7303926Z + export 'COMMIT_MESSAGES=+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor+ d0ea2938333efb927419758dbe4978a63db8ee06 Update on Add a cpp wrapper for Inductor+ 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on Add a cpp wrapper for Inductor+ b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on Add a cpp wrapper for Inductor+ 33a9875107144a3f813754949eaf40670de3434b Update on Add a cpp wrapper for Inductor+ 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on Add a cpp wrapper for Inductor+ a625ea99d524c33246a1d576404d80266a4f37b4 Update on Add a cpp wrapper for Inductor+ 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on Add a cpp wrapper for Inductor+ cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on Add a cpp wrapper for Inductor+ 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on Add a cpp wrapper for Inductor+ b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on Add a cpp wrapper for Inductor+ 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on Add a cpp wrapper for Inductor+ d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output+ 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction+ 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on Inductor cpp wrapper: support Reduction+ f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on Inductor cpp wrapper: support Reduction+ 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input+ 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call+ f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel+ 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper+ 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on Inductor cpp wrapper: cache the wrapper+ 29da460ce369580d10e337591399928b435347cb Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on Inductor cpp wrapper: cache the wrapper+ e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on Inductor cpp wrapper: cache the wrapper+ ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on Inductor cpp wrapper: cache the wrapper+ 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on Inductor cpp wrapper: cache the wrapper+ f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on Inductor cpp wrapper: cache the wrapper+ c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace+ f861398f8594246d54ee7ebc38eaf2463ad25252 Update on Inductor cpp wrapper: support ConvBinaryInplace+ 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary+ 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on Inductor cpp wrapper: support LinearUnary+ 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on Inductor cpp wrapper: support LinearUnary+ bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on Inductor cpp wrapper: support LinearUnary+ b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on Inductor cpp wrapper: support LinearUnary' 2023-03-31T05:29:10.7322366Z + COMMIT_MESSAGES='+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor+ d0ea2938333efb927419758dbe4978a63db8ee06 Update on Add a cpp wrapper for Inductor+ 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on Add a cpp wrapper for Inductor+ b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on Add a cpp wrapper for Inductor+ 33a9875107144a3f813754949eaf40670de3434b Update on Add a cpp wrapper for Inductor+ 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on Add a cpp wrapper for Inductor+ a625ea99d524c33246a1d576404d80266a4f37b4 Update on Add a cpp wrapper for Inductor+ 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on Add a cpp wrapper for Inductor+ cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on Add a cpp wrapper for Inductor+ 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on Add a cpp wrapper for Inductor+ b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on Add a cpp wrapper for Inductor+ 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on Add a cpp wrapper for Inductor+ d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output+ 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction+ 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on Inductor cpp wrapper: support Reduction+ f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on Inductor cpp wrapper: support Reduction+ 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input+ 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call+ f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel+ 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper+ 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on Inductor cpp wrapper: cache the wrapper+ 29da460ce369580d10e337591399928b435347cb Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on Inductor cpp wrapper: cache the wrapper+ e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on Inductor cpp wrapper: cache the wrapper+ ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on Inductor cpp wrapper: cache the wrapper+ 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on Inductor cpp wrapper: cache the wrapper+ f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on Inductor cpp wrapper: cache the wrapper+ c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace+ f861398f8594246d54ee7ebc38eaf2463ad25252 Update on Inductor cpp wrapper: support ConvBinaryInplace+ 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary+ 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on Inductor cpp wrapper: support LinearUnary+ 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on Inductor cpp wrapper: support LinearUnary+ bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on Inductor cpp wrapper: support LinearUnary+ b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on Inductor cpp wrapper: support LinearUnary' 2023-03-31T05:29:10.7329907Z + export 'PR_BODY=Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom):* #89743* #94977* #94964* __->__ #97655* #90755cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire' 2023-03-31T05:29:10.7330916Z + PR_BODY='Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom):* #89743* #94977* #94964* __->__ #97655* #90755cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire' 2023-03-31T05:29:10.7331416Z +++ nproc --ignore=2 2023-03-31T05:29:10.7480075Z ++ docker run -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e PR_BODY -e COMMIT_MESSAGES -e CONTINUE_THROUGH_ERROR -e PYTORCH_RETRY_TEST_CASES -e PYTORCH_OVERRIDE_FLAKY_SIGNAL -e PR_LABELS -e MAX_JOBS=6 -e SCCACHE_BUCKET -e SCCACHE_S3_KEY_PREFIX -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 --env-file=/tmp/github_env_4571766038 --ulimit stack=10485760:83886080 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=1g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T05:29:22.9310065Z + container_name=ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T05:29:22.9310809Z + echo DOCKER_CONTAINER_ID=ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T05:29:22.9314787Z ++ echo dist/torch-2.1.0a0+git4af0d9f-cp38-cp38-linux_x86_64.whl 2023-03-31T05:29:22.9316790Z + docker exec -t ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db sh -c 'pip install dist/torch-2.1.0a0+git4af0d9f-cp38-cp38-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2023-03-31T05:29:23.8937563Z Processing ./dist/torch-2.1.0a0+git4af0d9f-cp38-cp38-linux_x86_64.whl 2023-03-31T05:29:24.3105185Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (3.10.7) 2023-03-31T05:29:24.3109023Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (2.6.3) 2023-03-31T05:29:24.3112360Z Requirement already satisfied: sympy in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (1.11.1) 2023-03-31T05:29:24.3115774Z Requirement already satisfied: typing-extensions in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (4.5.0) 2023-03-31T05:29:24.3118938Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (3.1.2) 2023-03-31T05:29:24.3129966Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch==2.1.0a0+git4af0d9f) (3.3.0) 2023-03-31T05:29:24.3191253Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from opt-einsum>=3.3->torch==2.1.0a0+git4af0d9f) (1.21.2) 2023-03-31T05:29:24.3288754Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from jinja2->torch==2.1.0a0+git4af0d9f) (2.1.2) 2023-03-31T05:29:24.3451828Z Requirement already satisfied: mpmath>=0.19 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from sympy->torch==2.1.0a0+git4af0d9f) (1.3.0) 2023-03-31T05:29:25.0746937Z Installing collected packages: torch 2023-03-31T05:29:31.1208535Z Successfully installed torch-2.1.0a0+git4af0d9f 2023-03-31T05:29:31.2040544Z + echo 'Environment variables:' 2023-03-31T05:29:31.2040857Z Environment variables: 2023-03-31T05:29:31.2041046Z + env 2023-03-31T05:29:31.2059636Z SHARD_NUMBER=1 2023-03-31T05:29:31.2060024Z UCC_HOME=/usr 2023-03-31T05:29:31.2060490Z BUILD_ENVIRONMENT=linux-bionic-py3.8-clang9 2023-03-31T05:29:31.2060837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2023-03-31T05:29:31.2061064Z INSTALLED_DB=yes 2023-03-31T05:29:31.2061332Z HOSTNAME=ad599e049a6e 2023-03-31T05:29:31.2061729Z GITHUB_REF_NAME=97655/merge 2023-03-31T05:29:31.2062142Z GITHUB_API_URL=https://api.github.com 2023-03-31T05:29:31.2062470Z GITHUB_REPOSITORY_OWNER_ID=21003710 2023-03-31T05:29:31.2062765Z OPENSSL_DIR=/opt/openssl 2023-03-31T05:29:31.2062945Z UCC_COMMIT= 2023-03-31T05:29:31.2063413Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2063745Z CUDA_PATH=/usr/local/cuda 2023-03-31T05:29:31.2064125Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2023-03-31T05:29:31.2064479Z GITHUB_RUN_ATTEMPT=1 2023-03-31T05:29:31.2064669Z INSTALLED_THRIFT= 2023-03-31T05:29:31.2064883Z TEST_CONFIG=default 2023-03-31T05:29:31.2065403Z GITHUB_REPOSITORY_OWNER=pytorch 2023-03-31T05:29:31.2065610Z GITHUB_ACTIONS=true 2023-03-31T05:29:31.2065874Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/pull.yml@refs/pull/97655/merge 2023-03-31T05:29:31.2066131Z CI=true 2023-03-31T05:29:31.2066436Z PYTORCH_OVERRIDE_FLAKY_SIGNAL=1 2023-03-31T05:29:31.2066631Z BRANCH=pull/97655 2023-03-31T05:29:31.2066888Z GITHUB_HEAD_REF=gh/chunyuan-w/41/head 2023-03-31T05:29:31.2067094Z UCX_COMMIT= 2023-03-31T05:29:31.2067283Z GITHUB_ACTOR=chunyuan-w 2023-03-31T05:29:31.2067478Z GITHUB_ACTION_REF= 2023-03-31T05:29:31.2067669Z GITHUB_ACTION=__self 2023-03-31T05:29:31.2067857Z GITHUB_REF_PROTECTED=false 2023-03-31T05:29:31.2068217Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2023-03-31T05:29:31.2068516Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2023-03-31T05:29:31.2070308Z *** 2023-03-31T05:29:31.2070483Z INSTALLED_VISION=yes 2023-03-31T05:29:31.2070682Z HOME=/var/lib/jenkins 2023-03-31T05:29:31.2071133Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2071458Z GITHUB_ACTION_REPOSITORY= 2023-03-31T05:29:31.2071673Z GITHUB_REF_TYPE=branch 2023-03-31T05:29:31.2071875Z GITHUB_RETENTION_DAYS=90 2023-03-31T05:29:31.2072171Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2023-03-31T05:29:31.2072637Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2072953Z DEBIAN_FRONTEND=noninteractive 2023-03-31T05:29:31.2073177Z GITHUB_REF=refs/pull/97655/merge 2023-03-31T05:29:31.2073406Z GITHUB_SHA=48284fffe3d9196187bcf2fdf764bf15956e81e8 2023-03-31T05:29:31.2073637Z INSTALLED_PROTOBUF=yes 2023-03-31T05:29:31.2073843Z ANACONDA_PYTHON_VERSION=3.8 2023-03-31T05:29:31.2074024Z DESIRED_CUDA= 2023-03-31T05:29:31.2074221Z GITHUB_REPOSITORY_ID=65600975 2023-03-31T05:29:31.2074428Z GITHUB_RUN_ID=4571766038 2023-03-31T05:29:31.2074616Z CONTINUE_THROUGH_ERROR=False 2023-03-31T05:29:31.2074857Z GITHUB_SERVER_URL=https://github.com 2023-03-31T05:29:31.2090438Z MAX_JOBS=6 2023-03-31T05:29:31.2090725Z GITHUB_ACTOR_ID=65992142 2023-03-31T05:29:31.2091138Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2023-03-31T05:29:31.2091431Z UCX_HOME=/usr 2023-03-31T05:29:31.2091630Z PYTORCH_RETRY_TEST_CASES=1 2023-03-31T05:29:31.2091897Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2023-03-31T05:29:31.2092163Z BASE_SHA=ea1d5ab619567884455a87f53d7eefb5b60a4c0b 2023-03-31T05:29:31.2092382Z INSTALLED_ANDROID= 2023-03-31T05:29:31.2093106Z PR_BODY=Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom):* #89743* #94977* #94964* __->__ #97655* #90755cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire 2023-03-31T05:29:31.2093674Z GITHUB_BASE_REF=gh/chunyuan-w/41/base 2023-03-31T05:29:31.2093865Z TERM=xterm 2023-03-31T05:29:31.2094040Z XLA_CUDA= 2023-03-31T05:29:31.2094228Z TORCH_CUDA_ARCH_LIST=Maxwell 2023-03-31T05:29:31.2094427Z OPENSSL_ROOT_DIR=/opt/openssl 2023-03-31T05:29:31.2095006Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2095318Z GITHUB_JOB=test 2023-03-31T05:29:31.2095502Z SCCACHE_S3_KEY_PREFIX=pull 2023-03-31T05:29:31.2102945Z COMMIT_MESSAGES=+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor+ d0ea2938333efb927419758dbe4978a63db8ee06 Update on Add a cpp wrapper for Inductor+ 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on Add a cpp wrapper for Inductor+ b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on Add a cpp wrapper for Inductor+ 33a9875107144a3f813754949eaf40670de3434b Update on Add a cpp wrapper for Inductor+ 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on Add a cpp wrapper for Inductor+ a625ea99d524c33246a1d576404d80266a4f37b4 Update on Add a cpp wrapper for Inductor+ 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on Add a cpp wrapper for Inductor+ cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on Add a cpp wrapper for Inductor+ 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on Add a cpp wrapper for Inductor+ b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on Add a cpp wrapper for Inductor+ 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on Add a cpp wrapper for Inductor+ d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output+ 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction+ 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on Inductor cpp wrapper: support Reduction+ f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on Inductor cpp wrapper: support Reduction+ 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input+ 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call+ f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel+ 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper+ 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on Inductor cpp wrapper: cache the wrapper+ 29da460ce369580d10e337591399928b435347cb Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on Inductor cpp wrapper: cache the wrapper+ e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on Inductor cpp wrapper: cache the wrapper+ ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on Inductor cpp wrapper: cache the wrapper+ 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on Inductor cpp wrapper: cache the wrapper+ f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on Inductor cpp wrapper: cache the wrapper+ c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace+ f861398f8594246d54ee7ebc38eaf2463ad25252 Update on Inductor cpp wrapper: support ConvBinaryInplace+ 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary+ 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on Inductor cpp wrapper: support LinearUnary+ 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on Inductor cpp wrapper: support LinearUnary+ bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on Inductor cpp wrapper: support LinearUnary+ b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on Inductor cpp wrapper: support LinearUnary 2023-03-31T05:29:31.2109866Z NUM_TEST_SHARDS=3 2023-03-31T05:29:31.2110055Z PR_NUMBER=97655 2023-03-31T05:29:31.2110488Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2110788Z SHLVL=1 2023-03-31T05:29:31.2110983Z GITHUB_REPOSITORY=pytorch/pytorch 2023-03-31T05:29:31.2111221Z SHA1=4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:29:31.2111451Z GITHUB_EVENT_NAME=pull_request 2023-03-31T05:29:31.2111722Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2023-03-31T05:29:31.2111932Z GITHUB_RUN_NUMBER=98082 2023-03-31T05:29:31.2112132Z GITHUB_WORKFLOW=pull 2023-03-31T05:29:31.2112486Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.8/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2023-03-31T05:29:31.2112863Z GITHUB_WORKFLOW_SHA=48284fffe3d9196187bcf2fdf764bf15956e81e8 2023-03-31T05:29:31.2113218Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T05:29:31.2113512Z GITHUB_TRIGGERING_ACTOR=chunyuan-w 2023-03-31T05:29:31.2113739Z SKIP_SCCACHE_INITIALIZATION=1 2023-03-31T05:29:31.2113925Z _=/usr/bin/env 2023-03-31T05:29:31.2114220Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2023-03-31T05:29:31.2252786Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch 2023-03-31T05:29:31.2253528Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/bin 2023-03-31T05:29:31.2254181Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib 2023-03-31T05:29:31.2254873Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/test 2023-03-31T05:29:31.2255288Z + BUILD_DIR=build 2023-03-31T05:29:31.2255583Z + BUILD_RENAMED_DIR=build_renamed 2023-03-31T05:29:31.2255918Z + BUILD_BIN_DIR=build/bin 2023-03-31T05:29:31.2256219Z + export VALGRIND=ON 2023-03-31T05:29:31.2256489Z + VALGRIND=ON 2023-03-31T05:29:31.2256906Z + [[ linux-bionic-py3.8-clang9 == *clang9* ]] 2023-03-31T05:29:31.2257288Z + export VALGRIND=OFF 2023-03-31T05:29:31.2257582Z + VALGRIND=OFF 2023-03-31T05:29:31.2257999Z + [[ linux-bionic-py3.8-clang9 != *bazel* ]] 2023-03-31T05:29:31.2258450Z ++ realpath build/custom_test_artifacts 2023-03-31T05:29:31.2278840Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2023-03-31T05:29:31.2281924Z ++ dirname .ci/pytorch/test.sh 2023-03-31T05:29:31.2303894Z + source .ci/pytorch/common.sh 2023-03-31T05:29:31.2307810Z +++ dirname .ci/pytorch/common.sh 2023-03-31T05:29:31.2316069Z ++ source .ci/pytorch/common_utils.sh 2023-03-31T05:29:31.2318542Z +++ declare -f -t trap_add 2023-03-31T05:29:31.2324431Z ++ set -ex 2023-03-31T05:29:31.2324964Z ++ [[ linux-bionic-py3.8-clang9 == *rocm* ]] 2023-03-31T05:29:31.2325330Z ++ BUILD_TEST_LIBTORCH=0 2023-03-31T05:29:31.2325760Z + echo 'Environment variables' 2023-03-31T05:29:31.2326051Z Environment variables 2023-03-31T05:29:31.2326242Z + env 2023-03-31T05:29:31.2331621Z SHARD_NUMBER=1 2023-03-31T05:29:31.2332170Z UCC_HOME=/usr 2023-03-31T05:29:31.2332804Z BUILD_ENVIRONMENT=linux-bionic-py3.8-clang9 2023-03-31T05:29:31.2333257Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2023-03-31T05:29:31.2333630Z INSTALLED_DB=yes 2023-03-31T05:29:31.2333861Z HOSTNAME=ad599e049a6e 2023-03-31T05:29:31.2334165Z GITHUB_REF_NAME=97655/merge 2023-03-31T05:29:31.2334499Z GITHUB_API_URL=https://api.github.com 2023-03-31T05:29:31.2334878Z GITHUB_REPOSITORY_OWNER_ID=21003710 2023-03-31T05:29:31.2335236Z OPENSSL_DIR=/opt/openssl 2023-03-31T05:29:31.2335554Z UCC_COMMIT= 2023-03-31T05:29:31.2336260Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2336631Z CUDA_PATH=/usr/local/cuda 2023-03-31T05:29:31.2337037Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2023-03-31T05:29:31.2337316Z GITHUB_RUN_ATTEMPT=1 2023-03-31T05:29:31.2337509Z INSTALLED_THRIFT= 2023-03-31T05:29:31.2337699Z TEST_CONFIG=default 2023-03-31T05:29:31.2337892Z GITHUB_REPOSITORY_OWNER=pytorch 2023-03-31T05:29:31.2338098Z GITHUB_ACTIONS=true 2023-03-31T05:29:31.2338377Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/pull.yml@refs/pull/97655/merge 2023-03-31T05:29:31.2338634Z CI=true 2023-03-31T05:29:31.2338810Z PYTORCH_OVERRIDE_FLAKY_SIGNAL=1 2023-03-31T05:29:31.2339024Z BRANCH=pull/97655 2023-03-31T05:29:31.2339277Z GITHUB_HEAD_REF=gh/chunyuan-w/41/head 2023-03-31T05:29:31.2339623Z UCX_COMMIT= 2023-03-31T05:29:31.2339950Z GITHUB_ACTOR=chunyuan-w 2023-03-31T05:29:31.2340219Z GITHUB_ACTION_REF= 2023-03-31T05:29:31.2340404Z GITHUB_ACTION=__self 2023-03-31T05:29:31.2340595Z VALGRIND=OFF 2023-03-31T05:29:31.2340789Z GITHUB_REF_PROTECTED=false 2023-03-31T05:29:31.2341137Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2023-03-31T05:29:31.2341437Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2023-03-31T05:29:31.2341756Z *** 2023-03-31T05:29:31.2341916Z INSTALLED_VISION=yes 2023-03-31T05:29:31.2342108Z HOME=/var/lib/jenkins 2023-03-31T05:29:31.2342538Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2342970Z GITHUB_ACTION_REPOSITORY= 2023-03-31T05:29:31.2343177Z GITHUB_REF_TYPE=branch 2023-03-31T05:29:31.2343381Z GITHUB_RETENTION_DAYS=90 2023-03-31T05:29:31.2343679Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2023-03-31T05:29:31.2344387Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2344723Z DEBIAN_FRONTEND=noninteractive 2023-03-31T05:29:31.2344945Z GITHUB_REF=refs/pull/97655/merge 2023-03-31T05:29:31.2345177Z GITHUB_SHA=48284fffe3d9196187bcf2fdf764bf15956e81e8 2023-03-31T05:29:31.2345413Z INSTALLED_PROTOBUF=yes 2023-03-31T05:29:31.2345619Z ANACONDA_PYTHON_VERSION=3.8 2023-03-31T05:29:31.2345803Z DESIRED_CUDA= 2023-03-31T05:29:31.2346023Z GITHUB_REPOSITORY_ID=65600975 2023-03-31T05:29:31.2346228Z GITHUB_RUN_ID=4571766038 2023-03-31T05:29:31.2346418Z CONTINUE_THROUGH_ERROR=False 2023-03-31T05:29:31.2346656Z GITHUB_SERVER_URL=https://github.com 2023-03-31T05:29:31.2346872Z MAX_JOBS=6 2023-03-31T05:29:31.2347038Z GITHUB_ACTOR_ID=65992142 2023-03-31T05:29:31.2347397Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2023-03-31T05:29:31.2347914Z UCX_HOME=/usr 2023-03-31T05:29:31.2348262Z PYTORCH_RETRY_TEST_CASES=1 2023-03-31T05:29:31.2348577Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2023-03-31T05:29:31.2348853Z BASE_SHA=ea1d5ab619567884455a87f53d7eefb5b60a4c0b 2023-03-31T05:29:31.2349175Z INSTALLED_ANDROID= 2023-03-31T05:29:31.2349892Z PR_BODY=Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom):* #89743* #94977* #94964* __->__ #97655* #90755cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbell10 @desertfire 2023-03-31T05:29:31.2350619Z GITHUB_BASE_REF=gh/chunyuan-w/41/base 2023-03-31T05:29:31.2350829Z TERM=xterm 2023-03-31T05:29:31.2351000Z XLA_CUDA= 2023-03-31T05:29:31.2351175Z TORCH_CUDA_ARCH_LIST=Maxwell 2023-03-31T05:29:31.2351388Z OPENSSL_ROOT_DIR=/opt/openssl 2023-03-31T05:29:31.2351825Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2352119Z GITHUB_JOB=test 2023-03-31T05:29:31.2352315Z SCCACHE_S3_KEY_PREFIX=pull 2023-03-31T05:29:31.2359722Z COMMIT_MESSAGES=+ 57d358489e4e888a9c430cd9ad8c247d77d5f66d Add a cpp wrapper for Inductor+ d0ea2938333efb927419758dbe4978a63db8ee06 Update on Add a cpp wrapper for Inductor+ 83e66a3615b0513aacb0fc636858b7afcffd00ae Update on Add a cpp wrapper for Inductor+ b07c4b359d58f47fd85f7d108595f8b61cec2c15 Update on Add a cpp wrapper for Inductor+ 33a9875107144a3f813754949eaf40670de3434b Update on Add a cpp wrapper for Inductor+ 5d93900e2788bbe9b9bf26dde2be0cda72447d77 Update on Add a cpp wrapper for Inductor+ a625ea99d524c33246a1d576404d80266a4f37b4 Update on Add a cpp wrapper for Inductor+ 4a199ab6be7abdc8ae78d9e838e2bb2b08754c7f Update on Add a cpp wrapper for Inductor+ cc8d918b44fee07ea2a8a3f3fd9ff3585a8b4622 Update on Add a cpp wrapper for Inductor+ 5a696d27da6280846b5a5ed02a00943a9f79e18c Update on Add a cpp wrapper for Inductor+ b8510900d810c958946e96535ef876e5a2e52dfe Update base for Update on Add a cpp wrapper for Inductor+ 4afea82244f2447577eb115c529f437a24a6a0dd Update base for Update on Add a cpp wrapper for Inductor+ d1103ebc91af6af793b2f123639e070db370d19f Inductor cpp wrapper: support None as output+ 29b19ffb738c728c6950ec25e77bfc28aa622c70 Inductor cpp wrapper: support Reduction+ 1d253695a3cfce77c47ffe77252e3ee33b98e186 Update base for Update on Inductor cpp wrapper: support Reduction+ f8b829c7d2337200f3b1ec3d1e43f69bac38f4cd Update base for Update on Inductor cpp wrapper: support Reduction+ 24c8a57794a1d9632a90c938b2d0eea93968cc9f Inductor cpp wrapper: support more dtypes of input+ 925efec5c46074a68c9b0f96714dff10fe18e2f2 Inductor cpp wrapper: support bmm, mm, addmm extern call+ f47e3658857f2e77fdce06dd6551e7e4441a070f Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 795de2c44e1a76795d40a07fed295d624e1b3c32 Update base for Update on Inductor cpp wrapper: support bmm, mm, addmm extern call+ 6ebcf41416b354732013f18a469d72408bfc7ee8 Inductor cpp wrapper: cache the loading of the kernel+ 56e8652266070cf938e215c36c46d48742410d80 Inductor cpp wrapper: cache the wrapper+ 967e7b183af9552a5cacf13d0e6e714b5f1e732b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 06dc771aad38673333162a17197c49f3930e9c8c Update base for Update on Inductor cpp wrapper: cache the wrapper+ 29da460ce369580d10e337591399928b435347cb Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3c77f0299ea901bfafc98346e2229c3d6ecc1419 Update base for Update on Inductor cpp wrapper: cache the wrapper+ e8c751e0d59060632f34bd4bdd224c03b132c183 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 45f386463593855df52ab0ef6e310f626752b439 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c50326295215453201fc078eaa1d30483caf65d2 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 60bfd01c9a2e01820d283e4d145a5e5c583859cd Update base for Update on Inductor cpp wrapper: cache the wrapper+ ca62f56b5d772549ec5c9c1acc90123f1803ee9a Update base for Update on Inductor cpp wrapper: cache the wrapper+ 5aee3e01bb89bb3226a9539fb4892a3d695a2e2e Update base for Update on Inductor cpp wrapper: cache the wrapper+ f6eb2e1a916550d051324e7085e4ae97df08c4d3 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 89a053cf2f5ad53f113676371adcac7337993f38 Update base for Update on Inductor cpp wrapper: cache the wrapper+ c998f7f7335ef741b5100fbaf93565168ac31195 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4cb335f8bd53cf363b8a8da2318fd76103c41020 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8932819a639a4d8b158ec1ab50663a397511a7bb Update base for Update on Inductor cpp wrapper: cache the wrapper+ c2bd178328354fb333a0c6554ae5d2806cac661b Update base for Update on Inductor cpp wrapper: cache the wrapper+ 4c69c1fb129da445598c0e0660bc8a1fc6ec9689 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 3e93b8aa308a6470fedb188a7facb71653125e96 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 221127a667b478d27f7596877f9eb36f3b60cf27 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 8d8a8c8d12d1b275cf66adec30256a474c1fd0d1 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 579b3f8c839b7a2aa8cb998dd1fee065ad006621 Update base for Update on Inductor cpp wrapper: cache the wrapper+ 2fcb3185e4c2768308e30ba1137c039553011b77 Inductor cpp wrapper: support ConvBinaryInplace+ f861398f8594246d54ee7ebc38eaf2463ad25252 Update on Inductor cpp wrapper: support ConvBinaryInplace+ 8f5510806a6a39d09ee4c3e1f2eaaff9d6f62428 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 041ad77a92fcebbede76152e92dbf7f87ef1ab06 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 2237ca4f75e9611949c1722dcc59d38e04c5cf96 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b01cbd6eb399ef3f84f155ad381ec5a1a7bf6b82 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 0b219351aef70f807b024583723f33c65ac603fb Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ e76284e659e006a2d8cfe10a5043cab133f1cf0a Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ b9c67da09b774e9ac44b728b2227a5b8e348783e Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 352fde6f3ec9e6881adc0a28bfb7ad21065f75cc Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 4489e7742b1bbb8c89b12614ac9bdf07e1141910 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 7ff5e53a63c9977c6c66673426288e3d5c739061 Update base for Update on Inductor cpp wrapper: support ConvBinaryInplace+ 9490dc802ab9abfeb69697b388ef92b23ad7420c Inductor cpp wrapper: support LinearUnary+ 3daefc80636782fd8223f819186feaa56f2e91b0 Update base for Update on Inductor cpp wrapper: support LinearUnary+ 74eca977d9b5b59c458d27dffcc0982457028715 Update base for Update on Inductor cpp wrapper: support LinearUnary+ bafde4a79fa8372750af58c85740234d86e1ce08 Update base for Update on Inductor cpp wrapper: support LinearUnary+ b8491db18f75e5188ccfa17954faa5b0e84be754 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ce7574f52d466d043b709bad748941f396d8b7d2 Update base for Update on Inductor cpp wrapper: support LinearUnary+ ea1d5ab619567884455a87f53d7eefb5b60a4c0b Update base for Update on Inductor cpp wrapper: support LinearUnary 2023-03-31T05:29:31.2367036Z NUM_TEST_SHARDS=3 2023-03-31T05:29:31.2367218Z PR_NUMBER=97655 2023-03-31T05:29:31.2367666Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_02ede36d-bac3-44c7-a044-9517fcc8230e 2023-03-31T05:29:31.2367964Z SHLVL=1 2023-03-31T05:29:31.2368149Z GITHUB_REPOSITORY=pytorch/pytorch 2023-03-31T05:29:31.2368558Z SHA1=4af0d9f59b3f9dcba6d4663724824d1665313b5d 2023-03-31T05:29:31.2368904Z GITHUB_EVENT_NAME=pull_request 2023-03-31T05:29:31.2369175Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2023-03-31T05:29:31.2369395Z GITHUB_RUN_NUMBER=98082 2023-03-31T05:29:31.2369708Z GITHUB_WORKFLOW=pull 2023-03-31T05:29:31.2370057Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.8/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2023-03-31T05:29:31.2370503Z GITHUB_WORKFLOW_SHA=48284fffe3d9196187bcf2fdf764bf15956e81e8 2023-03-31T05:29:31.2370875Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T05:29:31.2371168Z GITHUB_TRIGGERING_ACTOR=chunyuan-w 2023-03-31T05:29:31.2371381Z SKIP_SCCACHE_INITIALIZATION=1 2023-03-31T05:29:31.2371585Z _=/usr/bin/env 2023-03-31T05:29:31.2371803Z + echo 'Testing pytorch' 2023-03-31T05:29:31.2371980Z Testing pytorch 2023-03-31T05:29:31.2372183Z + export LANG=C.UTF-8 2023-03-31T05:29:31.2372380Z + LANG=C.UTF-8 2023-03-31T05:29:31.2438012Z + PR_NUMBER=97655 2023-03-31T05:29:31.2438393Z + [[ default == \d\e\f\a\u\l\t ]] 2023-03-31T05:29:31.2438779Z + export CUDA_VISIBLE_DEVICES=0 2023-03-31T05:29:31.2439147Z + CUDA_VISIBLE_DEVICES=0 2023-03-31T05:29:31.2439519Z + export HIP_VISIBLE_DEVICES=0 2023-03-31T05:29:31.2439875Z + HIP_VISIBLE_DEVICES=0 2023-03-31T05:29:31.2440233Z + [[ default == \d\i\s\t\r\i\b\u\t\e\d ]] 2023-03-31T05:29:31.2440552Z + [[ default == \s\l\o\w ]] 2023-03-31T05:29:31.2441160Z + [[ linux-bionic-py3.8-clang9 == *slow-gradcheck* ]] 2023-03-31T05:29:31.2441590Z + [[ linux-bionic-py3.8-clang9 == *cuda* ]] 2023-03-31T05:29:31.2441877Z + [[ linux-bionic-py3.8-clang9 == *rocm* ]] 2023-03-31T05:29:31.2442103Z + [[ default == *crossref* ]] 2023-03-31T05:29:31.2442376Z + [[ linux-bionic-py3.8-clang9 == *rocm* ]] 2023-03-31T05:29:31.2442667Z + [[ linux-bionic-py3.8-clang9 != *-bazel-* ]] 2023-03-31T05:29:31.2442949Z + pip_install --user ninja==1.10.2 2023-03-31T05:29:31.2443460Z + pip install --progress-bar off --user ninja==1.10.2 2023-03-31T05:29:31.6979245Z Collecting ninja==1.10.2 2023-03-31T05:29:31.7139167Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2023-03-31T05:29:32.3556334Z Installing collected packages: ninja 2023-03-31T05:29:32.3640718Z  WARNING: The script ninja is installed in '/var/lib/jenkins/.local/bin' which is not on PATH. 2023-03-31T05:29:32.3641263Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2023-03-31T05:29:32.3695684Z Successfully installed ninja-1.10.2 2023-03-31T05:29:32.4439389Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.8/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2023-03-31T05:29:32.4440255Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.8/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2023-03-31T05:29:32.4441127Z + [[ linux-bionic-py3.8-clang9 == *asan* ]] 2023-03-31T05:29:32.4441589Z + [[ linux-bionic-py3.8-clang9 == *-tsan* ]] 2023-03-31T05:29:32.4442076Z + [[ default == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2023-03-31T05:29:32.4442326Z + [[ default == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2023-03-31T05:29:32.4444162Z + DYNAMO_BENCHMARK_FLAGS=() 2023-03-31T05:29:32.4444504Z + [[ default == *dynamo_eager* ]] 2023-03-31T05:29:32.4444824Z + [[ default == *aot_eager* ]] 2023-03-31T05:29:32.4445077Z + [[ default == *inductor* ]] 2023-03-31T05:29:32.4445285Z + [[ default == *dynamic* ]] 2023-03-31T05:29:32.4445584Z + [[ default == *cpu_accuracy* ]] 2023-03-31T05:29:32.4445878Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2023-03-31T05:29:32.4452814Z + [[ linux-bionic-py3.8-clang9 == *tbb* ]] 2023-03-31T05:29:32.4466069Z + [[ linux-bionic-py3.8-clang9 == *libtorch* ]] 2023-03-31T05:29:32.4471538Z + [[ linux-bionic-py3.8-clang9 == *-bazel-* ]] 2023-03-31T05:29:32.4472014Z + [[ linux-bionic-py3.8-clang9 == *-tsan* ]] 2023-03-31T05:29:32.4472234Z + cd test 2023-03-31T05:29:32.4472520Z + python -c 'import torch; print(torch.__config__.show())' 2023-03-31T05:29:34.2408715Z PyTorch built with: 2023-03-31T05:29:34.2409666Z - GCC 4.2 2023-03-31T05:29:34.2410089Z - C++ Version: 201703 2023-03-31T05:29:34.2410518Z - clang 9.0.0 2023-03-31T05:29:34.2411542Z - Intel(R) oneAPI Math Kernel Library Version 2021.4-Product Build 20210904 for Intel(R) 64 architecture applications 2023-03-31T05:29:34.2412449Z - Intel(R) MKL-DNN v2.7.3 (Git Hash 6dbeffbae1f23cbbeae17adb7b5b13f1f37c080e) 2023-03-31T05:29:34.2413075Z - OpenMP 201107 (a.k.a. OpenMP 3.1) 2023-03-31T05:29:34.2413646Z - LAPACK is enabled (usually provided by MKL) 2023-03-31T05:29:34.2414146Z - NNPACK is enabled 2023-03-31T05:29:34.2414618Z - CPU capability usage: AVX2 2023-03-31T05:29:34.2420055Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, CXX_COMPILER=/opt/cache/bin/clang++, CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DUSE_FBGEMM -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=braced-scalar-init -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wvla-extension -Wno-range-loop-analysis -Wno-pass-failed -Wno-error=pedantic -Wno-error=old-style-cast -Wconstant-conversion -Wno-invalid-partial-specialization -Wno-unused-private-field -Wno-missing-braces -Wunused-lambda-capture -Qunused-arguments -fcolor-diagnostics -fdiagnostics-color=always -Werror -fno-math-errno -fno-trapping-math -Werror=format, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, TORCH_DISABLE_GPU_ASSERTS=ON, TORCH_VERSION=2.1.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, 2023-03-31T05:29:34.2423964Z 2023-03-31T05:29:34.4762956Z + cd test 2023-03-31T05:29:34.4763976Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2023-03-31T05:29:35.5414189Z ATen/Parallel: 2023-03-31T05:29:35.5432282Z at::get_num_threads() : 4 2023-03-31T05:29:35.5432774Z at::get_num_interop_threads() : 4 2023-03-31T05:29:35.5433282Z OpenMP 201107 (a.k.a. OpenMP 3.1) 2023-03-31T05:29:35.5433688Z omp_get_max_threads() : 4 2023-03-31T05:29:35.5434625Z Intel(R) oneAPI Math Kernel Library Version 2021.4-Product Build 20210904 for Intel(R) 64 architecture applications 2023-03-31T05:29:35.5435216Z mkl_get_max_threads() : 4 2023-03-31T05:29:35.5435885Z Intel(R) MKL-DNN v2.7.3 (Git Hash 6dbeffbae1f23cbbeae17adb7b5b13f1f37c080e) 2023-03-31T05:29:35.5436475Z std::thread::hardware_concurrency() : 8 2023-03-31T05:29:35.5436917Z Environment variables: 2023-03-31T05:29:35.5437318Z OMP_NUM_THREADS : [not set] 2023-03-31T05:29:35.5437709Z MKL_NUM_THREADS : [not set] 2023-03-31T05:29:35.5438131Z ATen parallel backend: OpenMP 2023-03-31T05:29:35.5438395Z 2023-03-31T05:29:35.7740530Z + [[ default == *backward* ]] 2023-03-31T05:29:35.7740839Z + [[ default == *xla* ]] 2023-03-31T05:29:35.7741199Z + [[ default == \j\i\t\_\l\e\g\a\c\y ]] 2023-03-31T05:29:35.7741880Z + [[ linux-bionic-py3.8-clang9 == *libtorch* ]] 2023-03-31T05:29:35.7742297Z + [[ default == distributed ]] 2023-03-31T05:29:35.7742648Z + [[ default == deploy ]] 2023-03-31T05:29:35.7743084Z + [[ default == *inductor_distributed* ]] 2023-03-31T05:29:35.7743429Z + [[ default == *huggingface* ]] 2023-03-31T05:29:35.7743756Z + [[ default == *timm* ]] 2023-03-31T05:29:35.7744069Z + [[ default == *torchbench* ]] 2023-03-31T05:29:35.7744390Z + [[ default == *inductor* ]] 2023-03-31T05:29:35.7744672Z + [[ default == *dynamo* ]] 2023-03-31T05:29:35.7744990Z + [[ default == *dynamo* ]] 2023-03-31T05:29:35.7745319Z + [[ 1 == 1 ]] 2023-03-31T05:29:35.7745648Z + [[ 3 -gt 1 ]] 2023-03-31T05:29:35.7745927Z + test_without_numpy 2023-03-31T05:29:35.7746251Z ++ dirname .ci/pytorch/test.sh 2023-03-31T05:29:35.7754163Z + pushd .ci/pytorch 2023-03-31T05:29:35.7754909Z ~/workspace/.ci/pytorch ~/workspace 2023-03-31T05:29:35.7755851Z + python -c 'import sys;sys.path.insert(0, '\''fake_numpy'\'');from unittest import TestCase;import torch;x=torch.randn(3,3);TestCase().assertRaises(RuntimeError, lambda: x.numpy())' 2023-03-31T05:29:36.8098481Z :1: UserWarning: Failed to initialize NumPy: Sorry PyTorch, but our NumPy is in the other folder (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_numpy.cpp:84.) 2023-03-31T05:29:37.0355761Z + python -c 'import sys;sys.path.insert(0, '\''fake_numpy'\'');import torch;print(torch.tensor([torch.tensor(0.), torch.tensor(1.)]))' 2023-03-31T05:29:38.0430885Z :1: UserWarning: Failed to initialize NumPy: Sorry PyTorch, but our NumPy is in the other folder (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/utils/tensor_numpy.cpp:84.) 2023-03-31T05:29:38.0440119Z tensor([0., 1.]) 2023-03-31T05:29:38.2651074Z + popd 2023-03-31T05:29:38.2651389Z ~/workspace 2023-03-31T05:29:38.2651775Z + install_torchvision 2023-03-31T05:29:38.2652037Z + local commit 2023-03-31T05:29:38.2654758Z ++ get_pinned_commit vision 2023-03-31T05:29:38.2655177Z ++ cat .github/ci_commit_pins/vision.txt 2023-03-31T05:29:38.2684145Z + commit=0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:38.2684782Z + pip_install --no-use-pep517 --user git+https://github.com/pytorch/vision.git@0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:38.2685392Z + pip install --progress-bar off --no-use-pep517 --user git+https://github.com/pytorch/vision.git@0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:38.6309807Z Collecting git+https://github.com/pytorch/vision.git@0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:38.6315410Z Cloning https://github.com/pytorch/vision.git (to revision 0387b8821d67ca62d57e3b228ade45371c0af79d) to /tmp/pip-req-build-8ti3rqy6 2023-03-31T05:29:38.6594690Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-8ti3rqy6 2023-03-31T05:29:40.5313805Z Running command git rev-parse -q --verify 'sha^0387b8821d67ca62d57e3b228ade45371c0af79d' 2023-03-31T05:29:40.5357799Z Running command git fetch -q https://github.com/pytorch/vision.git 0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:41.6307983Z Running command git checkout -q 0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:41.9294512Z Resolved https://github.com/pytorch/vision.git to commit 0387b8821d67ca62d57e3b228ade45371c0af79d 2023-03-31T05:29:44.0557205Z Preparing metadata (setup.py) ... [?25l- done 2023-03-31T05:29:44.0624009Z [?25hRequirement already satisfied: numpy in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torchvision==0.16.0a0+0387b88) (1.21.2) 2023-03-31T05:29:44.0627706Z Requirement already satisfied: requests in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torchvision==0.16.0a0+0387b88) (2.28.2) 2023-03-31T05:29:44.0631296Z Requirement already satisfied: torch in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torchvision==0.16.0a0+0387b88) (2.1.0a0+git4af0d9f) 2023-03-31T05:29:44.0637767Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torchvision==0.16.0a0+0387b88) (9.4.0) 2023-03-31T05:29:44.0815915Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from requests->torchvision==0.16.0a0+0387b88) (1.26.15) 2023-03-31T05:29:44.0823310Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from requests->torchvision==0.16.0a0+0387b88) (3.1.0) 2023-03-31T05:29:44.0829502Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from requests->torchvision==0.16.0a0+0387b88) (3.4) 2023-03-31T05:29:44.0835516Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from requests->torchvision==0.16.0a0+0387b88) (2022.12.7) 2023-03-31T05:29:44.0887234Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch->torchvision==0.16.0a0+0387b88) (2.6.3) 2023-03-31T05:29:44.0891236Z Requirement already satisfied: typing-extensions in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch->torchvision==0.16.0a0+0387b88) (4.5.0) 2023-03-31T05:29:44.0894622Z Requirement already satisfied: sympy in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch->torchvision==0.16.0a0+0387b88) (1.11.1) 2023-03-31T05:29:44.0898339Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch->torchvision==0.16.0a0+0387b88) (3.10.7) 2023-03-31T05:29:44.0902098Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from torch->torchvision==0.16.0a0+0387b88) (3.1.2) 2023-03-31T05:29:44.1147639Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from jinja2->torch->torchvision==0.16.0a0+0387b88) (2.1.2) 2023-03-31T05:29:44.1303797Z Requirement already satisfied: mpmath>=0.19 in /opt/conda/envs/py_3.8/lib/python3.8/site-packages (from sympy->torch->torchvision==0.16.0a0+0387b88) (1.3.0) 2023-03-31T05:29:44.1398558Z Building wheels for collected packages: torchvision 2023-03-31T05:31:17.4800323Z Building wheel for torchvision (setup.py) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / done 2023-03-31T05:31:17.5366445Z [?25h Created wheel for torchvision: filename=torchvision-0.16.0a0+0387b88-cp38-cp38-linux_x86_64.whl size=22047994 sha256=3832f1496c4fe82700e8eea1413d57d943068e81006c55b33d58338e3d58bb70 2023-03-31T05:31:17.5367000Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/da/c6/a8/4e25658b6c5b62d2c0e2469af08abadbc782bba1c21b7615a7 2023-03-31T05:31:17.5414395Z Successfully built torchvision 2023-03-31T05:31:18.1255499Z Installing collected packages: torchvision 2023-03-31T05:31:18.8069679Z Successfully installed torchvision-0.16.0a0+0387b88 2023-03-31T05:31:18.9479808Z + test_python_shard 1 2023-03-31T05:31:18.9480396Z + [[ -z 3 ]] 2023-03-31T05:31:18.9480867Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --shard 1 3 --verbose 2023-03-31T05:31:20.6555525Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:20.6594872Z Ignoring disabled issues: [] 2023-03-31T05:31:20.6801711Z Found test time stats from artifacts 2023-03-31T05:31:20.6858545Z Selected tests: 2023-03-31T05:31:20.6858790Z test_decomp 2023-03-31T05:31:20.6858962Z test_ops_jit 2023-03-31T05:31:20.6859156Z test_ops_fwd_gradients 2023-03-31T05:31:20.6861144Z test_fx 2023-03-31T05:31:20.6861431Z test_nn 2023-03-31T05:31:20.6861619Z test_linalg 2023-03-31T05:31:20.6861888Z test_cpp_extensions_jit 2023-03-31T05:31:20.6862330Z nn/test_pooling 2023-03-31T05:31:20.6862656Z test_autograd 2023-03-31T05:31:20.6862964Z test_torch 2023-03-31T05:31:20.6863182Z test_cpp_api_parity 2023-03-31T05:31:20.6863382Z test_mobile_optimizer 2023-03-31T05:31:20.6863602Z test_cpp_extensions_open_device_registration 2023-03-31T05:31:20.6863853Z test_namedtuple_return_api 2023-03-31T05:31:20.6864101Z test_fake_tensor 2023-03-31T05:31:20.6864269Z test_prims 2023-03-31T05:31:20.6864475Z functorch/test_memory_efficient_fusion 2023-03-31T05:31:20.6864691Z test_show_pickle 2023-03-31T05:31:20.6864879Z test_cpp_extensions_aot_ninja 2023-03-31T05:31:20.6866494Z test_jit_cuda_fuser 2023-03-31T05:31:20.6866814Z _nvfuser/test_torchscript 2023-03-31T05:31:20.6867032Z functorch/test_functionalize 2023-03-31T05:31:20.6867260Z inductor/test_kernel_benchmark 2023-03-31T05:31:20.6867516Z inductor/test_perf 2023-03-31T05:31:20.6867858Z inductor/test_triton_wrapper 2023-03-31T05:31:20.6868213Z lazy/test_meta_kernel 2023-03-31T05:31:20.6868523Z test_cuda_nvml_based_avail 2023-03-31T05:31:20.6868827Z test_cuda_trace 2023-03-31T05:31:20.6869312Z test_matmul_cuda 2023-03-31T05:31:20.6869481Z test_quantization 2023-03-31T05:31:20.6869657Z test_jit 2023-03-31T05:31:20.6869824Z test_foreach 2023-03-31T05:31:20.6870747Z dynamo/test_minifier 2023-03-31T05:31:20.6871066Z inductor/test_cpp_wrapper 2023-03-31T05:31:20.6871524Z test_tensorboard 2023-03-31T05:31:20.6871761Z test_transformers 2023-03-31T05:31:20.6871943Z test_stateless 2023-03-31T05:31:20.6872123Z test_ao_sparsity 2023-03-31T05:31:20.6872291Z nn/test_embedding 2023-03-31T05:31:20.6872498Z inductor/test_standalone_compile 2023-03-31T05:31:20.6872700Z test_datapipe 2023-03-31T05:31:20.6872874Z test_functionalization 2023-03-31T05:31:20.6873089Z profiler/test_memory_profiler 2023-03-31T05:31:20.6873284Z test_weak 2023-03-31T05:31:20.6873442Z test_mkl_verbose 2023-03-31T05:31:20.6873630Z dynamo/test_export 2023-03-31T05:31:20.6873824Z dynamo/test_aot_autograd 2023-03-31T05:31:20.6874022Z dynamo/test_logging 2023-03-31T05:31:20.6874323Z test_import_stats 2023-03-31T05:31:20.6874552Z test_mkldnn_verbose 2023-03-31T05:31:20.6874792Z dynamo/test_modules 2023-03-31T05:31:20.6875155Z functorch/test_control_flow 2023-03-31T05:31:20.6875531Z dynamo/test_subgraphs 2023-03-31T05:31:20.6875822Z test_maskedtensor 2023-03-31T05:31:20.6875995Z test_package 2023-03-31T05:31:20.6876193Z functorch/test_vmap_registrations 2023-03-31T05:31:20.6876398Z test_module_init 2023-03-31T05:31:20.6876564Z test_nestedtensor 2023-03-31T05:31:20.6880235Z dynamo/test_optimizers 2023-03-31T05:31:20.6880649Z test_logging 2023-03-31T05:31:20.6880982Z test_scatter_gather_ops 2023-03-31T05:31:20.6881357Z test_public_bindings 2023-03-31T05:31:20.6881663Z dynamo/test_unspec 2023-03-31T05:31:20.6882005Z dynamo/test_functions 2023-03-31T05:31:20.6882307Z test_python_dispatch 2023-03-31T05:31:20.6882483Z test_shape_ops 2023-03-31T05:31:20.6882666Z dynamo/test_backends 2023-03-31T05:31:20.6882855Z test_bundled_inputs 2023-03-31T05:31:20.6883235Z test_futures 2023-03-31T05:31:20.6883421Z test_namedtensor 2023-03-31T05:31:20.6883634Z test_tensorexpr_pybind 2023-03-31T05:31:20.6883812Z test_model_dump 2023-03-31T05:31:20.6884008Z dynamo/test_python_autograd 2023-03-31T05:31:20.6884211Z test_fx_reinplace_pass 2023-03-31T05:31:20.6884393Z test_function_schema 2023-03-31T05:31:20.6884587Z test_fx_passes 2023-03-31T05:31:20.6884767Z dynamo/test_global 2023-03-31T05:31:20.6884940Z export/test_export 2023-03-31T05:31:20.6885128Z dynamo/test_interop 2023-03-31T05:31:20.6885327Z functorch/test_minifier 2023-03-31T05:31:20.6885501Z test_subclass 2023-03-31T05:31:20.6885683Z test_native_functions 2023-03-31T05:31:20.6885868Z test_autocast 2023-03-31T05:31:20.6886030Z dynamo/test_nops 2023-03-31T05:31:20.6886221Z test_functional_optim 2023-03-31T05:31:20.6886409Z nn/test_dropout 2023-03-31T05:31:20.6886579Z lazy/test_debug_util 2023-03-31T05:31:20.6886778Z nn/test_parametrization 2023-03-31T05:31:20.6886990Z dynamo/test_export_mutations 2023-03-31T05:31:20.6887172Z test_pytree 2023-03-31T05:31:20.6887359Z test_per_overload_api 2023-03-31T05:31:20.6887546Z test_native_mha 2023-03-31T05:31:20.6887715Z nn/test_module_hooks 2023-03-31T05:31:20.6887900Z test_dlpack 2023-03-31T05:31:20.6888085Z test_numpy_interop 2023-03-31T05:31:20.6888281Z distributions/test_constraints 2023-03-31T05:31:20.6888483Z nn/test_pruning 2023-03-31T05:31:20.6997316Z Prioritized test from test file changes. 2023-03-31T05:31:20.6997692Z reordering tests for PR: 2023-03-31T05:31:20.6998226Z prioritized: ['test_linalg', 'test_autograd', 'test_torch', 'inductor/test_cpp_wrapper', 'test_nestedtensor', 'nn/test_module_hooks'] 2023-03-31T05:31:20.7002151Z the rest: ['test_decomp', 'test_ops_jit', 'test_ops_fwd_gradients', 'test_fx', 'test_nn', 'test_cpp_extensions_jit', 'nn/test_pooling', 'test_cpp_api_parity', 'test_mobile_optimizer', 'test_cpp_extensions_open_device_registration', 'test_namedtuple_return_api', 'test_fake_tensor', 'test_prims', 'functorch/test_memory_efficient_fusion', 'test_show_pickle', 'test_cpp_extensions_aot_ninja', 'test_jit_cuda_fuser', '_nvfuser/test_torchscript', 'functorch/test_functionalize', 'inductor/test_kernel_benchmark', 'inductor/test_perf', 'inductor/test_triton_wrapper', 'lazy/test_meta_kernel', 'test_cuda_nvml_based_avail', 'test_cuda_trace', 'test_matmul_cuda', 'test_quantization', 'test_jit', 'test_foreach', 'dynamo/test_minifier', 'test_tensorboard', 'test_transformers', 'test_stateless', 'test_ao_sparsity', 'nn/test_embedding', 'inductor/test_standalone_compile', 'test_datapipe', 'test_functionalization', 'profiler/test_memory_profiler', 'test_weak', 'test_mkl_verbose', 'dynamo/test_export', 'dynamo/test_aot_autograd', 'dynamo/test_logging', 'test_import_stats', 'test_mkldnn_verbose', 'dynamo/test_modules', 'functorch/test_control_flow', 'dynamo/test_subgraphs', 'test_maskedtensor', 'test_package', 'functorch/test_vmap_registrations', 'test_module_init', 'dynamo/test_optimizers', 'test_logging', 'test_scatter_gather_ops', 'test_public_bindings', 'dynamo/test_unspec', 'dynamo/test_functions', 'test_python_dispatch', 'test_shape_ops', 'dynamo/test_backends', 'test_bundled_inputs', 'test_futures', 'test_namedtensor', 'test_tensorexpr_pybind', 'test_model_dump', 'dynamo/test_python_autograd', 'test_fx_reinplace_pass', 'test_function_schema', 'test_fx_passes', 'dynamo/test_global', 'export/test_export', 'dynamo/test_interop', 'functorch/test_minifier', 'test_subclass', 'test_native_functions', 'test_autocast', 'dynamo/test_nops', 'test_functional_optim', 'nn/test_dropout', 'lazy/test_debug_util', 'nn/test_parametrization', 'dynamo/test_export_mutations', 'test_pytree', 'test_per_overload_api', 'test_native_mha', 'test_dlpack', 'test_numpy_interop', 'distributions/test_constraints', 'nn/test_pruning'] 2023-03-31T05:31:20.7004787Z 2023-03-31T05:31:20.7005189Z Downloading https://ossci-metrics.s3.amazonaws.com/slow-tests.json to /var/lib/jenkins/workspace/test/.pytorch-slow-tests.json 2023-03-31T05:31:20.7547511Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/workspace/test/.pytorch-disabled-tests.json 2023-03-31T05:31:20.8024050Z parallel (file granularity) tests: 2023-03-31T05:31:20.8024731Z inductor/test_cpp_wrapper 2023-03-31T05:31:20.8024983Z test_nestedtensor 2023-03-31T05:31:20.8025411Z nn/test_module_hooks 2023-03-31T05:31:20.8028851Z functorch/test_functionalize 2023-03-31T05:31:20.8029234Z inductor/test_kernel_benchmark 2023-03-31T05:31:20.8029587Z inductor/test_perf 2023-03-31T05:31:20.8029931Z inductor/test_triton_wrapper 2023-03-31T05:31:20.8030265Z lazy/test_meta_kernel 2023-03-31T05:31:20.8030553Z test_matmul_cuda 2023-03-31T05:31:20.8030929Z test_quantization 2023-03-31T05:31:20.8031252Z test_jit 2023-03-31T05:31:20.8031524Z test_foreach 2023-03-31T05:31:20.8031842Z dynamo/test_minifier 2023-03-31T05:31:20.8032166Z test_tensorboard 2023-03-31T05:31:20.8032468Z test_transformers 2023-03-31T05:31:20.8032785Z test_stateless 2023-03-31T05:31:20.8033089Z test_ao_sparsity 2023-03-31T05:31:20.8033389Z nn/test_embedding 2023-03-31T05:31:20.8033761Z inductor/test_standalone_compile 2023-03-31T05:31:20.8033964Z test_datapipe 2023-03-31T05:31:20.8034141Z test_functionalization 2023-03-31T05:31:20.8034352Z profiler/test_memory_profiler 2023-03-31T05:31:20.8034548Z test_weak 2023-03-31T05:31:20.8034719Z test_mkl_verbose 2023-03-31T05:31:20.8034901Z dynamo/test_export 2023-03-31T05:31:20.8035094Z dynamo/test_aot_autograd 2023-03-31T05:31:20.8035277Z dynamo/test_logging 2023-03-31T05:31:20.8035463Z test_import_stats 2023-03-31T05:31:20.8035684Z test_mkldnn_verbose 2023-03-31T05:31:20.8035869Z dynamo/test_modules 2023-03-31T05:31:20.8036051Z functorch/test_control_flow 2023-03-31T05:31:20.8036253Z dynamo/test_subgraphs 2023-03-31T05:31:20.8036440Z test_maskedtensor 2023-03-31T05:31:20.8036604Z test_package 2023-03-31T05:31:20.8036803Z functorch/test_vmap_registrations 2023-03-31T05:31:20.8037007Z test_module_init 2023-03-31T05:31:20.8037183Z dynamo/test_optimizers 2023-03-31T05:31:20.8037372Z test_logging 2023-03-31T05:31:20.8037782Z test_scatter_gather_ops 2023-03-31T05:31:20.8037968Z test_public_bindings 2023-03-31T05:31:20.8038157Z dynamo/test_unspec 2023-03-31T05:31:20.8038350Z dynamo/test_functions 2023-03-31T05:31:20.8038528Z test_python_dispatch 2023-03-31T05:31:20.8038807Z test_shape_ops 2023-03-31T05:31:20.8038994Z dynamo/test_backends 2023-03-31T05:31:20.8039169Z test_bundled_inputs 2023-03-31T05:31:20.8039351Z test_futures 2023-03-31T05:31:20.8039526Z test_namedtensor 2023-03-31T05:31:20.8039703Z test_tensorexpr_pybind 2023-03-31T05:31:20.8039895Z test_model_dump 2023-03-31T05:31:20.8040089Z dynamo/test_python_autograd 2023-03-31T05:31:20.8040279Z test_fx_reinplace_pass 2023-03-31T05:31:20.8040473Z test_function_schema 2023-03-31T05:31:20.8040659Z test_fx_passes 2023-03-31T05:31:20.8040825Z dynamo/test_global 2023-03-31T05:31:20.8041014Z export/test_export 2023-03-31T05:31:20.8041201Z dynamo/test_interop 2023-03-31T05:31:20.8041380Z functorch/test_minifier 2023-03-31T05:31:20.8041571Z test_subclass 2023-03-31T05:31:20.8041759Z test_native_functions 2023-03-31T05:31:20.8041929Z test_autocast 2023-03-31T05:31:20.8042107Z dynamo/test_nops 2023-03-31T05:31:20.8042294Z test_functional_optim 2023-03-31T05:31:20.8042468Z nn/test_dropout 2023-03-31T05:31:20.8042658Z lazy/test_debug_util 2023-03-31T05:31:20.8042852Z nn/test_parametrization 2023-03-31T05:31:20.8043252Z dynamo/test_export_mutations 2023-03-31T05:31:20.8043446Z test_pytree 2023-03-31T05:31:20.8043637Z test_per_overload_api 2023-03-31T05:31:20.8043821Z test_native_mha 2023-03-31T05:31:20.8043983Z test_dlpack 2023-03-31T05:31:20.8044158Z test_numpy_interop 2023-03-31T05:31:20.8044360Z distributions/test_constraints 2023-03-31T05:31:20.8044545Z nn/test_pruning 2023-03-31T05:31:20.8044743Z serial (file granularity) tests: 2023-03-31T05:31:20.8044938Z test_linalg 2023-03-31T05:31:20.8045093Z test_autograd 2023-03-31T05:31:20.8045261Z test_torch 2023-03-31T05:31:20.8045427Z test_decomp 2023-03-31T05:31:20.8045581Z test_ops_jit 2023-03-31T05:31:20.8045766Z test_ops_fwd_gradients 2023-03-31T05:31:20.8045942Z test_fx 2023-03-31T05:31:20.8046087Z test_nn 2023-03-31T05:31:20.8046264Z test_cpp_extensions_jit 2023-03-31T05:31:20.8046453Z nn/test_pooling 2023-03-31T05:31:20.8046626Z test_cpp_api_parity 2023-03-31T05:31:20.8046816Z test_mobile_optimizer 2023-03-31T05:31:20.8047040Z test_cpp_extensions_open_device_registration 2023-03-31T05:31:20.8047257Z test_namedtuple_return_api 2023-03-31T05:31:20.8047453Z test_fake_tensor 2023-03-31T05:31:20.8047627Z test_prims 2023-03-31T05:31:20.8047814Z functorch/test_memory_efficient_fusion 2023-03-31T05:31:20.8048019Z test_show_pickle 2023-03-31T05:31:20.8048217Z test_cpp_extensions_aot_ninja 2023-03-31T05:31:20.8048402Z test_jit_cuda_fuser 2023-03-31T05:31:20.8048596Z _nvfuser/test_torchscript 2023-03-31T05:31:20.8048797Z test_cuda_nvml_based_avail 2023-03-31T05:31:20.8048972Z test_cuda_trace 2023-03-31T05:31:22.4442091Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:22.4468157Z Ignoring disabled issues: [] 2023-03-31T05:31:22.4619042Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:22.4619897Z Running inductor/test_cpp_wrapper ... [2023-03-31 05:31:22.461796] 2023-03-31T05:31:22.4623275Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'inductor/test_cpp_wrapper.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:22.462099] 2023-03-31T05:31:22.4645256Z Ignoring disabled issues: [] 2023-03-31T05:31:22.4799440Z Running test_nestedtensor ... [2023-03-31 05:31:22.479627] 2023-03-31T05:31:22.4801511Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_nestedtensor.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:22.479929] 2023-03-31T05:31:28.0329824Z 2023-03-31T05:31:28.0330588Z Expand the folded group to see the log file of test_nestedtensor 2023-03-31T05:31:28.0337619Z ##[group]PRINTING LOG FILE of test_nestedtensor (/var/lib/jenkins/workspace/test/test-reports/test_nestedtensor_z2d2ouyq.log) 2023-03-31T05:31:28.0339216Z Test results will be stored in test-reports/python-pytest/test_nestedtensor/test_nestedtensor-fca67e1907f77aa4.xml 2023-03-31T05:31:28.0339904Z ============================= test session starts ============================== 2023-03-31T05:31:28.0340678Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:31:28.0341221Z cachedir: .pytest_cache 2023-03-31T05:31:28.0342188Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:31:28.0342913Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:31:28.0343793Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:31:28.0344430Z collecting ... collected 229 items 2023-03-31T05:31:28.0403485Z Running 229 items in this shard: test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_copy_, test/test_nestedtensor.py::TestNestedTensor::test_default_nested_tensor, test/test_nestedtensor.py::TestNestedTensor::test_dim, test/test_nestedtensor.py::TestNestedTensor::test_fill_, test/test_nestedtensor.py::TestNestedTensor::test_is_contiguous, test/test_nestedtensor.py::TestNestedTensor::test_like_functions_ones_like, test/test_nestedtensor.py::TestNestedTensor::test_like_functions_randn_like, test/test_nestedtensor.py::TestNestedTensor::test_like_functions_zeros_like, test/test_nestedtensor.py::TestNestedTensor::test_nested_namespace, test/test_nestedtensor.py::TestNestedTensor::test_nested_tensor, test/test_nestedtensor.py::TestNestedTensor::test_nested_tensor_matching_dim, test/test_nestedtensor.py::TestNestedTensor::test_numel, test/test_nestedtensor.py::TestNestedTensor::test_repr_string, test/test_nestedtensor.py::TestNestedTensor::test_size, test/test_nestedtensor.py::TestNestedTensor::test_size_dim, test/test_nestedtensor.py::TestNestedTensor::test_stride, test/test_nestedtensor.py::TestNestedTensor::test_to, test/test_nestedtensor.py::TestNestedTensor::test_to_padded_tensor_on_empty_tensor, test/test_nestedtensor.py::TestNestedTensor::test_unbind_0, test/test_nestedtensor.py::TestNestedTensor::test_unbind_1, test/test_nestedtensor.py::TestNestedTensor::test_unbind_3, test/test_nestedtensor.py::TestNestedTensor::test_unbind_4, test/test_nestedtensor.py::TestNestedTensor::test_unbind_dim, test/test_nestedtensor.py::TestNestedTensor::test_zero_, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_abs__cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_abs_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_gelu__cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_gelu_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_neg_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_relu__cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_relu_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_sgn_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_silu__cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_silu_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_tanh__cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_tanh_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cpu_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cpu_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_clone_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_clone_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_contiguous_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_contiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_device_checks_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_embedding_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_layer_norm_breaking_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_layer_norm_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_with_bmm_path_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_with_bmm_path_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_in_place_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_in_place_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_False_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_False_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_True_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_True_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_128_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_128_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_256_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_256_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_384_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_384_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_8_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_8_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_div_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_div_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_in_place_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_in_place_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_False_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_False_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_True_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_True_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sum_dim_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_scaled_dot_product_attention_input_dim_3_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_scaled_dot_product_attention_input_dim_4_cpu, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_output_size_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_output_size_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_simple_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_simple_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_then_from_padded_tensor_no_transform0213_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float64, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_abs_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_accumulate_grad_different_strides_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_as_nested_tensor_propagates_gradients_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_add_strided_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_for_add_op_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_for_sub_op_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_sub_strided_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_dropout_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_gelu_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_indexing_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_128_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_2_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_32_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_4_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_edge_case_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_1023_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_1024_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_128_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_256_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_2_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_32_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_4_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_512_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_513_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_bmm_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_bmm_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_list_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_mask_and_to_padded_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_padded_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_padded_fused_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_generates_leaf_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_plus_transpose_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_matmul_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_matmul_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_reshape_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_reshape_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_softmax_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_squeeze_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_squeeze_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_to_padded_tensor_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_transpose_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_transpose_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_unsqueeze_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_unsqueeze_gradcheck_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_relu_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_selu_backward_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_set_requires_grad_from_list_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_set_requires_grad_from_mask_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_to_buffer_series_ops_grad_with_broadcast_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_unbind_flow_through_cpu, test/test_nestedtensor.py::TestNestedTensorAutogradCPU::test_values_grad_with_broadcast_cpu 2023-03-31T05:31:28.0459190Z 2023-03-31T05:31:28.0459639Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_10 PASSED [ 0%] 2023-03-31T05:31:28.0460460Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_20 PASSED [ 0%] 2023-03-31T05:31:28.0461270Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_10 PASSED [ 1%] 2023-03-31T05:31:28.0462177Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_20 PASSED [ 1%] 2023-03-31T05:31:28.0462970Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10 PASSED [ 2%] 2023-03-31T05:31:28.0463785Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20 PASSED [ 2%] 2023-03-31T05:31:28.0464608Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10 PASSED [ 3%] 2023-03-31T05:31:28.0465422Z test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_20 PASSED [ 3%] 2023-03-31T05:31:28.0466188Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_10 PASSED [ 3%] 2023-03-31T05:31:28.0466985Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_20 PASSED [ 4%] 2023-03-31T05:31:28.0467780Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_10 PASSED [ 4%] 2023-03-31T05:31:28.0468570Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_2_max_seq_len_5_vocab_size_20 PASSED [ 5%] 2023-03-31T05:31:28.0469357Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10 PASSED [ 5%] 2023-03-31T05:31:28.0470404Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20 PASSED [ 6%] 2023-03-31T05:31:28.0471327Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10 PASSED [ 6%] 2023-03-31T05:31:28.0472111Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_20 PASSED [ 6%] 2023-03-31T05:31:28.0472902Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_3_vocab_size_10 PASSED [ 7%] 2023-03-31T05:31:28.0473730Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_3_vocab_size_20 PASSED [ 7%] 2023-03-31T05:31:28.0474547Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_5_vocab_size_10 PASSED [ 8%] 2023-03-31T05:31:28.0475367Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_5_vocab_size_20 PASSED [ 8%] 2023-03-31T05:31:28.0476172Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_10 PASSED [ 9%] 2023-03-31T05:31:28.0477000Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_20 PASSED [ 9%] 2023-03-31T05:31:28.0477822Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_5_vocab_size_10 PASSED [ 10%] 2023-03-31T05:31:28.0478642Z test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_5_vocab_size_20 PASSED [ 10%] 2023-03-31T05:31:28.0479353Z test_nestedtensor.py::TestNestedTensor::test_copy_ PASSED [ 10%] 2023-03-31T05:31:28.0480017Z test_nestedtensor.py::TestNestedTensor::test_default_nested_tensor PASSED [ 11%] 2023-03-31T05:31:28.0480671Z test_nestedtensor.py::TestNestedTensor::test_dim PASSED [ 11%] 2023-03-31T05:31:28.0481293Z test_nestedtensor.py::TestNestedTensor::test_fill_ PASSED [ 12%] 2023-03-31T05:31:28.0481926Z test_nestedtensor.py::TestNestedTensor::test_is_contiguous PASSED [ 12%] 2023-03-31T05:31:28.0482618Z test_nestedtensor.py::TestNestedTensor::test_like_functions_ones_like PASSED [ 13%] 2023-03-31T05:31:28.0483463Z test_nestedtensor.py::TestNestedTensor::test_like_functions_randn_like PASSED [ 13%] 2023-03-31T05:31:28.0484145Z test_nestedtensor.py::TestNestedTensor::test_like_functions_zeros_like PASSED [ 13%] 2023-03-31T05:31:28.0484823Z test_nestedtensor.py::TestNestedTensor::test_nested_namespace PASSED [ 14%] 2023-03-31T05:31:28.0485476Z test_nestedtensor.py::TestNestedTensor::test_nested_tensor PASSED [ 14%] 2023-03-31T05:31:28.0486142Z test_nestedtensor.py::TestNestedTensor::test_nested_tensor_matching_dim PASSED [ 15%] 2023-03-31T05:31:28.0486810Z test_nestedtensor.py::TestNestedTensor::test_numel PASSED [ 15%] 2023-03-31T05:31:28.0487459Z test_nestedtensor.py::TestNestedTensor::test_repr_string PASSED [ 16%] 2023-03-31T05:31:28.0488093Z test_nestedtensor.py::TestNestedTensor::test_size PASSED [ 16%] 2023-03-31T05:31:28.0488699Z test_nestedtensor.py::TestNestedTensor::test_size_dim PASSED [ 17%] 2023-03-31T05:31:28.0489326Z test_nestedtensor.py::TestNestedTensor::test_stride PASSED [ 17%] 2023-03-31T05:31:28.0489936Z test_nestedtensor.py::TestNestedTensor::test_to PASSED [ 17%] 2023-03-31T05:31:28.0490598Z test_nestedtensor.py::TestNestedTensor::test_to_padded_tensor_on_empty_tensor PASSED [ 18%] 2023-03-31T05:31:28.0491276Z test_nestedtensor.py::TestNestedTensor::test_unbind_0 PASSED [ 18%] 2023-03-31T05:31:28.0491903Z test_nestedtensor.py::TestNestedTensor::test_unbind_1 PASSED [ 19%] 2023-03-31T05:31:28.0492531Z test_nestedtensor.py::TestNestedTensor::test_unbind_3 PASSED [ 19%] 2023-03-31T05:31:28.0493144Z test_nestedtensor.py::TestNestedTensor::test_unbind_4 PASSED [ 20%] 2023-03-31T05:31:28.0494021Z test_nestedtensor.py::TestNestedTensor::test_unbind_dim PASSED [ 20%] 2023-03-31T05:31:28.0494784Z test_nestedtensor.py::TestNestedTensor::test_zero_ PASSED [ 20%] 2023-03-31T05:31:28.0495476Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_abs__cpu PASSED [ 21%] 2023-03-31T05:31:28.0496276Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_abs_cpu PASSED [ 21%] 2023-03-31T05:31:28.0497066Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_gelu__cpu PASSED [ 22%] 2023-03-31T05:31:28.0497866Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_gelu_cpu PASSED [ 22%] 2023-03-31T05:31:28.0498642Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_neg_cpu PASSED [ 23%] 2023-03-31T05:31:28.0499433Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_relu__cpu PASSED [ 23%] 2023-03-31T05:31:28.0500255Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_relu_cpu PASSED [ 24%] 2023-03-31T05:31:28.0501024Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_sgn_cpu PASSED [ 24%] 2023-03-31T05:31:28.0501921Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_silu__cpu PASSED [ 24%] 2023-03-31T05:31:28.0502718Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_silu_cpu PASSED [ 25%] 2023-03-31T05:31:28.0503505Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_tanh__cpu PASSED [ 25%] 2023-03-31T05:31:28.0504290Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_activations_tanh_cpu PASSED [ 26%] 2023-03-31T05:31:28.0505070Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cpu_cpu_float32 PASSED [ 26%] 2023-03-31T05:31:28.0505838Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cpu_cpu_float64 PASSED [ 27%] 2023-03-31T05:31:28.0506665Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T05:31:28.0507531Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T05:31:28.0508407Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_cuda_cpu_float64 SKIPPED (Only runs on cuda) [ 28%] 2023-03-31T05:31:28.0509280Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_noncontiguous_cpu_float32 PASSED [ 28%] 2023-03-31T05:31:28.0510096Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_bmm_noncontiguous_cpu_float64 PASSED [ 29%] 2023-03-31T05:31:28.0510900Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_clone_cpu_float16 PASSED [ 29%] 2023-03-31T05:31:28.0511672Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_clone_cpu_float32 PASSED [ 30%] 2023-03-31T05:31:28.0512469Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_contiguous_cpu_float16 PASSED [ 30%] 2023-03-31T05:31:28.0513265Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_contiguous_cpu_float32 PASSED [ 31%] 2023-03-31T05:31:28.0514063Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float16 PASSED [ 31%] 2023-03-31T05:31:28.0514869Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float32 PASSED [ 31%] 2023-03-31T05:31:28.0515658Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_detach_cpu_float64 PASSED [ 32%] 2023-03-31T05:31:28.0516422Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_device_checks_cpu PASSED [ 32%] 2023-03-31T05:31:28.0517205Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_cpu_float32 PASSED [ 33%] 2023-03-31T05:31:28.0517995Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_cpu_float64 PASSED [ 33%] 2023-03-31T05:31:28.0518793Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_noncontiguous_cpu_float32 PASSED [ 34%] 2023-03-31T05:31:28.0519652Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_dropout_noncontiguous_cpu_float64 PASSED [ 34%] 2023-03-31T05:31:28.0520634Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_embedding_cpu PASSED [ 34%] 2023-03-31T05:31:28.0521540Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float16 PASSED [ 35%] 2023-03-31T05:31:28.0522321Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float32 PASSED [ 35%] 2023-03-31T05:31:28.0523264Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_empty_like_cpu_float64 PASSED [ 36%] 2023-03-31T05:31:28.0524089Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_layer_norm_breaking_cpu_float32 PASSED [ 36%] 2023-03-31T05:31:28.0524906Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_layer_norm_cpu_float32 PASSED [ 37%] 2023-03-31T05:31:28.0525689Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_cpu_float32 PASSED [ 37%] 2023-03-31T05:31:28.0526485Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_cpu_float64 PASSED [ 37%] 2023-03-31T05:31:28.0527316Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_noncontiguous_cpu_float32 PASSED [ 38%] 2023-03-31T05:31:28.0528158Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_linear_noncontiguous_cpu_float64 PASSED [ 38%] 2023-03-31T05:31:28.0528975Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_cpu_float32 PASSED [ 39%] 2023-03-31T05:31:28.0529742Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_cpu_float64 PASSED [ 39%] 2023-03-31T05:31:28.0530539Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_noncontiguous_cpu_float32 PASSED [ 40%] 2023-03-31T05:31:28.0531385Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_noncontiguous_cpu_float64 PASSED [ 40%] 2023-03-31T05:31:28.0532209Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_with_bmm_path_cpu_float32 PASSED [ 41%] 2023-03-31T05:31:28.0533053Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_matmul_with_bmm_path_cpu_float64 PASSED [ 41%] 2023-03-31T05:31:28.0533922Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_in_place_cpu_float16 PASSED [ 41%] 2023-03-31T05:31:28.0534783Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_in_place_cpu_float32 PASSED [ 42%] 2023-03-31T05:31:28.0535660Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_False_cpu_float16 PASSED [ 42%] 2023-03-31T05:31:28.0536553Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_False_cpu_float32 PASSED [ 43%] 2023-03-31T05:31:28.0537451Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_True_cpu_float16 PASSED [ 43%] 2023-03-31T05:31:28.0538354Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_add_transpose_True_cpu_float32 PASSED [ 44%] 2023-03-31T05:31:28.0539210Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float16 PASSED [ 44%] 2023-03-31T05:31:28.0540064Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float32 PASSED [ 44%] 2023-03-31T05:31:28.0540915Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_chunk_cpu_float64 PASSED [ 45%] 2023-03-31T05:31:28.0541960Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_128_cpu_float16 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T05:31:28.0543002Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_128_cpu_float32 SKIPPED (Only runs on cuda) [ 46%] 2023-03-31T05:31:28.0544034Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_256_cpu_float16 SKIPPED (Only runs on cuda) [ 46%] 2023-03-31T05:31:28.0545041Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_256_cpu_float32 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:31:28.0546314Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_384_cpu_float16 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:31:28.0547517Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_384_cpu_float32 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:31:28.0548542Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_8_cpu_float16 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:31:28.0549582Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_dense_elementwise_embedding_dim_8_cpu_float32 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:31:28.0550485Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_div_cpu_float16 PASSED [ 49%] 2023-03-31T05:31:28.0551321Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_div_cpu_float32 PASSED [ 49%] 2023-03-31T05:31:28.0552159Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float16 PASSED [ 50%] 2023-03-31T05:31:28.0553026Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float32 PASSED [ 50%] 2023-03-31T05:31:28.0553892Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_cpu_float64 PASSED [ 51%] 2023-03-31T05:31:28.0554789Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float16 PASSED [ 51%] 2023-03-31T05:31:28.0555559Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float32 PASSED [ 51%] 2023-03-31T05:31:28.0556156Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_indexing_noncontiguous_cpu_float64 PASSED [ 52%] 2023-03-31T05:31:28.0556818Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_cpu_float16 PASSED [ 52%] 2023-03-31T05:31:28.0557433Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_cpu_float32 PASSED [ 53%] 2023-03-31T05:31:28.0558106Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_in_place_cpu_float16 PASSED [ 53%] 2023-03-31T05:31:28.0558741Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_mul_in_place_cpu_float32 PASSED [ 54%] 2023-03-31T05:31:28.0559419Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float16 PASSED [ 54%] 2023-03-31T05:31:28.0560069Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float32 PASSED [ 55%] 2023-03-31T05:31:28.0560753Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_split_with_sizes_cpu_float64 PASSED [ 55%] 2023-03-31T05:31:28.0561401Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_False_cpu_float16 PASSED [ 55%] 2023-03-31T05:31:28.0562076Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_False_cpu_float32 PASSED [ 56%] 2023-03-31T05:31:28.0562761Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_True_cpu_float16 PASSED [ 56%] 2023-03-31T05:31:28.0563628Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sub_transpose_True_cpu_float32 PASSED [ 57%] 2023-03-31T05:31:28.0564264Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_nested_tensor_sum_dim_cpu_float32 PASSED [ 57%] 2023-03-31T05:31:28.0564876Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float16 PASSED [ 58%] 2023-03-31T05:31:28.0565457Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float32 PASSED [ 58%] 2023-03-31T05:31:28.0566023Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_reshape_cpu_float64 PASSED [ 58%] 2023-03-31T05:31:28.0566660Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_scaled_dot_product_attention_input_dim_3_cpu PASSED [ 59%] 2023-03-31T05:31:28.0567342Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_scaled_dot_product_attention_input_dim_4_cpu PASSED [ 59%] 2023-03-31T05:31:28.0568184Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_cpu_float32 PASSED [ 60%] 2023-03-31T05:31:28.0568907Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_cpu_float64 PASSED [ 60%] 2023-03-31T05:31:28.0569532Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_noncontiguous_cpu_float32 PASSED [ 61%] 2023-03-31T05:31:28.0570145Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_softmax_noncontiguous_cpu_float64 PASSED [ 61%] 2023-03-31T05:31:28.0570768Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float16 PASSED [ 62%] 2023-03-31T05:31:28.0571355Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float32 PASSED [ 62%] 2023-03-31T05:31:28.0571981Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_squeeze_unsqueeze_cpu_float64 PASSED [ 62%] 2023-03-31T05:31:28.0572592Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float16 PASSED [ 63%] 2023-03-31T05:31:28.0573203Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float32 PASSED [ 63%] 2023-03-31T05:31:28.0573865Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim2_cpu_float64 PASSED [ 64%] 2023-03-31T05:31:28.0574493Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float16 PASSED [ 64%] 2023-03-31T05:31:28.0575124Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float32 PASSED [ 65%] 2023-03-31T05:31:28.0575689Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim3_cpu_float64 PASSED [ 65%] 2023-03-31T05:31:28.0576248Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float16 PASSED [ 65%] 2023-03-31T05:31:28.0576849Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float32 PASSED [ 66%] 2023-03-31T05:31:28.0577526Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_dim4_cpu_float64 PASSED [ 66%] 2023-03-31T05:31:28.0578198Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float16 PASSED [ 67%] 2023-03-31T05:31:28.0578904Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float32 PASSED [ 67%] 2023-03-31T05:31:28.0579575Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_noncontiguous_cpu_float64 PASSED [ 68%] 2023-03-31T05:31:28.0580228Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_output_size_cpu_float16 PASSED [ 68%] 2023-03-31T05:31:28.0580855Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_output_size_cpu_float32 PASSED [ 68%] 2023-03-31T05:31:28.0581493Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_simple_cpu_float16 PASSED [ 69%] 2023-03-31T05:31:28.0582256Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_padded_tensor_simple_cpu_float32 PASSED [ 69%] 2023-03-31T05:31:28.0582930Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_to_then_from_padded_tensor_no_transform0213_cpu_float32 PASSED [ 70%] 2023-03-31T05:31:28.0583575Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float16 PASSED [ 70%] 2023-03-31T05:31:28.0584166Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float32 PASSED [ 71%] 2023-03-31T05:31:28.0584768Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_cpu_float64 PASSED [ 71%] 2023-03-31T05:31:28.0585357Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float16 PASSED [ 72%] 2023-03-31T05:31:28.0585985Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float32 PASSED [ 72%] 2023-03-31T05:31:28.0586665Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_transpose_inference_mode_interaction_cpu_float64 PASSED [ 72%] 2023-03-31T05:31:28.0587555Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float16 PASSED [ 73%] 2023-03-31T05:31:28.0588333Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float32 PASSED [ 73%] 2023-03-31T05:31:28.0589002Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_unbind_noncontiguous_cpu_float64 PASSED [ 74%] 2023-03-31T05:31:28.0589603Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float16 PASSED [ 74%] 2023-03-31T05:31:28.0590190Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float32 PASSED [ 75%] 2023-03-31T05:31:28.0590771Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_cpu_float64 PASSED [ 75%] 2023-03-31T05:31:28.0591392Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float16 PASSED [ 75%] 2023-03-31T05:31:28.0592072Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float32 PASSED [ 76%] 2023-03-31T05:31:28.0592726Z test_nestedtensor.py::TestNestedTensorDeviceTypeCPU::test_view_inference_mode_interaction_cpu_float64 PASSED [ 76%] 2023-03-31T05:31:28.0593369Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_abs_backward_cpu PASSED [ 77%] 2023-03-31T05:31:28.0593975Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_accumulate_grad_different_strides_cpu PASSED [ 77%] 2023-03-31T05:31:28.0594662Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_as_nested_tensor_propagates_gradients_cpu PASSED [ 78%] 2023-03-31T05:31:28.0595322Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_add_strided_cpu PASSED [ 78%] 2023-03-31T05:31:28.0595906Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_for_add_op_cpu PASSED [ 79%] 2023-03-31T05:31:28.0596485Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_for_sub_op_cpu PASSED [ 79%] 2023-03-31T05:31:28.0597069Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_backward_sub_strided_cpu PASSED [ 79%] 2023-03-31T05:31:28.0597637Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_dropout_backward_cpu PASSED [ 80%] 2023-03-31T05:31:28.0598193Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_gelu_backward_cpu PASSED [ 80%] 2023-03-31T05:31:28.0598747Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_indexing_backward_cpu PASSED [ 81%] 2023-03-31T05:31:28.0599333Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_128_cpu PASSED [ 81%] 2023-03-31T05:31:28.0599938Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_2_cpu PASSED [ 82%] 2023-03-31T05:31:28.0600590Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_32_cpu PASSED [ 82%] 2023-03-31T05:31:28.0601192Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_5d_size_4_cpu PASSED [ 82%] 2023-03-31T05:31:28.0601806Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_edge_case_cpu PASSED [ 83%] 2023-03-31T05:31:28.0602441Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_1023_cpu PASSED [ 83%] 2023-03-31T05:31:28.0603239Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_1024_cpu PASSED [ 84%] 2023-03-31T05:31:28.0603785Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_128_cpu PASSED [ 84%] 2023-03-31T05:31:28.0604195Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_256_cpu PASSED [ 85%] 2023-03-31T05:31:28.0604607Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_2_cpu PASSED [ 85%] 2023-03-31T05:31:28.0605025Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_32_cpu PASSED [ 86%] 2023-03-31T05:31:28.0605439Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_4_cpu PASSED [ 86%] 2023-03-31T05:31:28.0606007Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_512_cpu PASSED [ 86%] 2023-03-31T05:31:28.0606500Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_layer_norm_backward_size_513_cpu PASSED [ 87%] 2023-03-31T05:31:28.0606911Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_bmm_backward_cpu PASSED [ 87%] 2023-03-31T05:31:28.0607314Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_bmm_gradcheck_cpu PASSED [ 88%] 2023-03-31T05:31:28.0607726Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_list_cpu PASSED [ 88%] 2023-03-31T05:31:28.0608151Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_mask_and_to_padded_cpu PASSED [ 89%] 2023-03-31T05:31:28.0608577Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_padded_cpu PASSED [ 89%] 2023-03-31T05:31:28.0608989Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_from_padded_fused_cpu PASSED [ 89%] 2023-03-31T05:31:28.0609416Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_generates_leaf_cpu PASSED [ 90%] 2023-03-31T05:31:28.0609841Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_backward_cpu PASSED [ 90%] 2023-03-31T05:31:28.0610248Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_cpu PASSED [ 91%] 2023-03-31T05:31:28.0610656Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_linear_plus_transpose_cpu PASSED [ 91%] 2023-03-31T05:31:28.0611085Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_matmul_backward_cpu PASSED [ 92%] 2023-03-31T05:31:28.0611509Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_matmul_gradcheck_cpu PASSED [ 92%] 2023-03-31T05:31:28.0631748Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_reshape_backward_cpu PASSED [ 93%] 2023-03-31T05:31:28.0632361Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_reshape_gradcheck_cpu PASSED [ 93%] 2023-03-31T05:31:28.0632956Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_softmax_cpu PASSED [ 93%] 2023-03-31T05:31:28.0633595Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_squeeze_backward_cpu PASSED [ 94%] 2023-03-31T05:31:28.0634166Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_squeeze_gradcheck_cpu PASSED [ 94%] 2023-03-31T05:31:28.0634776Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_to_padded_tensor_cpu PASSED [ 95%] 2023-03-31T05:31:28.0635382Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_transpose_backward_cpu PASSED [ 95%] 2023-03-31T05:31:28.0636051Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_transpose_gradcheck_cpu PASSED [ 96%] 2023-03-31T05:31:28.0636747Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_unsqueeze_backward_cpu PASSED [ 96%] 2023-03-31T05:31:28.0637481Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_nested_tensor_unsqueeze_gradcheck_cpu PASSED [ 96%] 2023-03-31T05:31:28.0638110Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_relu_backward_cpu PASSED [ 97%] 2023-03-31T05:31:28.0638768Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_selu_backward_cpu PASSED [ 97%] 2023-03-31T05:31:28.0639457Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_set_requires_grad_from_list_cpu PASSED [ 98%] 2023-03-31T05:31:28.0640161Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_set_requires_grad_from_mask_cpu PASSED [ 98%] 2023-03-31T05:31:28.0640931Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_to_buffer_series_ops_grad_with_broadcast_cpu PASSED [ 99%] 2023-03-31T05:31:28.0641608Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_unbind_flow_through_cpu PASSED [ 99%] 2023-03-31T05:31:28.0642268Z test_nestedtensor.py::TestNestedTensorAutogradCPU::test_values_grad_with_broadcast_cpu PASSED [100%] 2023-03-31T05:31:28.0642859Z 2023-03-31T05:31:28.0643785Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-fca67e1907f77aa4.xml - 2023-03-31T05:31:28.0644636Z ======================= 218 passed, 11 skipped in 2.47s ======================== 2023-03-31T05:31:28.0645166Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:31:28.0645530Z 2023-03-31T05:31:28.0645993Z ##[endgroup] 2023-03-31T05:31:28.0646586Z FINISHED PRINTING LOG FILE of test_nestedtensor (/var/lib/jenkins/workspace/test/test-reports/test_nestedtensor_z2d2ouyq.log) 2023-03-31T05:31:28.0646825Z 2023-03-31T05:31:30.0425908Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:30.0451259Z Ignoring disabled issues: [] 2023-03-31T05:31:30.0604231Z Running nn/test_module_hooks ... [2023-03-31 05:31:30.060157] 2023-03-31T05:31:30.0607313Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_module_hooks.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:30.060509] 2023-03-31T05:31:32.3362213Z 2023-03-31T05:31:32.3362795Z Expand the folded group to see the log file of nn/test_module_hooks 2023-03-31T05:31:32.3364139Z ##[group]PRINTING LOG FILE of nn/test_module_hooks (/var/lib/jenkins/workspace/test/test-reports/nn-test_module_hooks_o09de0fm.log) 2023-03-31T05:31:32.3365304Z Test results will be stored in test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-cf9ef0ba863acb32.xml 2023-03-31T05:31:32.3365938Z ============================= test session starts ============================== 2023-03-31T05:31:32.3366620Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:31:32.3367080Z cachedir: .pytest_cache 2023-03-31T05:31:32.3367811Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:31:32.3368452Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:31:32.3369250Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:31:32.3369774Z collecting ... collected 37 items 2023-03-31T05:31:32.3376163Z Running 37 items in this shard: test/nn/test_module_hooks.py::TestModuleHooks::test_bw_hook_warning_for_non_tensor_or_tuple, test/nn/test_module_hooks.py::TestModuleHooks::test_forward_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_forward_pre_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_full_backward_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_full_backward_pre_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_kwarg_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_mixed_hooks, test/nn/test_module_hooks.py::TestModuleHooks::test_remove_kwarg_hooks, test/nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_module_pre_hook, test/nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_post_hook, test/nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_post_hook_backward_compatibility, test/nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_pre_hook, test/nn/test_module_hooks.py::TestStateDictHooks::test_no_extra_ref_to_module, test/nn/test_module_hooks.py::TestStateDictHooks::test_pickled_hook, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_global_and_local_hooks_order, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_backward_global_hook_writeable, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_forward_forward_hook_removable, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_forward_preforward_hook_removable, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_forward_preforward_hook_writeable, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_hook_invalid_outputs, test/nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_hooks, test/nn/test_module_hooks.py::TestModuleHookNN::test_backward_hooks_interaction, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_backward_size, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_backward_writeable, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_buffer_registration, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_cpp, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_extra_input, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_forward_preforward_writable, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_inplace, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_invalid_outputs, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_last_arg_requires_grad, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_no_requires_grad, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_non_full_warning, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_parameter_registration, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_requires_grad, test/nn/test_module_hooks.py::TestModuleHookNN::test_hook_submodule_registration, test/nn/test_module_hooks.py::TestModuleHookNN::test_hooks 2023-03-31T05:31:32.3382368Z 2023-03-31T05:31:32.3382729Z nn/test_module_hooks.py::TestModuleHooks::test_bw_hook_warning_for_non_tensor_or_tuple PASSED [ 2%] 2023-03-31T05:31:32.3383359Z nn/test_module_hooks.py::TestModuleHooks::test_forward_hooks PASSED [ 5%] 2023-03-31T05:31:32.3383926Z nn/test_module_hooks.py::TestModuleHooks::test_forward_pre_hooks PASSED [ 8%] 2023-03-31T05:31:32.3384499Z nn/test_module_hooks.py::TestModuleHooks::test_full_backward_hooks PASSED [ 10%] 2023-03-31T05:31:32.3385102Z nn/test_module_hooks.py::TestModuleHooks::test_full_backward_pre_hooks PASSED [ 13%] 2023-03-31T05:31:32.3385708Z nn/test_module_hooks.py::TestModuleHooks::test_kwarg_hooks PASSED [ 16%] 2023-03-31T05:31:32.3386245Z nn/test_module_hooks.py::TestModuleHooks::test_mixed_hooks PASSED [ 18%] 2023-03-31T05:31:32.3386828Z nn/test_module_hooks.py::TestModuleHooks::test_remove_kwarg_hooks PASSED [ 21%] 2023-03-31T05:31:32.3387462Z nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_module_pre_hook PASSED [ 24%] 2023-03-31T05:31:32.3388121Z nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_post_hook PASSED [ 27%] 2023-03-31T05:31:32.3388801Z nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_post_hook_backward_compatibility PASSED [ 29%] 2023-03-31T05:31:32.3389472Z nn/test_module_hooks.py::TestStateDictHooks::test_load_state_dict_pre_hook PASSED [ 32%] 2023-03-31T05:31:32.3390071Z nn/test_module_hooks.py::TestStateDictHooks::test_no_extra_ref_to_module PASSED [ 35%] 2023-03-31T05:31:32.3390680Z nn/test_module_hooks.py::TestStateDictHooks::test_pickled_hook PASSED [ 37%] 2023-03-31T05:31:32.3391273Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_global_and_local_hooks_order PASSED [ 40%] 2023-03-31T05:31:32.3391987Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_backward_global_hook_writeable PASSED [ 43%] 2023-03-31T05:31:32.3392732Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_forward_forward_hook_removable PASSED [ 45%] 2023-03-31T05:31:32.3393445Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_forward_preforward_hook_removable PASSED [ 48%] 2023-03-31T05:31:32.3394178Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_forward_preforward_hook_writeable PASSED [ 51%] 2023-03-31T05:31:32.3394903Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_hook_invalid_outputs PASSED [ 54%] 2023-03-31T05:31:32.3395573Z nn/test_module_hooks.py::TestModuleGlobalHooks::test_module_global_hooks PASSED [ 56%] 2023-03-31T05:31:32.3396202Z nn/test_module_hooks.py::TestModuleHookNN::test_backward_hooks_interaction PASSED [ 59%] 2023-03-31T05:31:32.3396807Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_backward_size PASSED [ 62%] 2023-03-31T05:31:32.3397632Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_backward_writeable PASSED [ 64%] 2023-03-31T05:31:32.3398209Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_buffer_registration PASSED [ 67%] 2023-03-31T05:31:32.3398898Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_cpp PASSED [ 70%] 2023-03-31T05:31:32.3399443Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_extra_input PASSED [ 72%] 2023-03-31T05:31:32.3400084Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_forward_preforward_writable PASSED [ 75%] 2023-03-31T05:31:32.3400658Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_inplace PASSED [ 78%] 2023-03-31T05:31:32.3401266Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_invalid_outputs PASSED [ 81%] 2023-03-31T05:31:32.3401848Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_last_arg_requires_grad PASSED [ 83%] 2023-03-31T05:31:32.3402456Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_no_requires_grad PASSED [ 86%] 2023-03-31T05:31:32.3403204Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_non_full_warning PASSED [ 89%] 2023-03-31T05:31:32.3403820Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_parameter_registration PASSED [ 91%] 2023-03-31T05:31:32.3404419Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_requires_grad PASSED [ 94%] 2023-03-31T05:31:32.3405000Z nn/test_module_hooks.py::TestModuleHookNN::test_hook_submodule_registration PASSED [ 97%] 2023-03-31T05:31:32.3405607Z nn/test_module_hooks.py::TestModuleHookNN::test_hooks PASSED [100%] 2023-03-31T05:31:32.3405901Z 2023-03-31T05:31:32.3406621Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-cf9ef0ba863acb32.xml - 2023-03-31T05:31:32.3407279Z ============================== 37 passed in 0.18s ============================== 2023-03-31T05:31:32.3407813Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:31:32.3408160Z 2023-03-31T05:31:32.3408600Z ##[endgroup] 2023-03-31T05:31:32.3409356Z FINISHED PRINTING LOG FILE of nn/test_module_hooks (/var/lib/jenkins/workspace/test/test-reports/nn-test_module_hooks_o09de0fm.log) 2023-03-31T05:31:32.3409818Z 2023-03-31T05:31:34.3494058Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:34.3521271Z Ignoring disabled issues: [] 2023-03-31T05:31:34.3678987Z Running functorch/test_functionalize ... [2023-03-31 05:31:34.367499] 2023-03-31T05:31:34.3680992Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'functorch/test_functionalize.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:34.367841] 2023-03-31T05:31:38.5303351Z 2023-03-31T05:31:38.5303932Z Expand the folded group to see the log file of functorch/test_functionalize 2023-03-31T05:31:38.5305046Z ##[group]PRINTING LOG FILE of functorch/test_functionalize (/var/lib/jenkins/workspace/test/test-reports/functorch-test_functionalize_xd19c82b.log) 2023-03-31T05:31:38.5305936Z Test results will be stored in test-reports/python-pytest/functorch.test_functionalize/functorch.test_functionalize-93e9c96bd792f149.xml 2023-03-31T05:31:38.5306341Z ============================= test session starts ============================== 2023-03-31T05:31:38.5306823Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:31:38.5307099Z cachedir: .pytest_cache 2023-03-31T05:31:38.5307594Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:31:38.5307970Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:31:38.5308495Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:31:38.5308842Z collecting ... collected 0 items 2023-03-31T05:31:38.5309480Z Running 0 items in this shard: 2023-03-31T05:31:38.5309734Z 2023-03-31T05:31:38.5310396Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_functionalize/functorch.test_functionalize-93e9c96bd792f149.xml - 2023-03-31T05:31:38.5310927Z ============================ no tests ran in 0.01s ============================= 2023-03-31T05:31:38.5311253Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:31:38.5311443Z 2023-03-31T05:31:38.5311701Z ##[endgroup] 2023-03-31T05:31:38.5312171Z FINISHED PRINTING LOG FILE of functorch/test_functionalize (/var/lib/jenkins/workspace/test/test-reports/functorch-test_functionalize_xd19c82b.log) 2023-03-31T05:31:38.5312435Z 2023-03-31T05:31:40.6598042Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:40.6626696Z Ignoring disabled issues: [] 2023-03-31T05:31:40.6787094Z Running inductor/test_kernel_benchmark ... [2023-03-31 05:31:40.678305] 2023-03-31T05:31:40.6789202Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'inductor/test_kernel_benchmark.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:40.678689] 2023-03-31T05:31:47.3468100Z 2023-03-31T05:31:47.3468681Z Expand the folded group to see the log file of inductor/test_kernel_benchmark 2023-03-31T05:31:47.3469906Z ##[group]PRINTING LOG FILE of inductor/test_kernel_benchmark (/var/lib/jenkins/workspace/test/test-reports/inductor-test_kernel_benchmark_vyifgg2y.log) 2023-03-31T05:31:47.3470873Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:47.3471054Z 2023-03-31T05:31:47.3471288Z ##[endgroup] 2023-03-31T05:31:47.3471753Z FINISHED PRINTING LOG FILE of inductor/test_kernel_benchmark (/var/lib/jenkins/workspace/test/test-reports/inductor-test_kernel_benchmark_vyifgg2y.log) 2023-03-31T05:31:47.3472021Z 2023-03-31T05:31:49.4503508Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:49.4531587Z Ignoring disabled issues: [] 2023-03-31T05:31:49.4690303Z Running inductor/test_perf ... [2023-03-31 05:31:49.468721] 2023-03-31T05:31:49.4692723Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'inductor/test_perf.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:49.469032] 2023-03-31T05:31:56.3358784Z 2023-03-31T05:31:56.3359370Z Expand the folded group to see the log file of inductor/test_perf 2023-03-31T05:31:56.3360281Z ##[group]PRINTING LOG FILE of inductor/test_perf (/var/lib/jenkins/workspace/test/test-reports/inductor-test_perf_i2c6xq48.log) 2023-03-31T05:31:56.3360823Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:56.3360985Z 2023-03-31T05:31:56.3361213Z ##[endgroup] 2023-03-31T05:31:56.3361638Z FINISHED PRINTING LOG FILE of inductor/test_perf (/var/lib/jenkins/workspace/test/test-reports/inductor-test_perf_i2c6xq48.log) 2023-03-31T05:31:56.3361903Z 2023-03-31T05:31:58.4976182Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:31:58.5004508Z Ignoring disabled issues: [] 2023-03-31T05:31:58.5164588Z Running inductor/test_triton_wrapper ... [2023-03-31 05:31:58.516070] 2023-03-31T05:31:58.5166422Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'inductor/test_triton_wrapper.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:31:58.516407] 2023-03-31T05:32:05.1454482Z 2023-03-31T05:32:05.1455064Z Expand the folded group to see the log file of inductor/test_triton_wrapper 2023-03-31T05:32:05.1456139Z ##[group]PRINTING LOG FILE of inductor/test_triton_wrapper (/var/lib/jenkins/workspace/test/test-reports/inductor-test_triton_wrapper_mltnltdg.log) 2023-03-31T05:32:05.1456752Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:32:05.1457246Z 2023-03-31T05:32:05.1457614Z ##[endgroup] 2023-03-31T05:32:05.1458118Z FINISHED PRINTING LOG FILE of inductor/test_triton_wrapper (/var/lib/jenkins/workspace/test/test-reports/inductor-test_triton_wrapper_mltnltdg.log) 2023-03-31T05:32:05.1458384Z 2023-03-31T05:32:07.2609491Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:32:07.2638211Z Ignoring disabled issues: [] 2023-03-31T05:32:07.2796789Z Running lazy/test_meta_kernel ... [2023-03-31 05:32:07.279321] 2023-03-31T05:32:07.2799337Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'lazy/test_meta_kernel.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:32:07.279685] 2023-03-31T05:32:08.9420940Z 2023-03-31T05:32:08.9421593Z Expand the folded group to see the log file of lazy/test_meta_kernel 2023-03-31T05:32:08.9422599Z ##[group]PRINTING LOG FILE of lazy/test_meta_kernel (/var/lib/jenkins/workspace/test/test-reports/lazy-test_meta_kernel_4yeljjve.log) 2023-03-31T05:32:08.9422894Z 2023-03-31T05:32:08.9423126Z ##[endgroup] 2023-03-31T05:32:08.9423652Z FINISHED PRINTING LOG FILE of lazy/test_meta_kernel (/var/lib/jenkins/workspace/test/test-reports/lazy-test_meta_kernel_4yeljjve.log) 2023-03-31T05:32:08.9423905Z 2023-03-31T05:32:11.0573243Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:32:11.0601886Z Ignoring disabled issues: [] 2023-03-31T05:32:11.0760356Z Running test_matmul_cuda ... [2023-03-31 05:32:11.075626] 2023-03-31T05:32:11.0761332Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_matmul_cuda.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:32:11.075924] 2023-03-31T05:32:13.2159994Z 2023-03-31T05:32:13.2160558Z Expand the folded group to see the log file of test_matmul_cuda 2023-03-31T05:32:13.2161671Z ##[group]PRINTING LOG FILE of test_matmul_cuda (/var/lib/jenkins/workspace/test/test-reports/test_matmul_cuda_0rhh271z.log) 2023-03-31T05:32:13.2162484Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7f2fa785972ff8f.xml 2023-03-31T05:32:13.2162849Z ============================= test session starts ============================== 2023-03-31T05:32:13.2163412Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:32:13.2163681Z cachedir: .pytest_cache 2023-03-31T05:32:13.2164119Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:32:13.2164496Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:32:13.2164935Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:32:13.2165256Z collecting ... collected 0 items 2023-03-31T05:32:13.2165476Z Running 0 items in this shard: 2023-03-31T05:32:13.2165608Z 2023-03-31T05:32:13.2165962Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7f2fa785972ff8f.xml - 2023-03-31T05:32:13.2166320Z ============================ no tests ran in 0.01s ============================= 2023-03-31T05:32:13.2166646Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:32:13.2166861Z 2023-03-31T05:32:13.2167114Z ##[endgroup] 2023-03-31T05:32:13.2167512Z FINISHED PRINTING LOG FILE of test_matmul_cuda (/var/lib/jenkins/workspace/test/test-reports/test_matmul_cuda_0rhh271z.log) 2023-03-31T05:32:13.2167745Z 2023-03-31T05:32:15.3194349Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:32:15.3221786Z Ignoring disabled issues: [] 2023-03-31T05:32:15.3382204Z Running test_quantization ... [2023-03-31 05:32:15.337817] 2023-03-31T05:32:15.3383503Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_quantization.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:32:15.338119] 2023-03-31T05:48:33.5530986Z 2023-03-31T05:48:33.5531492Z Expand the folded group to see the log file of test_quantization 2023-03-31T05:48:33.5534363Z ##[group]PRINTING LOG FILE of test_quantization (/var/lib/jenkins/workspace/test/test-reports/test_quantization_zpcgt5ju.log) 2023-03-31T05:48:33.5537523Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:48:33.5538581Z Test results will be stored in test-reports/python-pytest/test_quantization/test_quantization-816f8ae0c160c36a.xml 2023-03-31T05:48:33.5539516Z ============================= test session starts ============================== 2023-03-31T05:48:33.5540307Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:48:33.5540788Z cachedir: .pytest_cache 2023-03-31T05:48:33.5541605Z hypothesis profile 'pytorch_ci' -> database=None, deadline=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:48:33.5607581Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:48:33.5608573Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:48:33.5609134Z collecting ... collected 1012 items 2023-03-31T05:48:33.5794611Z Running 1012 items in this shard: test/test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool, test/test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool2d_nhwc, test/test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool3d_ndhwc, test/test_quantization.py::TestQuantizedOps::test_add_scalar_relu, test/test_quantization.py::TestQuantizedOps::test_advanced_indexing, test/test_quantization.py::TestQuantizedOps::test_avg_pool2d, test/test_quantization.py::TestQuantizedOps::test_avg_pool2d_nhwc, test/test_quantization.py::TestQuantizedOps::test_avg_pool3d, test/test_quantization.py::TestQuantizedOps::test_avg_pool3d_nhwc, test/test_quantization.py::TestQuantizedOps::test_batch_norm, test/test_quantization.py::TestQuantizedOps::test_batch_norm_relu, test/test_quantization.py::TestQuantizedOps::test_cat, test/test_quantization.py::TestQuantizedOps::test_cat_nhwc, test/test_quantization.py::TestQuantizedOps::test_channel_shuffle, test/test_quantization.py::TestQuantizedOps::test_custom_module_lstm, test/test_quantization.py::TestQuantizedOps::test_custom_module_multi_head_attention, test/test_quantization.py::TestQuantizedOps::test_empty_batch, test/test_quantization.py::TestQuantizedOps::test_equal, test/test_quantization.py::TestQuantizedOps::test_group_norm, test/test_quantization.py::TestQuantizedOps::test_hardswish, test/test_quantization.py::TestQuantizedOps::test_hardtanh, test/test_quantization.py::TestQuantizedOps::test_instance_norm, test/test_quantization.py::TestQuantizedOps::test_interpolate, test/test_quantization.py::TestQuantizedOps::test_interpolate3d, test/test_quantization.py::TestQuantizedOps::test_leaky_relu, test/test_quantization.py::TestQuantizedOps::test_leaky_relu_observed_output, test/test_quantization.py::TestQuantizedOps::test_linear_bias_unpack, test/test_quantization.py::TestQuantizedOps::test_max_pool1d, test/test_quantization.py::TestQuantizedOps::test_max_pool2d, test/test_quantization.py::TestQuantizedOps::test_max_pool2d_cudnn, test/test_quantization.py::TestQuantizedOps::test_max_pool2d_nhwc, test/test_quantization.py::TestQuantizedOps::test_mean, test/test_quantization.py::TestQuantizedOps::test_mul_scalar_relu, test/test_quantization.py::TestQuantizedOps::test_qadd_broadcast, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_cudnn, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_cudnn_nhwc, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_different_qparams, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_same_qparams, test/test_quantization.py::TestQuantizedOps::test_qcelu, test/test_quantization.py::TestQuantizedOps::test_qclamp, test/test_quantization.py::TestQuantizedOps::test_qelu, test/test_quantization.py::TestQuantizedOps::test_qgelu, test/test_quantization.py::TestQuantizedOps::test_qhardsigmoid, test/test_quantization.py::TestQuantizedOps::test_qlayer_norm, test/test_quantization.py::TestQuantizedOps::test_qmatmul, test/test_quantization.py::TestQuantizedOps::test_qmul_broadcast, test/test_quantization.py::TestQuantizedOps::test_qmul_relu_different_qparams, test/test_quantization.py::TestQuantizedOps::test_qmul_relu_same_qparams, test/test_quantization.py::TestQuantizedOps::test_qprelu, test/test_quantization.py::TestQuantizedOps::test_qrelu, test/test_quantization.py::TestQuantizedOps::test_qrelu6, test/test_quantization.py::TestQuantizedOps::test_qsoftmax, test/test_quantization.py::TestQuantizedOps::test_qsoftmax_qnnpack, test/test_quantization.py::TestQuantizedOps::test_qtanh, test/test_quantization.py::TestQuantizedOps::test_qthreshold, test/test_quantization.py::TestQuantizedOps::test_qtopk, test/test_quantization.py::TestQuantizedOps::test_quantized_equal, test/test_quantization.py::TestQuantizedOps::test_quantized_mean_qnnpack, test/test_quantization.py::TestQuantizedOps::test_sigmoid, test/test_quantization.py::TestQuantizedOps::test_sigmoid_non_observed, test/test_quantization.py::TestQuantizedOps::test_std, test/test_quantization.py::TestQNNPackOps::test_adaptive_avg_pool2d, test/test_quantization.py::TestQNNPackOps::test_avg_pool2d, test/test_quantization.py::TestQNNPackOps::test_hardtanh, test/test_quantization.py::TestQNNPackOps::test_mean, test/test_quantization.py::TestQNNPackOps::test_qnnpack_add, test/test_quantization.py::TestQNNPackOps::test_qnnpack_add_broadcast, test/test_quantization.py::TestQNNPackOps::test_qnnpack_maxpool2d, test/test_quantization.py::TestQNNPackOps::test_qnnpack_mul, test/test_quantization.py::TestQNNPackOps::test_qnnpack_relu, test/test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid, test/test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid_sweep, test/test_quantization.py::TestQNNPackOps::test_qnnpack_tanh, test/test_quantization.py::TestQuantizedLinear::test_qlinear, test/test_quantization.py::TestQuantizedLinear::test_qlinear_cudnn, test/test_quantization.py::TestQuantizedLinear::test_qlinear_leaky_relu, test/test_quantization.py::TestQuantizedLinear::test_qlinear_qnnpack_free_memory_and_unpack, test/test_quantization.py::TestQuantizedLinear::test_qlinear_relu, test/test_quantization.py::TestQuantizedLinear::test_qlinear_tanh, test/test_quantization.py::TestQuantizedLinear::test_qlinear_unpack, test/test_quantization.py::TestQuantizedLinear::test_qlinear_with_input_q_dq_qweight_dq_output_fp32, test/test_quantization.py::TestQuantizedConv::test_benchmark, test/test_quantization.py::TestQuantizedConv::test_conv_reorder_issue_onednn, test/test_quantization.py::TestQuantizedConv::test_conv_transpose_reorder_issue_onednn, test/test_quantization.py::TestQuantizedConv::test_qconv1d, test/test_quantization.py::TestQuantizedConv::test_qconv1d_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv1d_relu, test/test_quantization.py::TestQuantizedConv::test_qconv1d_relu_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv1d_unpack, test/test_quantization.py::TestQuantizedConv::test_qconv2d, test/test_quantization.py::TestQuantizedConv::test_qconv2d_add, test/test_quantization.py::TestQuantizedConv::test_qconv2d_add_relu, test/test_quantization.py::TestQuantizedConv::test_qconv2d_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv2d_relu, test/test_quantization.py::TestQuantizedConv::test_qconv2d_relu_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv2d_unpack, test/test_quantization.py::TestQuantizedConv::test_qconv3d, test/test_quantization.py::TestQuantizedConv::test_qconv3d_relu, test/test_quantization.py::TestQuantizedConv::test_qconv3d_unpack, test/test_quantization.py::TestQuantizedConv::test_qconv_transpose1d, test/test_quantization.py::TestQuantizedConv::test_qconv_transpose2d, test/test_quantization.py::TestQuantizedConv::test_qconv_transpose3d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv1d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv2d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv3d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose1d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose2d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose3d, test/test_quantization.py::TestDynamicQuantizedOps::test_linear_prepack_fp16_numerics, test/test_quantization.py::TestDynamicQuantizedOps::test_qlinear, test/test_quantization.py::TestDynamicQuantizedOps::test_qlinear_dynamic_fp16, test/test_quantization.py::TestDynamicQuantizedOps::test_qlinear_legacy, test/test_quantization.py::TestDynamicQuantizedOps::test_qlstmGRU, test/test_quantization.py::TestDynamicQuantizedOps::test_qrnncell, test/test_quantization.py::TestComparatorOps::test_compare_tensor_scalar, test/test_quantization.py::TestComparatorOps::test_compare_tensor_tensor, test/test_quantization.py::TestPadding::test_constant_padNd, test/test_quantization.py::TestPadding::test_reflection_pad1d, test/test_quantization.py::TestPadding::test_reflection_pad2d, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_2d_indices, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2bit, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2bit_unpack, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2d_indices, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_4bit, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_4bit_unpack, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_byte, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_byte_unpack, test/test_quantization.py::TestQuantizedFunctionalOps::test_conv1d_api, test/test_quantization.py::TestQuantizedFunctionalOps::test_conv2d_api, test/test_quantization.py::TestQuantizedFunctionalOps::test_conv3d_api, test/test_quantization.py::TestQuantizedFunctionalOps::test_grid_sample, test/test_quantization.py::TestQuantizedFunctionalOps::test_relu_api, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_fake_quant_control, test/test_quantization.py::TestFakeQuantizeOps::test_fake_quant_per_channel_qparam_range, test/test_quantization.py::TestFakeQuantizeOps::test_fake_quant_preserves_qparam_shapes_for_activations, test/test_quantization.py::TestFakeQuantizeOps::test_fixed_qparams_fq_module, test/test_quantization.py::TestFakeQuantizeOps::test_forward_backward_per_tensor_with_amp, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_half_precision_numerics, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_half_precision_numerics, test/test_quantization.py::TestFakeQuantizeOps::test_fq_module_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_fq_serializable_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_tensor_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_tensor_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_channel_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_channel_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_numerical_consistency_per_channel, test/test_quantization.py::TestFakeQuantizeOps::test_numerical_consistency_per_tensor, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_backward_op_fake_quant_off, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_backward_op, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_moving_avg, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_moving_avg_per_channel, test/test_quantization.py::TestQuantizedTensor::test_bfp16_quantize, test/test_quantization.py::TestQuantizedTensor::test_choose_qparams, test/test_quantization.py::TestQuantizedTensor::test_choose_qparams_optimized, test/test_quantization.py::TestQuantizedTensor::test_clone, test/test_quantization.py::TestQuantizedTensor::test_compare_per_channel_device_numerics, test/test_quantization.py::TestQuantizedTensor::test_compare_per_tensor_device_numerics, test/test_quantization.py::TestQuantizedTensor::test_cuda_quantization_does_not_pin_memory, test/test_quantization.py::TestQuantizedTensor::test_decomposed_dequantize_per_channel, test/test_quantization.py::TestQuantizedTensor::test_decomposed_dequantize_per_tensor, test/test_quantization.py::TestQuantizedTensor::test_decomposed_dynamic_quant_pattern, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_channel, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_tensor, test/test_quantization.py::TestQuantizedTensor::test_dequantize_fp16_cpu, test/test_quantization.py::TestQuantizedTensor::test_dequantize_fp16_cuda, test/test_quantization.py::TestQuantizedTensor::test_fp16_saturate_op, test/test_quantization.py::TestQuantizedTensor::test_jit_serialization, test/test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_creation_cpu, test/test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_creation_cuda, test/test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_to_memory_format, test/test_quantization.py::TestQuantizedTensor::test_per_channel_to_device, test/test_quantization.py::TestQuantizedTensor::test_per_tensor_qtensor_to_memory_format, test/test_quantization.py::TestQuantizedTensor::test_per_tensor_to_device, test/test_quantization.py::TestQuantizedTensor::test_pickle_checkpoint_qtensor, test/test_quantization.py::TestQuantizedTensor::test_qscheme_pickle, test/test_quantization.py::TestQuantizedTensor::test_qtensor_channel_float_assignment, test/test_quantization.py::TestQuantizedTensor::test_qtensor_copy, test/test_quantization.py::TestQuantizedTensor::test_qtensor_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_creation, test/test_quantization.py::TestQuantizedTensor::test_qtensor_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_dtypes, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel_nhwc, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_tensor, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_tensor_nhwc, test/test_quantization.py::TestQuantizedTensor::test_qtensor_float_assignment, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_select_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_select_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_int_repr, test/test_quantization.py::TestQuantizedTensor::test_qtensor_legacy_new_failure, test/test_quantization.py::TestQuantizedTensor::test_qtensor_load_save, test/test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_load_save, test/test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_permute, test/test_quantization.py::TestQuantizedTensor::test_qtensor_permute, test/test_quantization.py::TestQuantizedTensor::test_qtensor_quant_dequant, test/test_quantization.py::TestQuantizedTensor::test_qtensor_quantize_per_channel, test/test_quantization.py::TestQuantizedTensor::test_qtensor_reshape, test/test_quantization.py::TestQuantizedTensor::test_qtensor_resize, test/test_quantization.py::TestQuantizedTensor::test_qtensor_sub_byte_aligned_cols, test/test_quantization.py::TestQuantizedTensor::test_qtensor_sub_byte_not_aligned_cols, test/test_quantization.py::TestQuantizedTensor::test_qtensor_unsqueeze, test/test_quantization.py::TestQuantizedTensor::test_qtensor_view, test/test_quantization.py::TestQuantizedTensor::test_quant_pin_memory, test/test_quantization.py::TestQuantizedTensor::test_quantize_per_channel_float_qparams, test/test_quantization.py::TestQuantizedTensor::test_quantize_per_channel_sub_byte, test/test_quantization.py::TestQuantizedTensor::test_repeat, test/test_quantization.py::TestQuantizedTensor::test_torch_qtensor_deepcopy, test/test_quantization.py::TestFakeQuantize::test_fq_module_per_channel, test/test_quantization.py::TestFakeQuantize::test_fq_serializable_per_channel, test/test_quantization.py::TestFakeQuantize::test_quant_min_max_override, test/test_quantization.py::TestObserver::test_histogram_observer_consistent_buffer_shape, test/test_quantization.py::TestObserver::test_histogram_observer_save_load_state_dict, test/test_quantization.py::TestObserver::test_memoryless_minmaxobserver, test/test_quantization.py::TestObserver::test_memoryless_perchannelminmaxobserver, test/test_quantization.py::TestObserver::test_observer_qparams_respects_device_affinity, test/test_quantization.py::TestObserver::test_observer_scriptable, test/test_quantization.py::TestObserver::test_per_channel_observers, test/test_quantization.py::TestObserver::test_per_tensor_observers, test/test_quantization.py::TestObserver::test_save_load_state_dict_script, test/test_quantization.py::TestObserver::test_state_dict_respects_device_affinity, test/test_quantization.py::TestObserver::test_zero_numel, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d_serialization, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d_serialization, test/test_quantization.py::TestStaticQuantizedModule::test_channel_shuffle, test/test_quantization.py::TestStaticQuantizedModule::test_conv1d_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv1d_relu_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv2d_add, test/test_quantization.py::TestStaticQuantizedModule::test_conv2d_add_relu, test/test_quantization.py::TestStaticQuantizedModule::test_conv2d_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv2d_relu_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv3d_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv3d_relu_api, test/test_quantization.py::TestStaticQuantizedModule::test_dropout, test/test_quantization.py::TestStaticQuantizedModule::test_dropout_serialization, test/test_quantization.py::TestStaticQuantizedModule::test_elu, test/test_quantization.py::TestStaticQuantizedModule::test_embedding_api, test/test_quantization.py::TestStaticQuantizedModule::test_embedding_bag_api, test/test_quantization.py::TestStaticQuantizedModule::test_group_norm, test/test_quantization.py::TestStaticQuantizedModule::test_hard_swish, test/test_quantization.py::TestStaticQuantizedModule::test_instance_norm, test/test_quantization.py::TestStaticQuantizedModule::test_layer_norm, test/test_quantization.py::TestStaticQuantizedModule::test_leaky_relu, test/test_quantization.py::TestStaticQuantizedModule::test_linear, test/test_quantization.py::TestStaticQuantizedModule::test_linear_leaky_relu, test/test_quantization.py::TestStaticQuantizedModule::test_linear_relu, test/test_quantization.py::TestStaticQuantizedModule::test_linear_tanh, test/test_quantization.py::TestStaticQuantizedModule::test_pool_api, test/test_quantization.py::TestStaticQuantizedModule::test_prelu, test/test_quantization.py::TestStaticQuantizedModule::test_quant_dequant_api, test/test_quantization.py::TestStaticQuantizedModule::test_relu, test/test_quantization.py::TestStaticQuantizedModule::test_sigmoid, test/test_quantization.py::TestDynamicQuantizedModule::test_cell_api, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv1d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv2d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv3d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose1d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose2d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose3d, test/test_quantization.py::TestDynamicQuantizedModule::test_gru_api, test/test_quantization.py::TestDynamicQuantizedModule::test_linear_api, test/test_quantization.py::TestDynamicQuantizedModule::test_lstm_api, test/test_quantization.py::TestReferenceQuantizedModule::test_linear_decomposed_weight_custom_qmin_qmax, test/test_quantization.py::TestReferenceQuantizedModule::test_rnn, test/test_quantization.py::TestReferenceQuantizedModule::test_rnn_cell, test/test_quantization.py::TestReferenceQuantizedModule::test_sparse, test/test_quantization.py::TestRecordHistogramObserver::test_observer_scriptable, test/test_quantization.py::TestRecordHistogramObserver::test_record_observer, test/test_quantization.py::TestHistogramObserver::test_histogram_observer, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_against_reference, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_correct_numel, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_extreme_inputs, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_one_sided, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_same_inputs, test/test_quantization.py::TestHistogramObserver::test_observer_scriptable, test/test_quantization.py::TestDistributed::test_device_affinity, test/test_quantization.py::TestDistributed::test_fake_quant_preserves_buffers, test/test_quantization.py::TestDistributed::test_observers_preserve_buffers, test/test_quantization.py::TestDistributed::test_qat_convbn_fused_syncbn_replacement, test/test_quantization.py::TestDistributed::test_qat_data_parallel, test/test_quantization.py::TestDistributed::test_syncbn_preserves_qconfig, test/test_quantization.py::TestFusedObsFakeQuantModule::test_compare_fused_obs_fq_oss_module, test/test_quantization.py::TestFusedObsFakeQuantModule::test_default_fused_qat_config, test/test_quantization.py::TestFusedObsFakeQuantModule::test_embedding_bag_qat_config, test/test_quantization.py::TestFusedObsFakeQuantModule::test_embedding_qat_config, test/test_quantization.py::TestFusedObsFakeQuantModule::test_fused_mod_per_channel, test/test_quantization.py::TestFusedObsFakeQuantModule::test_fused_mod_reduce_range, test/test_quantization.py::TestFusedObsFakeQuantModule::test_fused_obs_fq_module, test/test_quantization.py::TestFusedObsFakeQuantModule::test_fused_obs_fq_moving_avg_module, test/test_quantization.py::TestBackendConfig::test_backend_config_from_dict, test/test_quantization.py::TestBackendConfig::test_backend_config_set_backend_pattern_config, test/test_quantization.py::TestBackendConfig::test_backend_config_set_name, test/test_quantization.py::TestBackendConfig::test_backend_config_to_dict, test/test_quantization.py::TestBackendConfig::test_backend_op_config_add_dtype_config, test/test_quantization.py::TestBackendConfig::test_backend_op_config_from_dict, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_extra_inputs_getter, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_fused_module, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_fuser_method, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_input_type_to_index, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_num_tensor_args_to_observation_type, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_observation_type, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_qat_module, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_reference_quantized_module, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_module, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_node_getter, test/test_quantization.py::TestBackendConfig::test_backend_op_config_to_dict, test/test_quantization.py::TestBackendConfig::test_dtype_config_from_dict, test/test_quantization.py::TestBackendConfig::test_dtype_config_to_dict, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_complex_args, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_default_kwargs, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_simple, test/test_quantization.py::TestUtils::test_quantize_weight_clamping_per_channel, test/test_quantization.py::TestUtils::test_quantize_weight_clamping_per_tensor, test/test_quantization.py::TestQuantizationDocs::test_quantization_doc_custom, test/test_quantization.py::TestQuantizationDocs::test_quantization_doc_fx, test/test_quantization.py::TestQuantizationDocs::test_quantization_doc_ptdq, test/test_quantization.py::TestQuantizationDocs::test_quantization_doc_ptsq, test/test_quantization.py::TestQuantizationDocs::test_quantization_doc_qat, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_activations, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_activations_in_non_leaf_module_list, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_fails_early, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_qconfig_none, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_custom_module_class, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_dequant_stub, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_forward_hooks_preserved, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_manual, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_mha_batch_first_attr_is_copied_in_prepare, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_nested1, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_nested2, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_nested3, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_normalization, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding_bag, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_quantwrapper_attaches_qconfig_to_dequant, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_resnet_base, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_save_load_state_dict, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_single_layer, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_skip_quant, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_two_layers, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_embedding_ops_dynamic, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_forward_hooks_preserved, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_linear_relu_fusion, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested1, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested2, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested3, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_per_channel_linear_quantize, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_quantized_rnn, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_quantized_rnn_cell, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_single_layer, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_two_layers, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_type_match_rule, test/test_quantization.py::TestQuantizeEagerOps::test_conv_1d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_2d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_3d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_1d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_2d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_3d, test/test_quantization.py::TestQuantizeEagerOps::test_functional_module, test/test_quantization.py::TestQuantizeEagerOps::test_int16_reference_module, test/test_quantization.py::TestQuantizeEagerOps::test_leaky_relu, test/test_quantization.py::TestQuantizeEagerOps::test_linear, test/test_quantization.py::TestQuantizeEagerOps::test_relu, test/test_quantization.py::TestQuantizeEagerQAT::test_add_scalar_uses_input_qparams, test/test_quantization.py::TestQuantizeEagerQAT::test_conv_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_conv_linear_symm, test/test_quantization.py::TestQuantizeEagerQAT::test_defused_embedding_bag_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_dropout, test/test_quantization.py::TestQuantizeEagerQAT::test_dynamic_qat_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_embedding_bag_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_embedding_qat_qconfig_equal, test/test_quantization.py::TestQuantizeEagerQAT::test_eval_only_fake_quant, test/test_quantization.py::TestQuantizeEagerQAT::test_forward_hooks_preserved, test/test_quantization.py::TestQuantizeEagerQAT::test_manual, test/test_quantization.py::TestQuantizeEagerQAT::test_mul_scalar_uses_input_qparams, test/test_quantization.py::TestQuantizeEagerQAT::test_qat_embedding_bag_errors, test/test_quantization.py::TestQuantizeEagerQAT::test_train_save_load_eval, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_conv_bn_folded_vs_unfolded, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_conv_bn_relu, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_fixed_qparam_ops, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_leaky_relu, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_numerics, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_symm_numerics, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_workflow, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_relu, test/test_quantization.py::TestFuseEager::test_forward_hooks_preserved, test/test_quantization.py::TestFuseEager::test_fuse_function_customization, test/test_quantization.py::TestFuseEager::test_fuse_module_eval, test/test_quantization.py::TestFuseEager::test_fuse_module_train, test/test_quantization.py::TestFuseEager::test_fusion_conv_with_bias, test/test_quantization.py::TestFuseEager::test_fusion_convtranspose_bn_eval, test/test_quantization.py::TestFuseEager::test_fusion_linear_bn_eval, test/test_quantization.py::TestFuseEager::test_fusion_sequential_model_eval, test/test_quantization.py::TestFuseEager::test_fusion_sequential_model_train, test/test_quantization.py::TestModelNumericsEager::test_fake_quant_true_quant_compare, test/test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_channel, test/test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_tensor, test/test_quantization.py::TestModelNumericsEager::test_weight_only_activation_only_fakequant, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_functional_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_lstm_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_functional_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_linear_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_lstm_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_partial, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_submodule_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_lstm_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_mobilenet_v2, test/test_quantization.py::TestNumericSuiteEager::test_mobilenet_v3, test/test_quantization.py::TestNumericSuiteEager::test_output_logger, test/test_quantization.py::TestNumericSuiteEager::test_shadow_logger, test/test_quantization.py::TestEqualizeEager::test_converged, test/test_quantization.py::TestEqualizeEager::test_cross_layer_equalization, test/test_quantization.py::TestEqualizeEager::test_equalize, test/test_quantization.py::TestEqualizeEager::test_equalize_fused_convrelu, test/test_quantization.py::TestEqualizeEager::test_equalize_fused_linearrelu, test/test_quantization.py::TestBiasCorrectionEager::test_conv_chain, test/test_quantization.py::TestBiasCorrectionEager::test_linear_chain, test/test_quantization.py::TestFuseFx::test_fuse_addtional_fuser_method, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_by_default, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_lowering, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_onednn, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_relu, test/test_quantization.py::TestFuseFx::test_fuse_convtranspose_bn_eval, test/test_quantization.py::TestFuseFx::test_fuse_custom_pattern, test/test_quantization.py::TestFuseFx::test_fuse_linear_bn_eval, test/test_quantization.py::TestFuseFx::test_fuse_linear_bn_leaky_relu_onednn, test/test_quantization.py::TestFuseFx::test_fuse_linear_tanh_for_onednn_backend, test/test_quantization.py::TestFuseFx::test_fuse_module_relu, test/test_quantization.py::TestFuseFx::test_fusion_pattern_with_matchallnode, test/test_quantization.py::TestFuseFx::test_fusion_pattern_with_multiple_inputs, test/test_quantization.py::TestFuseFx::test_linear_bn_leaky_relu_not_fused_by_default, test/test_quantization.py::TestFuseFx::test_linear_tanh_not_fused_by_default, test/test_quantization.py::TestFuseFx::test_problematic_fuse_example, test/test_quantization.py::TestFuseFx::test_qconfig_fused_module, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_dynamic_quant, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_per_channel_quant, test/test_quantization.py::TestQuantizeFx::test_assert_on_size_after_quant_layer, test/test_quantization.py::TestQuantizeFx::test_attention, test/test_quantization.py::TestQuantizeFx::test_backend_config_check_for_weight_and_bias, test/test_quantization.py::TestQuantizeFx::test_backend_config_quantization_range, test/test_quantization.py::TestQuantizeFx::test_backend_config_scale_min, test/test_quantization.py::TestQuantizeFx::test_change_backend_config_for_fixed_qparam_ops, test/test_quantization.py::TestQuantizeFx::test_channel_shuffle_lowering, test/test_quantization.py::TestQuantizeFx::test_conv_bn_relu, test/test_quantization.py::TestQuantizeFx::test_conv_linear_not_reference, test/test_quantization.py::TestQuantizeFx::test_conv_linear_reference, test/test_quantization.py::TestQuantizeFx::test_conv_lowering, test/test_quantization.py::TestQuantizeFx::test_conv_transpose_not_reference, test/test_quantization.py::TestQuantizeFx::test_conv_transpose_reference, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_observed_to_quantized_mapping, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_preserved_attributes, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_convert_qconfig_mapping, test/test_quantization.py::TestQuantizeFx::test_convtranspose_per_channel_fails_early, test/test_quantization.py::TestQuantizeFx::test_copy_node_has_shared_actpp_instance, test/test_quantization.py::TestQuantizeFx::test_custom_module_class, test/test_quantization.py::TestQuantizeFx::test_custom_module_class_input_has_duplicate_nodes, test/test_quantization.py::TestQuantizeFx::test_custom_module_class_input_has_multiple_users, test/test_quantization.py::TestQuantizeFx::test_deepcopy_preserve_attributes, test/test_quantization.py::TestQuantizeFx::test_default_qconfig_mapping_override_global, test/test_quantization.py::TestQuantizeFx::test_default_quant_after_none_qconfig, test/test_quantization.py::TestQuantizeFx::test_dequantize, test/test_quantization.py::TestQuantizeFx::test_dict_output, test/test_quantization.py::TestQuantizeFx::test_dynamic_linear_input_multiple_use, test/test_quantization.py::TestQuantizeFx::test_dynamic_quant_fp16, test/test_quantization.py::TestQuantizeFx::test_dynamic_quant_weight_observer, test/test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion, test/test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion_multiple_uses, test/test_quantization.py::TestQuantizeFx::test_fold_quant_dequant, test/test_quantization.py::TestQuantizeFx::test_fp32_input_fp32_output, test/test_quantization.py::TestQuantizeFx::test_fp32_input_quantized_output, test/test_quantization.py::TestQuantizeFx::test_fp32_sum, test/test_quantization.py::TestQuantizeFx::test_fuse_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_fuse_custom_config_set_preserved_attributes, test/test_quantization.py::TestQuantizeFx::test_fuse_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_fused_module_qat_swap, test/test_quantization.py::TestQuantizeFx::test_fusion_pattern_unquantized, test/test_quantization.py::TestQuantizeFx::test_get_default_qconfig_valid_backend, test/test_quantization.py::TestQuantizeFx::test_get_executorch_backend_config, test/test_quantization.py::TestQuantizeFx::test_getattr_with_nontensor_result, test/test_quantization.py::TestQuantizeFx::test_linear_bn, test/test_quantization.py::TestQuantizeFx::test_linear_leaky_relu_lowering, test/test_quantization.py::TestQuantizeFx::test_linear_qint8_activation, test/test_quantization.py::TestQuantizeFx::test_linear_shape_view, test/test_quantization.py::TestQuantizeFx::test_linear_size_view, test/test_quantization.py::TestQuantizeFx::test_linear_tanh_lowering, test/test_quantization.py::TestQuantizeFx::test_lowering_functional_conv_with_kwargs, test/test_quantization.py::TestQuantizeFx::test_lowering_functional_linear_with_kwargs, test/test_quantization.py::TestQuantizeFx::test_masked_fill_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_match_pattern_with_multiple_args, test/test_quantization.py::TestQuantizeFx::test_mixed_dtypes, test/test_quantization.py::TestQuantizeFx::test_mul_add_fp16_config, test/test_quantization.py::TestQuantizeFx::test_no_obs_between_unmatched_node_and_copy_node, test/test_quantization.py::TestQuantizeFx::test_non_traceable_module, test/test_quantization.py::TestQuantizeFx::test_not_used, test/test_quantization.py::TestQuantizeFx::test_observer_fqn, test/test_quantization.py::TestQuantizeFx::test_output_lists_and_dicts, test/test_quantization.py::TestQuantizeFx::test_packed_weight_fused_op, test/test_quantization.py::TestQuantizeFx::test_pattern_match, test/test_quantization.py::TestQuantizeFx::test_pattern_match_constant, test/test_quantization.py::TestQuantizeFx::test_permute_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_float_to_observed_mapping, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_input_quantized_indexes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_classes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_names, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_output_quantized_indexes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_preserved_attributes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_standalone_module_class, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_standalone_module_name, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_prepare_mode, test/test_quantization.py::TestQuantizeFx::test_prepared_model_deepcopy, test/test_quantization.py::TestQuantizeFx::test_preserve_attributes, test/test_quantization.py::TestQuantizeFx::test_preserve_qconfig, test/test_quantization.py::TestQuantizeFx::test_preserve_tuple, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_split_tuple_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_tuple_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_list_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_split_list_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_split_tuple_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_tuple_args, test/test_quantization.py::TestQuantizeFx::test_qat_and_script, test/test_quantization.py::TestQuantizeFx::test_qat_prepare_device_affinity, test/test_quantization.py::TestQuantizeFx::test_qat_skip_untraced, test/test_quantization.py::TestQuantizeFx::test_qconfig_dict_setup, test/test_quantization.py::TestQuantizeFx::test_qconfig_dict_with_fused_modules, test/test_quantization.py::TestQuantizeFx::test_qconfig_for_call_func, test/test_quantization.py::TestQuantizeFx::test_qconfig_for_call_method, test/test_quantization.py::TestQuantizeFx::test_qconfig_function, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_from_dict, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_repr, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_global, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_object_type_order, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_regex, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_object_type, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_to_dict, test/test_quantization.py::TestQuantizeFx::test_qconfig_module_name_object_type_order, test/test_quantization.py::TestQuantizeFx::test_qconfig_module_name_regex, test/test_quantization.py::TestQuantizeFx::test_qconfig_module_type, test/test_quantization.py::TestQuantizeFx::test_qconfig_none, test/test_quantization.py::TestQuantizeFx::test_qconfig_precedence, test/test_quantization.py::TestQuantizeFx::test_qconfig_qat_module_type, test/test_quantization.py::TestQuantizeFx::test_qnnpack_backend_config, test/test_quantization.py::TestQuantizeFx::test_qparams_buffers, test/test_quantization.py::TestQuantizeFx::test_qparams_fqn, test/test_quantization.py::TestQuantizeFx::test_quant_output_always_observed, test/test_quantization.py::TestQuantizeFx::test_quantized_input_fp32_output, test/test_quantization.py::TestQuantizeFx::test_quantized_input_quantized_output, test/test_quantization.py::TestQuantizeFx::test_quantized_model_type, test/test_quantization.py::TestQuantizeFx::test_ref_conv_module, test/test_quantization.py::TestQuantizeFx::test_ref_linear_module, test/test_quantization.py::TestQuantizeFx::test_register_patterns, test/test_quantization.py::TestQuantizeFx::test_relu_lowering, test/test_quantization.py::TestQuantizeFx::test_remove_qconfig, test/test_quantization.py::TestQuantizeFx::test_repeat_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_reroute_tuple_getitem_patterns, test/test_quantization.py::TestQuantizeFx::test_reshape_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_return_none, test/test_quantization.py::TestQuantizeFx::test_reuse_input_qconfig, test/test_quantization.py::TestQuantizeFx::test_save_observer_state_dict, test/test_quantization.py::TestQuantizeFx::test_sequential, test/test_quantization.py::TestQuantizeFx::test_shape_followed_by_quantized_op, test/test_quantization.py::TestQuantizeFx::test_size_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_stack_trace_preserved_linear, test/test_quantization.py::TestQuantizeFx::test_standalone_module_float_interface, test/test_quantization.py::TestQuantizeFx::test_standalone_module_quantized_interface, test/test_quantization.py::TestQuantizeFx::test_state_dict, test/test_quantization.py::TestQuantizeFx::test_static_lstm, test/test_quantization.py::TestQuantizeFx::test_static_lstm_consume_tuple, test/test_quantization.py::TestQuantizeFx::test_static_lstm_with_custom_fixed_qparams, test/test_quantization.py::TestQuantizeFx::test_sub_scalar, test/test_quantization.py::TestQuantizeFx::test_symmetric_qnnpack_qconfig_mapping, test/test_quantization.py::TestQuantizeFx::test_torch_transpose_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_torch_unsqueeze_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_trace_quantize_per_tensor, test/test_quantization.py::TestQuantizeFx::test_transpose_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_unsqueeze__nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_unsqueeze_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_view_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFxOps::test_add, test/test_quantization.py::TestQuantizeFxOps::test_add_relu, test/test_quantization.py::TestQuantizeFxOps::test_add_relu_multiple_uses_of_relu, test/test_quantization.py::TestQuantizeFxOps::test_ave_pool_with_custom_cfg, test/test_quantization.py::TestQuantizeFxOps::test_bmm, test/test_quantization.py::TestQuantizeFxOps::test_bmm_int_reference, test/test_quantization.py::TestQuantizeFxOps::test_boolean_tensor, test/test_quantization.py::TestQuantizeFxOps::test_cat, test/test_quantization.py::TestQuantizeFxOps::test_chunk, test/test_quantization.py::TestQuantizeFxOps::test_clamp, test/test_quantization.py::TestQuantizeFxOps::test_conv_module, test/test_quantization.py::TestQuantizeFxOps::test_conv_transpose_1d, test/test_quantization.py::TestQuantizeFxOps::test_conv_transpose_2d, test/test_quantization.py::TestQuantizeFxOps::test_copy_node_fp32_input, test/test_quantization.py::TestQuantizeFxOps::test_div, test/test_quantization.py::TestQuantizeFxOps::test_elu, test/test_quantization.py::TestQuantizeFxOps::test_embedding, test/test_quantization.py::TestQuantizeFxOps::test_embedding_bag, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_fp16, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_qint8, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_wrong_qconfig, test/test_quantization.py::TestQuantizeFxOps::test_float_functional, test/test_quantization.py::TestQuantizeFxOps::test_functional_conv, test/test_quantization.py::TestQuantizeFxOps::test_functional_linear, test/test_quantization.py::TestQuantizeFxOps::test_gelu_normal, test/test_quantization.py::TestQuantizeFxOps::test_gelu_reference, test/test_quantization.py::TestQuantizeFxOps::test_general_shape_ops, test/test_quantization.py::TestQuantizeFxOps::test_general_value_ops, test/test_quantization.py::TestQuantizeFxOps::test_getitem, test/test_quantization.py::TestQuantizeFxOps::test_hardswish, test/test_quantization.py::TestQuantizeFxOps::test_instance_norm, test/test_quantization.py::TestQuantizeFxOps::test_int8_input_no_unnecessary_fq, test/test_quantization.py::TestQuantizeFxOps::test_layer_norm, test/test_quantization.py::TestQuantizeFxOps::test_leaky_relu, test/test_quantization.py::TestQuantizeFxOps::test_linear_dynamic_fp16, test/test_quantization.py::TestQuantizeFxOps::test_linear_module, test/test_quantization.py::TestQuantizeFxOps::test_linear_static_fp16, test/test_quantization.py::TestQuantizeFxOps::test_mish_reference, test/test_quantization.py::TestQuantizeFxOps::test_mul, test/test_quantization.py::TestQuantizeFxOps::test_mul_relu, test/test_quantization.py::TestQuantizeFxOps::test_multiple_qconfigs_for_single_value, test/test_quantization.py::TestQuantizeFxOps::test_narrow, test/test_quantization.py::TestQuantizeFxOps::test_norm_weight_bias, test/test_quantization.py::TestQuantizeFxOps::test_pixel_shuffle, test/test_quantization.py::TestQuantizeFxOps::test_pixel_unshuffle, test/test_quantization.py::TestQuantizeFxOps::test_prelu, test/test_quantization.py::TestQuantizeFxOps::test_qbatch_norm, test/test_quantization.py::TestQuantizeFxOps::test_qbatch_norm_relu, test/test_quantization.py::TestQuantizeFxOps::test_qmatmul, test/test_quantization.py::TestQuantizeFxOps::test_quantized_add_qat, test/test_quantization.py::TestQuantizeFxOps::test_quantized_conv_relu, test/test_quantization.py::TestQuantizeFxOps::test_quantized_mul_qat, test/test_quantization.py::TestQuantizeFxOps::test_ref_pattern_multi_use, test/test_quantization.py::TestQuantizeFxOps::test_reshape_fp16, test/test_quantization.py::TestQuantizeFxOps::test_rnn, test/test_quantization.py::TestQuantizeFxOps::test_rnn_cell, test/test_quantization.py::TestQuantizeFxOps::test_silu_reference, test/test_quantization.py::TestQuantizeFxOps::test_softmax_normal, test/test_quantization.py::TestQuantizeFxOps::test_softmax_reference, test/test_quantization.py::TestQuantizeFxOps::test_sub, test/test_quantization.py::TestQuantizeFxOps::test_sum, test/test_quantization.py::TestQuantizeFxModels::test_model_dropout, test/test_quantization.py::TestQuantizeFxModels::test_prepare_serialize_switch_device_convert, test/test_quantization.py::TestQuantizeFxModels::test_qat_embedding_linear, test/test_quantization.py::TestQuantizeFxModels::test_qat_embeddingbag_linear, test/test_quantization.py::TestQuantizeFxModels::test_qat_functional_linear, test/test_quantization.py::TestQuantizeFxModels::test_resnet18_ddp, test/test_quantization.py::TestQuantizeFxModels::test_resnet_base, test/test_quantization.py::TestQuantizeFxModels::test_static_gpu_convert_basic, test/test_quantization.py::TestQuantizeFxModels::test_switch_device_prepare_convert, test/test_quantization.py::TestQuantizeFxModels::test_torchvision, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_annotations_int, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_correct_output_replacement, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_graph_argument_order, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_internal_pattern_nodes_cannot_have_users_that_are_not_matched, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_multiple_pattern_match, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_is_entire_graph, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_output_pattern_node_can_have_users_that_are_not_matched, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_placeholder_matching, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_preserves_logic, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_replaces_referenced_submodules, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_single_pattern_match, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_traced_as_callable, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_with_oneliner_pattern, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_writer_replace_consecutive_submodules, test/test_quantization.py::TestQuantizePT2E::test_qconfig_module_type, test/test_quantization.py::TestQuantizePT2E::test_qconfig_none, test/test_quantization.py::TestQuantizePT2E::test_rearrange_weight_observer_for_decomposed_linear, test/test_quantization.py::TestQuantizePT2EModels::test_resnet18, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_inductor_backend_config_conv, test/test_quantization.py::TestFXGraphMatcher::test_dict_return_type, test/test_quantization.py::TestFXGraphMatcher::test_matching_failure_node_count, test/test_quantization.py::TestFXGraphMatcher::test_matching_failure_node_type, test/test_quantization.py::TestFXGraphMatcher::test_methods, test/test_quantization.py::TestFXGraphMatcher::test_nodes_before_cat, test/test_quantization.py::TestFXGraphMatcher::test_nodes_with_equal_types_get_matched, test/test_quantization.py::TestFXGraphMatcher::test_op_relationship_mapping, test/test_quantization.py::TestFXGraphMatcher::test_results_order, test/test_quantization.py::TestFXGraphMatcher::test_simple_fun, test/test_quantization.py::TestFXGraphMatcher::test_simple_fusion, test/test_quantization.py::TestFXGraphMatcher::test_simple_mod, test/test_quantization.py::TestFXGraphMatcher::test_simple_mod_multi, test/test_quantization.py::TestFXGraphMatcher::test_simple_tensor_ops, test/test_quantization.py::TestFXGraphMatcher::test_user_defined_function, test/test_quantization.py::TestFXGraphMatcherModels::test_mobilenet_v2, test/test_quantization.py::TestFXGraphMatcherModels::test_mobilenet_v2_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_loggers_cuda, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_mul_inputs_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_cuda, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_fun_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_meth_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_mod_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_mod_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extend_logger_results_with_comparison, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_conv_fun_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_conv_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_cuda, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_fqn, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_linear_fun_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_linear_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_mod_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_mod_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_fp16_shadows_fp32, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_fp32_coverage, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_fp32_simple, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_int8_fun, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_int8_mod, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_layer_names, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_shadow_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_vs_linear_fp16_shadow_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_weights, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_kwargs_shadow, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_loggers_preserve_qat_numerics, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_logging_inputs, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fqn, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fun_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_meth_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_mod_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_mod_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_mul_add_cat_stack_skips_shadowing, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_io_dtype_coverage, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_either_fp32_or_int8_input, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_only_kwargs_skips_shadowing, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_ops_with_same_fp32_and_int8_signature, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_shadow_activations_fqn, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_shadow_loggers_preserve_qat_numerics, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_unsupported_op_copy_skips_shadowing, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_defined_function, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_module, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_module_scriptable, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_conv_bn_relu_fusion_fp32, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_conv_bn_relu_fusion_quant, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_functions, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_fp32_fp32, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_fp32_quant, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_fp32, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_quant, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_mobilenet_v2, test/test_quantization.py::TestFXNumericSuiteNShadows::test_conv_bn_relu_mod, test/test_quantization.py::TestFXNumericSuiteNShadows::test_custom_functions_and_tracer, test/test_quantization.py::TestFXNumericSuiteNShadows::test_extract_weights_linear, test/test_quantization.py::TestFXNumericSuiteNShadows::test_functions, test/test_quantization.py::TestFXNumericSuiteNShadows::test_linear_mod, test/test_quantization.py::TestFXNumericSuiteNShadows::test_linear_relu_mod, test/test_quantization.py::TestFXNumericSuiteNShadows::test_logger_enabled_and_save_activations_flags, test/test_quantization.py::TestFXNumericSuiteNShadows::test_mobilenet_v2, test/test_quantization.py::TestFXNumericSuiteNShadows::test_partial_qconfig_mapping, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_deduplication, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_end_to_end, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_from_list, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_insert_padding, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_ordering, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_repr, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_retroactive_padding, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_conv, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_conv, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_conv, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_mobilenet_v2, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_resnet18, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_sparsenn_compare_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_sparsenn_shadow, test/test_quantization.py::TestFxModelReportDetector::test_conv_sub_class_considered, test/test_quantization.py::TestFxModelReportDetector::test_fusion_layer_in_sequential, test/test_quantization.py::TestFxModelReportDetector::test_multi_linear_model_without_per_channel, test/test_quantization.py::TestFxModelReportDetector::test_multiple_q_config_options, test/test_quantization.py::TestFxModelReportDetector::test_qat_aware_model_example, test/test_quantization.py::TestFxModelReportDetector::test_sequential_model_format, test/test_quantization.py::TestFxModelReportDetector::test_simple_conv, test/test_quantization.py::TestFxModelReportObserver::test_observer_after_relu, test/test_quantization.py::TestFxModelReportObserver::test_random_epochs_and_batches, test/test_quantization.py::TestFxModelReportObserver::test_single_batch_of_ones, test/test_quantization.py::TestFxModelReportObserver::test_zero_tensor_errors, test/test_quantization.py::TestFxModelReportDetectDynamicStatic::test_nested_detection_case, test/test_quantization.py::TestFxModelReportClass::test_constructor, test/test_quantization.py::TestFxModelReportClass::test_equalization_mapping_generation, test/test_quantization.py::TestFxModelReportClass::test_generate_report, test/test_quantization.py::TestFxModelReportClass::test_generate_visualizer, test/test_quantization.py::TestFxModelReportClass::test_prepare_model_callibration, test/test_quantization.py::TestFxModelReportClass::test_qconfig_mapping_generation, test/test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_determine_points, test/test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen, test/test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen_empty, test/test_quantization.py::TestFxDetectOutliers::test_all_outlier_report_gen, test/test_quantization.py::TestFxDetectOutliers::test_multiple_run_consistent_spike_outlier_report_gen, test/test_quantization.py::TestFxDetectOutliers::test_no_outlier_report_gen, test/test_quantization.py::TestFxDetectOutliers::test_outlier_detection_determine_points, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_match_with_report, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_no_match, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_single_feat_match, test/test_quantization.py::TestFxModelReportVisualizer::test_get_modules_and_features, test/test_quantization.py::TestEqualizeFx::test_input_weight_eq_observer, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_activation_values, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_branching, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_convert, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_equalization_scales, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_graphs, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_prepare, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_results, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_weights_bias, test/test_quantization.py::TestEqualizeFx::test_selective_equalization, test/test_quantization.py::TestSerialization::test_conv2d, test/test_quantization.py::TestSerialization::test_conv2d_graph, test/test_quantization.py::TestSerialization::test_conv2d_graph_v2, test/test_quantization.py::TestSerialization::test_conv2d_graph_v3, test/test_quantization.py::TestSerialization::test_conv2d_nobias, test/test_quantization.py::TestSerialization::test_conv2d_nobias_graph, test/test_quantization.py::TestSerialization::test_conv2d_nobias_graph_v2, test/test_quantization.py::TestSerialization::test_conv2d_nobias_graph_v3, test/test_quantization.py::TestSerialization::test_conv2d_relu, test/test_quantization.py::TestSerialization::test_conv3d, test/test_quantization.py::TestSerialization::test_conv3d_relu, test/test_quantization.py::TestSerialization::test_default_qat_qconfig, test/test_quantization.py::TestSerialization::test_linear, test/test_quantization.py::TestSerialization::test_linear_dynamic, test/test_quantization.py::TestSerialization::test_linear_relu, test/test_quantization.py::TestSerialization::test_linear_relu_package_quantization_transforms, test/test_quantization.py::TestSerialization::test_lstm, test/test_quantization.py::TestSerialization::test_per_channel_observer, test/test_quantization.py::TestSerialization::test_per_tensor_observer, test/test_quantization.py::TestQuantizeJit::test_conv, test/test_quantization.py::TestQuantizeJit::test_conv_bn, test/test_quantization.py::TestQuantizeJit::test_conv_transpose, test/test_quantization.py::TestQuantizeJit::test_linear_dynamic_fp16, test/test_quantization.py::TestQuantizeJit::test_nested, test/test_quantization.py::TestQuantizeJit::test_observer_with_ignored_function, test/test_quantization.py::TestQuantizeJit::test_single_linear, test/test_quantization.py::TestQuantizeJit::test_single_linear_dynamic, test/test_quantization.py::TestQuantizeJit::test_skip_quant, test/test_quantization.py::TestQuantizeJitPasses::test_conv_trace, test/test_quantization.py::TestQuantizeJitPasses::test_convtranspose_trace, test/test_quantization.py::TestQuantizeJitPasses::test_dedup_module_uses, test/test_quantization.py::TestQuantizeJitPasses::test_finalize_debug, test/test_quantization.py::TestQuantizeJitPasses::test_finalize_for_linear, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_complex_cases, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_in_submodule, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_no_fusion, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_shared_classtype, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial_nobias, test/test_quantization.py::TestQuantizeJitPasses::test_fuse_linear, test/test_quantization.py::TestQuantizeJitPasses::test_inplace_option, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_child_qconfig, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_general_ops, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_if, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_if_consistent_observation, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_nested_if, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_reused_weight, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_interface, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_interface_unshare_type, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed_for_function, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed_in_submodule, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_shared_class_type, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_skip_values, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_weight_dtype, test/test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant, test/test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant_shared_class_type, test/test_quantization.py::TestQuantizeJitPasses::test_interface_with_fork, test/test_quantization.py::TestQuantizeJitPasses::test_module_list, test/test_quantization.py::TestQuantizeJitPasses::test_quantize_fork_wait, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequant_same_value, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize_in_block, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_quantize_for_if, test/test_quantization.py::TestQuantizeJitPasses::test_skip_dequant_constant_prop, test/test_quantization.py::TestQuantizeJitPasses::test_swap_functional_linear, test/test_quantization.py::TestQuantizeJitOps::test_cat_linear, test/test_quantization.py::TestQuantizeJitOps::test_clamp, test/test_quantization.py::TestQuantizeJitOps::test_conv_with_benchmark_flag, test/test_quantization.py::TestQuantizeJitOps::test_dequantize_tuple, test/test_quantization.py::TestQuantizeJitOps::test_elu, test/test_quantization.py::TestQuantizeJitOps::test_general_shape_ops, test/test_quantization.py::TestQuantizeJitOps::test_general_value_ops, test/test_quantization.py::TestQuantizeJitOps::test_group_norm, test/test_quantization.py::TestQuantizeJitOps::test_hardswish, test/test_quantization.py::TestQuantizeJitOps::test_instance_norm, test/test_quantization.py::TestQuantizeJitOps::test_layer_norm, test/test_quantization.py::TestQuantizeJitOps::test_linear, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNFuncInplaceRelu, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNFuncRelu, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNRelu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_alpha, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_relu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_relu_alpha, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_scalar, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_scalar_relu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_cat, test/test_quantization.py::TestQuantizeJitOps::test_quantized_conv, test/test_quantization.py::TestQuantizeJitOps::test_quantized_conv_relu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_mul, test/test_quantization.py::TestQuantizeJitOps::test_quantized_mul_relu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_mul_scalar, test/test_quantization.py::TestQuantizeJitOps::test_quantized_mul_scalar_relu, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_convert_dynamic_fp16, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_multi_op, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_quant_multi_uses, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_shared_weights, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_weight_observer, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_with_if, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_insert_quant_dequant_linear_dynamic, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_prepare_dynamic, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_prepare_dynamic_child_qconfig, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_quantize_dynamic_fp16, test/test_quantization.py::TestQuantizeDynamicJitOps::test_embedding_bag, test/test_quantization.py::TestQuantizeDynamicJitOps::test_embedding_bag_padding_idx_error, test/test_quantization.py::TestQuantizeDynamicJitOps::test_linear, test/test_quantization.py::TestFusionPasses::test_quantized_add_relu_fusion, test/test_quantization.py::TestDeprecatedJitQuantized::test_erase_class_tensor_shapes, test/test_quantization.py::TestDeprecatedJitQuantized::test_quantization_modules, test/test_quantization.py::TestDeprecatedJitQuantized::test_rnn_cell_quantized, test/test_quantization.py::TestDeprecatedJitQuantized::test_rnn_quantized, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_fake_quantize, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_fuse_modules, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_fuser_method_mappings, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_observer, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_qconfig, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quant_type, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quantization_mappings, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize_jit, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_stubs, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_utils, test/test_quantization.py::TestAOMigrationNNQuantized::test_functional_import, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_conv, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_dynamic_linear, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_embedding_ops, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_linear, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantizable_activation, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantizable_rnn, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantized_dynamic_import, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_activation, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_batchnorm, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_conv, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_dropout, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_embedding_ops, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_functional_modules, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_import, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_linear, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_normalization, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_utils, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_qat, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_quantized, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_conv_fused, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_fused, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_bn_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_conv_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_linear_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_nn_intrinsic_fused, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_no_import_nn_intrinsic_quantized_dynamic, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_convert, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_equalize, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fuse, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fusion_patterns, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_graph_module, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_match_utils, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_pattern_utils, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_prepare, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_quantization_patterns, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_utils, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_quantize_fx, test/test_quantization.py::TestBits::test_subclass, test/test_quantization.py::TestBits::test_types 2023-03-31T05:48:33.5969624Z 2023-03-31T05:48:33.5970419Z test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5971445Z test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool2d_nhwc <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5972387Z test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool3d_ndhwc <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5973356Z test_quantization.py::TestQuantizedOps::test_add_scalar_relu <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5974327Z test_quantization.py::TestQuantizedOps::test_advanced_indexing <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5975226Z test_quantization.py::TestQuantizedOps::test_avg_pool2d <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5976080Z test_quantization.py::TestQuantizedOps::test_avg_pool2d_nhwc <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5977006Z test_quantization.py::TestQuantizedOps::test_avg_pool3d <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5977954Z test_quantization.py::TestQuantizedOps::test_avg_pool3d_nhwc <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5978887Z test_quantization.py::TestQuantizedOps::test_batch_norm <- test/quantization/core/test_quantized_op.py PASSED [ 0%] 2023-03-31T05:48:33.5980417Z test_quantization.py::TestQuantizedOps::test_batch_norm_relu <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5981539Z test_quantization.py::TestQuantizedOps::test_cat <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5982380Z test_quantization.py::TestQuantizedOps::test_cat_nhwc <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5983496Z test_quantization.py::TestQuantizedOps::test_channel_shuffle <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5984705Z test_quantization.py::TestQuantizedOps::test_custom_module_lstm <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 1%] 2023-03-31T05:48:33.5986088Z test_quantization.py::TestQuantizedOps::test_custom_module_multi_head_attention <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 1%] 2023-03-31T05:48:33.5987289Z test_quantization.py::TestQuantizedOps::test_empty_batch <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 1%] 2023-03-31T05:48:33.5988371Z test_quantization.py::TestQuantizedOps::test_equal <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5989237Z test_quantization.py::TestQuantizedOps::test_group_norm <- test/quantization/core/test_quantized_op.py PASSED [ 1%] 2023-03-31T05:48:33.5990408Z test_quantization.py::TestQuantizedOps::test_hardswish <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 1%] 2023-03-31T05:48:33.5991520Z test_quantization.py::TestQuantizedOps::test_hardtanh <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.5992398Z test_quantization.py::TestQuantizedOps::test_instance_norm <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.5993320Z test_quantization.py::TestQuantizedOps::test_interpolate <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.5994282Z test_quantization.py::TestQuantizedOps::test_interpolate3d <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.5995208Z test_quantization.py::TestQuantizedOps::test_leaky_relu <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.5996323Z test_quantization.py::TestQuantizedOps::test_leaky_relu_observed_output <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 2%] 2023-03-31T05:48:33.6019747Z test_quantization.py::TestQuantizedOps::test_linear_bias_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 2%] 2023-03-31T05:48:33.6020766Z test_quantization.py::TestQuantizedOps::test_max_pool1d <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.6021664Z test_quantization.py::TestQuantizedOps::test_max_pool2d <- test/quantization/core/test_quantized_op.py PASSED [ 2%] 2023-03-31T05:48:33.6022669Z test_quantization.py::TestQuantizedOps::test_max_pool2d_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (cudnn is not enabled.) [ 2%] 2023-03-31T05:48:33.6023735Z test_quantization.py::TestQuantizedOps::test_max_pool2d_nhwc <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6024784Z test_quantization.py::TestQuantizedOps::test_mean <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 3%] 2023-03-31T05:48:33.6025707Z test_quantization.py::TestQuantizedOps::test_mul_scalar_relu <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6028010Z test_quantization.py::TestQuantizedOps::test_qadd_broadcast <- test/quantization/core/test_quantized_op.py [W Resize.cpp:33] Warning: An output with one or more elements was resized since it had shape [1, 1, 4, 4], which does not match the required output shape [2, 1, 4, 4]. This behavior is deprecated, and in a future PyTorch release outputs will not be resized unless they have zero elements. You can explicitly reuse an out tensor t by resizing it, inplace, to zero elements with t.resize_(0). (function _resize_output_check) 2023-03-31T05:48:33.6029356Z PASSED [ 3%] 2023-03-31T05:48:33.6030143Z test_quantization.py::TestQuantizedOps::test_qadd_relu_cudnn <- test/quantization/core/test_quantized_op.py SKIPPED (cudnn is not enabled.) [ 3%] 2023-03-31T05:48:33.6031178Z test_quantization.py::TestQuantizedOps::test_qadd_relu_cudnn_nhwc <- test/quantization/core/test_quantized_op.py SKIPPED (cudnn is not enabled.) [ 3%] 2023-03-31T05:48:33.6032184Z test_quantization.py::TestQuantizedOps::test_qadd_relu_different_qparams <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6033153Z test_quantization.py::TestQuantizedOps::test_qadd_relu_same_qparams <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6034025Z test_quantization.py::TestQuantizedOps::test_qcelu <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6034874Z test_quantization.py::TestQuantizedOps::test_qclamp <- test/quantization/core/test_quantized_op.py PASSED [ 3%] 2023-03-31T05:48:33.6035729Z test_quantization.py::TestQuantizedOps::test_qelu <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6036578Z test_quantization.py::TestQuantizedOps::test_qgelu <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6037649Z test_quantization.py::TestQuantizedOps::test_qhardsigmoid <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 4%] 2023-03-31T05:48:33.6038673Z test_quantization.py::TestQuantizedOps::test_qlayer_norm <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6039559Z test_quantization.py::TestQuantizedOps::test_qmatmul <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6040493Z test_quantization.py::TestQuantizedOps::test_qmul_broadcast <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6041449Z test_quantization.py::TestQuantizedOps::test_qmul_relu_different_qparams <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6042395Z test_quantization.py::TestQuantizedOps::test_qmul_relu_same_qparams <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6043720Z test_quantization.py::TestQuantizedOps::test_qprelu <- test/quantization/core/test_quantized_op.py PASSED [ 4%] 2023-03-31T05:48:33.6044805Z test_quantization.py::TestQuantizedOps::test_qrelu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 4%] 2023-03-31T05:48:33.6045824Z test_quantization.py::TestQuantizedOps::test_qrelu6 <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6046742Z test_quantization.py::TestQuantizedOps::test_qsoftmax <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6047664Z test_quantization.py::TestQuantizedOps::test_qsoftmax_qnnpack <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6048953Z test_quantization.py::TestQuantizedOps::test_qtanh <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 5%] 2023-03-31T05:48:33.6050109Z test_quantization.py::TestQuantizedOps::test_qthreshold <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6050969Z test_quantization.py::TestQuantizedOps::test_qtopk <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6051916Z test_quantization.py::TestQuantizedOps::test_quantized_equal <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6053267Z test_quantization.py::TestQuantizedOps::test_quantized_mean_qnnpack <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6054384Z test_quantization.py::TestQuantizedOps::test_sigmoid <- test/quantization/core/test_quantized_op.py PASSED [ 5%] 2023-03-31T05:48:33.6055541Z test_quantization.py::TestQuantizedOps::test_sigmoid_non_observed <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 5%] 2023-03-31T05:48:33.6056771Z test_quantization.py::TestQuantizedOps::test_std <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 6%] 2023-03-31T05:48:33.6057803Z test_quantization.py::TestQNNPackOps::test_adaptive_avg_pool2d <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6058728Z test_quantization.py::TestQNNPackOps::test_avg_pool2d <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6059630Z test_quantization.py::TestQNNPackOps::test_hardtanh <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6060512Z test_quantization.py::TestQNNPackOps::test_mean <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6061422Z test_quantization.py::TestQNNPackOps::test_qnnpack_add <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6063562Z test_quantization.py::TestQNNPackOps::test_qnnpack_add_broadcast <- test/quantization/core/test_quantized_op.py [W Resize.cpp:33] Warning: An output with one or more elements was resized since it had shape [1, 1, 1, 1], which does not match the required output shape [1, 3, 4, 4]. This behavior is deprecated, and in a future PyTorch release outputs will not be resized unless they have zero elements. You can explicitly reuse an out tensor t by resizing it, inplace, to zero elements with t.resize_(0). (function _resize_output_check) 2023-03-31T05:48:33.6065532Z [W Resize.cpp:33] Warning: An output with one or more elements was resized since it had shape [1, 1, 1, 1, 1], which does not match the required output shape [1, 3, 4, 4, 4]. This behavior is deprecated, and in a future PyTorch release outputs will not be resized unless they have zero elements. You can explicitly reuse an out tensor t by resizing it, inplace, to zero elements with t.resize_(0). (function _resize_output_check) 2023-03-31T05:48:33.6067270Z [W Resize.cpp:33] Warning: An output with one or more elements was resized since it had shape [1, 1, 1, 1], which does not match the required output shape [1, 3, 4, 4]. This behavior is deprecated, and in a future PyTorch release outputs will not be resized unless they have zero elements. You can explicitly reuse an out tensor t by resizing it, inplace, to zero elements with t.resize_(0). (function _resize_output_check) 2023-03-31T05:48:33.6069011Z [W Resize.cpp:33] Warning: An output with one or more elements was resized since it had shape [1, 1, 1, 1, 1], which does not match the required output shape [1, 3, 4, 4, 4]. This behavior is deprecated, and in a future PyTorch release outputs will not be resized unless they have zero elements. You can explicitly reuse an out tensor t by resizing it, inplace, to zero elements with t.resize_(0). (function _resize_output_check) 2023-03-31T05:48:33.6070002Z PASSED [ 6%] 2023-03-31T05:48:33.6070764Z test_quantization.py::TestQNNPackOps::test_qnnpack_maxpool2d <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6071582Z test_quantization.py::TestQNNPackOps::test_qnnpack_mul <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6072475Z test_quantization.py::TestQNNPackOps::test_qnnpack_relu <- test/quantization/core/test_quantized_op.py PASSED [ 6%] 2023-03-31T05:48:33.6073406Z test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid <- test/quantization/core/test_quantized_op.py PASSED [ 7%] 2023-03-31T05:48:33.6074449Z test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid_sweep <- test/quantization/core/test_quantized_op.py PASSED [ 7%] 2023-03-31T05:48:33.6075628Z test_quantization.py::TestQNNPackOps::test_qnnpack_tanh <- test/quantization/core/test_quantized_op.py PASSED [ 7%] 2023-03-31T05:48:33.6076942Z test_quantization.py::TestQuantizedLinear::test_qlinear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 7%] 2023-03-31T05:48:33.6078176Z test_quantization.py::TestQuantizedLinear::test_qlinear_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py SKIPPED (cudnn is not enabled.) [ 7%] 2023-03-31T05:48:33.6079390Z test_quantization.py::TestQuantizedLinear::test_qlinear_leaky_relu <- test/quantization/core/test_quantized_op.py PASSED [ 7%] 2023-03-31T05:48:33.6080611Z test_quantization.py::TestQuantizedLinear::test_qlinear_qnnpack_free_memory_and_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 7%] 2023-03-31T05:48:33.6081779Z test_quantization.py::TestQuantizedLinear::test_qlinear_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 7%] 2023-03-31T05:48:33.6082862Z test_quantization.py::TestQuantizedLinear::test_qlinear_tanh <- test/quantization/core/test_quantized_op.py PASSED [ 7%] 2023-03-31T05:48:33.6084142Z test_quantization.py::TestQuantizedLinear::test_qlinear_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 7%] 2023-03-31T05:48:33.6085455Z test_quantization.py::TestQuantizedLinear::test_qlinear_with_input_q_dq_qweight_dq_output_fp32 <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py PASSED [ 8%] 2023-03-31T05:48:33.6086856Z test_quantization.py::TestQuantizedConv::test_benchmark <- test/quantization/core/test_quantized_op.py SKIPPED (used for local benchmarking, comment when we want to run it) [ 8%] 2023-03-31T05:48:33.6088103Z test_quantization.py::TestQuantizedConv::test_conv_reorder_issue_onednn <- test/quantization/core/test_quantized_op.py PASSED [ 8%] 2023-03-31T05:48:33.6089197Z test_quantization.py::TestQuantizedConv::test_conv_transpose_reorder_issue_onednn <- test/quantization/core/test_quantized_op.py PASSED [ 8%] 2023-03-31T05:48:33.6090326Z test_quantization.py::TestQuantizedConv::test_qconv1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 8%] 2023-03-31T05:48:33.6091686Z test_quantization.py::TestQuantizedConv::test_qconv1d_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py SKIPPED (cudnn is not enabled.) [ 8%] 2023-03-31T05:48:33.6093131Z test_quantization.py::TestQuantizedConv::test_qconv1d_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 8%] 2023-03-31T05:48:33.6094518Z test_quantization.py::TestQuantizedConv::test_qconv1d_relu_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py SKIPPED (cudnn is not enabled.) [ 8%] 2023-03-31T05:48:33.6095719Z test_quantization.py::TestQuantizedConv::test_qconv1d_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 8%] 2023-03-31T05:48:33.6096899Z test_quantization.py::TestQuantizedConv::test_qconv2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 8%] 2023-03-31T05:48:33.6097901Z test_quantization.py::TestQuantizedConv::test_qconv2d_add <- test/quantization/core/test_quantized_op.py PASSED [ 8%] 2023-03-31T05:48:33.6098836Z test_quantization.py::TestQuantizedConv::test_qconv2d_add_relu <- test/quantization/core/test_quantized_op.py PASSED [ 9%] 2023-03-31T05:48:33.6100503Z test_quantization.py::TestQuantizedConv::test_qconv2d_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py SKIPPED (cudnn is not enabled.) [ 9%] 2023-03-31T05:48:33.6101740Z test_quantization.py::TestQuantizedConv::test_qconv2d_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 9%] 2023-03-31T05:48:33.6103022Z test_quantization.py::TestQuantizedConv::test_qconv2d_relu_cudnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py SKIPPED (cudnn is not enabled.) [ 9%] 2023-03-31T05:48:33.6104397Z test_quantization.py::TestQuantizedConv::test_qconv2d_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 9%] 2023-03-31T05:48:33.6105430Z test_quantization.py::TestQuantizedConv::test_qconv3d <- test/quantization/core/test_quantized_op.py PASSED [ 9%] 2023-03-31T05:48:33.6106376Z test_quantization.py::TestQuantizedConv::test_qconv3d_relu <- test/quantization/core/test_quantized_op.py PASSED [ 9%] 2023-03-31T05:48:33.6107491Z test_quantization.py::TestQuantizedConv::test_qconv3d_unpack <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 9%] 2023-03-31T05:48:33.6108690Z test_quantization.py::TestQuantizedConv::test_qconv_transpose1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 9%] 2023-03-31T05:48:33.6110184Z test_quantization.py::TestQuantizedConv::test_qconv_transpose2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 9%] 2023-03-31T05:48:33.6111923Z test_quantization.py::TestQuantizedConv::test_qconv_transpose3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 10%] 2023-03-31T05:48:33.6113371Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6114653Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6115910Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6117177Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6118495Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6119829Z test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6121202Z test_quantization.py::TestDynamicQuantizedOps::test_linear_prepack_fp16_numerics <- test/quantization/core/test_quantized_op.py [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6122138Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6122764Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6123796Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6124612Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6125220Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6125846Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6126481Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6127102Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6127702Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6128331Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6128968Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6129593Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6130212Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6130850Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6131443Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6132067Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6132710Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6133342Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6133803Z PASSED [ 10%] 2023-03-31T05:48:33.6134769Z test_quantization.py::TestDynamicQuantizedOps::test_qlinear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 10%] 2023-03-31T05:48:33.6135906Z test_quantization.py::TestDynamicQuantizedOps::test_qlinear_dynamic_fp16 <- test/quantization/core/test_quantized_op.py PASSED [ 10%] 2023-03-31T05:48:33.6136940Z test_quantization.py::TestDynamicQuantizedOps::test_qlinear_legacy <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6138120Z test_quantization.py::TestDynamicQuantizedOps::test_qlstmGRU <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 11%] 2023-03-31T05:48:33.6139948Z test_quantization.py::TestDynamicQuantizedOps::test_qrnncell <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qlinear_dynamic.cpp:247] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function operator()) 2023-03-31T05:48:33.6140856Z PASSED [ 11%] 2023-03-31T05:48:33.6141705Z test_quantization.py::TestComparatorOps::test_compare_tensor_scalar <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6142763Z test_quantization.py::TestComparatorOps::test_compare_tensor_tensor <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6143893Z test_quantization.py::TestPadding::test_constant_padNd <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6144902Z test_quantization.py::TestPadding::test_reflection_pad1d <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6145808Z test_quantization.py::TestPadding::test_reflection_pad2d <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6146771Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6148109Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_2d_indices <- test/quantization/core/test_quantized_op.py PASSED [ 11%] 2023-03-31T05:48:33.6149439Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2bit <- test/quantization/core/test_quantized_op.py PASSED [ 12%] 2023-03-31T05:48:33.6150623Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2bit_unpack <- test/quantization/core/test_quantized_op.py SKIPPED (Test needs Caffe2) [ 12%] 2023-03-31T05:48:33.6151822Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2d_indices <- test/quantization/core/test_quantized_op.py PASSED [ 12%] 2023-03-31T05:48:33.6152943Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_4bit <- test/quantization/core/test_quantized_op.py PASSED [ 12%] 2023-03-31T05:48:33.6154113Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_4bit_unpack <- test/quantization/core/test_quantized_op.py SKIPPED (Test needs Caffe2) [ 12%] 2023-03-31T05:48:33.6155180Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_byte <- test/quantization/core/test_quantized_op.py PASSED [ 12%] 2023-03-31T05:48:33.6156313Z test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_byte_unpack <- test/quantization/core/test_quantized_op.py SKIPPED (Test needs Caffe2) [ 12%] 2023-03-31T05:48:33.6157584Z test_quantization.py::TestQuantizedFunctionalOps::test_conv1d_api <- test/quantization/core/test_quantized_functional.py PASSED [ 12%] 2023-03-31T05:48:33.6158715Z test_quantization.py::TestQuantizedFunctionalOps::test_conv2d_api <- test/quantization/core/test_quantized_functional.py PASSED [ 12%] 2023-03-31T05:48:33.6159891Z test_quantization.py::TestQuantizedFunctionalOps::test_conv3d_api <- test/quantization/core/test_quantized_functional.py PASSED [ 12%] 2023-03-31T05:48:33.6161010Z test_quantization.py::TestQuantizedFunctionalOps::test_grid_sample <- test/quantization/core/test_quantized_functional.py PASSED [ 13%] 2023-03-31T05:48:33.6162129Z test_quantization.py::TestQuantizedFunctionalOps::test_relu_api <- test/quantization/core/test_quantized_functional.py PASSED [ 13%] 2023-03-31T05:48:33.6163704Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 13%] 2023-03-31T05:48:33.6164942Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 13%] 2023-03-31T05:48:33.6166104Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cuda <- test/quantization/core/test_workflow_ops.py SKIPPED (No gpu is not available.) [ 13%] 2023-03-31T05:48:33.6167481Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (temporarily disable the test) [ 13%] 2023-03-31T05:48:33.6168689Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor_cachemask_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 13%] 2023-03-31T05:48:33.6169963Z test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor_cachemask_cuda <- test/quantization/core/test_workflow_ops.py SKIPPED (No gpu is not available.) [ 13%] 2023-03-31T05:48:33.6171211Z test_quantization.py::TestFakeQuantizeOps::test_fake_quant_control <- test/quantization/core/test_workflow_ops.py PASSED [ 13%] 2023-03-31T05:48:33.6172247Z test_quantization.py::TestFakeQuantizeOps::test_fake_quant_per_channel_qparam_range <- test/quantization/core/test_workflow_ops.py PASSED [ 13%] 2023-03-31T05:48:33.6173405Z test_quantization.py::TestFakeQuantizeOps::test_fake_quant_preserves_qparam_shapes_for_activations <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6174567Z test_quantization.py::TestFakeQuantizeOps::test_fixed_qparams_fq_module <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6176049Z test_quantization.py::TestFakeQuantizeOps::test_forward_backward_per_tensor_with_amp <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6177305Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6178432Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_cachemask_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6179671Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_cachemask_cuda <- test/quantization/core/test_workflow_ops.py SKIPPED (No gpu is not available.) [ 14%] 2023-03-31T05:48:33.6180885Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_half_precision_numerics <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6182111Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6183155Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 14%] 2023-03-31T05:48:33.6184398Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cuda <- test/quantization/core/test_workflow_ops.py SKIPPED (No gpu is not available.) [ 14%] 2023-03-31T05:48:33.6185533Z test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_half_precision_numerics <- test/quantization/core/test_workflow_ops.py PASSED [ 15%] 2023-03-31T05:48:33.6186592Z test_quantization.py::TestFakeQuantizeOps::test_fq_module_per_tensor <- test/quantization/core/test_workflow_ops.py PASSED [ 15%] 2023-03-31T05:48:33.6187584Z test_quantization.py::TestFakeQuantizeOps::test_fq_serializable_per_tensor <- test/quantization/core/test_workflow_ops.py PASSED [ 15%] 2023-03-31T05:48:33.6189167Z test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cpu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 15%] 2023-03-31T05:48:33.6190610Z test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cuda <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (No gpu is not available.) [ 15%] 2023-03-31T05:48:33.6191930Z test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_tensor_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 15%] 2023-03-31T05:48:33.6193221Z test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_tensor_cuda <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (No gpu is not available.) [ 15%] 2023-03-31T05:48:33.6194321Z test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_channel_cpu <- test/quantization/core/test_workflow_ops.py PASSED [ 15%] 2023-03-31T05:48:33.6195671Z test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_channel_cuda <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (No gpu is not available.) [ 15%] 2023-03-31T05:48:33.6197218Z test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cpu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 15%] 2023-03-31T05:48:33.6198607Z test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cuda <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (No gpu is not available.) [ 16%] 2023-03-31T05:48:33.6199938Z test_quantization.py::TestFakeQuantizeOps::test_numerical_consistency_per_channel <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6201085Z test_quantization.py::TestFakeQuantizeOps::test_numerical_consistency_per_tensor <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6202388Z test_quantization.py::TestFusedObsFakeQuant::test_fused_backward_op_fake_quant_off <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6204260Z test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_backward_op <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6204924Z test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_moving_avg <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6205557Z test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_moving_avg_per_channel <- test/quantization/core/test_workflow_ops.py PASSED [ 16%] 2023-03-31T05:48:33.6206135Z test_quantization.py::TestQuantizedTensor::test_bfp16_quantize <- test/quantization/core/test_quantized_tensor.py PASSED [ 16%] 2023-03-31T05:48:33.6206874Z test_quantization.py::TestQuantizedTensor::test_choose_qparams <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py SKIPPED (this is broken without changes to any relevant code, we need to remove hypothesis testing in CI) [ 16%] 2023-03-31T05:48:33.6207567Z test_quantization.py::TestQuantizedTensor::test_choose_qparams_optimized <- test/quantization/core/test_quantized_tensor.py PASSED [ 16%] 2023-03-31T05:48:33.6208128Z test_quantization.py::TestQuantizedTensor::test_clone <- test/quantization/core/test_quantized_tensor.py PASSED [ 16%] 2023-03-31T05:48:33.6208747Z test_quantization.py::TestQuantizedTensor::test_compare_per_channel_device_numerics <- test/quantization/core/test_quantized_tensor.py SKIPPED (CUDA is not available) [ 17%] 2023-03-31T05:48:33.6209429Z test_quantization.py::TestQuantizedTensor::test_compare_per_tensor_device_numerics <- test/quantization/core/test_quantized_tensor.py SKIPPED (CUDA is not available) [ 17%] 2023-03-31T05:48:33.6210115Z test_quantization.py::TestQuantizedTensor::test_cuda_quantization_does_not_pin_memory <- test/quantization/core/test_quantized_tensor.py SKIPPED (CUDA is not available) [ 17%] 2023-03-31T05:48:33.6210760Z test_quantization.py::TestQuantizedTensor::test_decomposed_dequantize_per_channel <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6211378Z test_quantization.py::TestQuantizedTensor::test_decomposed_dequantize_per_tensor <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6211975Z test_quantization.py::TestQuantizedTensor::test_decomposed_dynamic_quant_pattern <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6212589Z test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_channel <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6213192Z test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_tensor <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6213785Z test_quantization.py::TestQuantizedTensor::test_dequantize_fp16_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 17%] 2023-03-31T05:48:33.6214388Z test_quantization.py::TestQuantizedTensor::test_dequantize_fp16_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 17%] 2023-03-31T05:48:33.6214979Z test_quantization.py::TestQuantizedTensor::test_fp16_saturate_op <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6215551Z test_quantization.py::TestQuantizedTensor::test_jit_serialization <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6216149Z test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_creation_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6216791Z test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_creation_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 18%] 2023-03-31T05:48:33.6217428Z test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_to_memory_format <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6218246Z test_quantization.py::TestQuantizedTensor::test_per_channel_to_device <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 18%] 2023-03-31T05:48:33.6218877Z test_quantization.py::TestQuantizedTensor::test_per_tensor_qtensor_to_memory_format <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6219493Z test_quantization.py::TestQuantizedTensor::test_per_tensor_to_device <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 18%] 2023-03-31T05:48:33.6220103Z test_quantization.py::TestQuantizedTensor::test_pickle_checkpoint_qtensor <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6220684Z test_quantization.py::TestQuantizedTensor::test_qscheme_pickle <- test/quantization/core/test_quantized_tensor.py PASSED [ 18%] 2023-03-31T05:48:33.6221276Z test_quantization.py::TestQuantizedTensor::test_qtensor_channel_float_assignment <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6221852Z test_quantization.py::TestQuantizedTensor::test_qtensor_copy <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6222396Z test_quantization.py::TestQuantizedTensor::test_qtensor_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6222951Z test_quantization.py::TestQuantizedTensor::test_qtensor_creation <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6223643Z test_quantization.py::TestQuantizedTensor::test_qtensor_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 19%] 2023-03-31T05:48:33.6224211Z test_quantization.py::TestQuantizedTensor::test_qtensor_dtypes <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6224783Z test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6225381Z test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel_nhwc <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6225971Z test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_tensor <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6226544Z test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_tensor_nhwc <- test/quantization/core/test_quantized_tensor.py PASSED [ 19%] 2023-03-31T05:48:33.6227148Z test_quantization.py::TestQuantizedTensor::test_qtensor_float_assignment <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6227741Z test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6228369Z test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 20%] 2023-03-31T05:48:33.6228984Z test_quantization.py::TestQuantizedTensor::test_qtensor_index_select_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6229602Z test_quantization.py::TestQuantizedTensor::test_qtensor_index_select_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 20%] 2023-03-31T05:48:33.6230207Z test_quantization.py::TestQuantizedTensor::test_qtensor_int_repr <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6230784Z test_quantization.py::TestQuantizedTensor::test_qtensor_legacy_new_failure <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6231356Z test_quantization.py::TestQuantizedTensor::test_qtensor_load_save <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6231922Z test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cpu <- test/quantization/core/test_quantized_tensor.py PASSED [ 20%] 2023-03-31T05:48:33.6232640Z test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cuda <- test/quantization/core/test_quantized_tensor.py SKIPPED (No gpu is available.) [ 20%] 2023-03-31T05:48:33.6233321Z test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_load_save <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6233923Z test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_permute <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6234485Z test_quantization.py::TestQuantizedTensor::test_qtensor_permute <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6235051Z test_quantization.py::TestQuantizedTensor::test_qtensor_quant_dequant <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6235652Z test_quantization.py::TestQuantizedTensor::test_qtensor_quantize_per_channel <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6236237Z test_quantization.py::TestQuantizedTensor::test_qtensor_reshape <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6236784Z test_quantization.py::TestQuantizedTensor::test_qtensor_resize <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6237363Z test_quantization.py::TestQuantizedTensor::test_qtensor_sub_byte_aligned_cols <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6238014Z test_quantization.py::TestQuantizedTensor::test_qtensor_sub_byte_not_aligned_cols <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6238591Z test_quantization.py::TestQuantizedTensor::test_qtensor_unsqueeze <- test/quantization/core/test_quantized_tensor.py PASSED [ 21%] 2023-03-31T05:48:33.6239140Z test_quantization.py::TestQuantizedTensor::test_qtensor_view <- test/quantization/core/test_quantized_tensor.py PASSED [ 22%] 2023-03-31T05:48:33.6239752Z test_quantization.py::TestQuantizedTensor::test_quant_pin_memory <- test/quantization/core/test_quantized_tensor.py SKIPPED (CUDA is not available) [ 22%] 2023-03-31T05:48:33.6240382Z test_quantization.py::TestQuantizedTensor::test_quantize_per_channel_float_qparams <- test/quantization/core/test_quantized_tensor.py PASSED [ 22%] 2023-03-31T05:48:33.6240980Z test_quantization.py::TestQuantizedTensor::test_quantize_per_channel_sub_byte <- test/quantization/core/test_quantized_tensor.py PASSED [ 22%] 2023-03-31T05:48:33.6241521Z test_quantization.py::TestQuantizedTensor::test_repeat <- test/quantization/core/test_quantized_tensor.py PASSED [ 22%] 2023-03-31T05:48:33.6242083Z test_quantization.py::TestQuantizedTensor::test_torch_qtensor_deepcopy <- test/quantization/core/test_quantized_tensor.py PASSED [ 22%] 2023-03-31T05:48:33.6242652Z test_quantization.py::TestFakeQuantize::test_fq_module_per_channel <- test/quantization/core/test_workflow_module.py PASSED [ 22%] 2023-03-31T05:48:33.6243522Z test_quantization.py::TestFakeQuantize::test_fq_serializable_per_channel <- test/quantization/core/test_workflow_module.py PASSED [ 22%] 2023-03-31T05:48:33.6244097Z test_quantization.py::TestFakeQuantize::test_quant_min_max_override <- test/quantization/core/test_workflow_module.py PASSED [ 22%] 2023-03-31T05:48:33.6244688Z test_quantization.py::TestObserver::test_histogram_observer_consistent_buffer_shape <- test/quantization/core/test_workflow_module.py PASSED [ 22%] 2023-03-31T05:48:33.6245293Z test_quantization.py::TestObserver::test_histogram_observer_save_load_state_dict <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6245868Z test_quantization.py::TestObserver::test_memoryless_minmaxobserver <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6246447Z test_quantization.py::TestObserver::test_memoryless_perchannelminmaxobserver <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6247263Z test_quantization.py::TestObserver::test_observer_qparams_respects_device_affinity <- test/quantization/core/test_workflow_module.py SKIPPED (multi-GPU not supported) [ 23%] 2023-03-31T05:48:33.6247937Z test_quantization.py::TestObserver::test_observer_scriptable <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6248492Z test_quantization.py::TestObserver::test_per_channel_observers <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6249043Z test_quantization.py::TestObserver::test_per_tensor_observers <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6249585Z test_quantization.py::TestObserver::test_save_load_state_dict_script <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6250301Z test_quantization.py::TestObserver::test_state_dict_respects_device_affinity <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py SKIPPED (multi-GPU not supported) [ 23%] 2023-03-31T05:48:33.6250939Z test_quantization.py::TestObserver::test_zero_numel <- test/quantization/core/test_workflow_module.py PASSED [ 23%] 2023-03-31T05:48:33.6251499Z test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6252088Z test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d_serialization <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6252681Z test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6253281Z test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d_serialization <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6253883Z test_quantization.py::TestStaticQuantizedModule::test_channel_shuffle <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6254550Z test_quantization.py::TestStaticQuantizedModule::test_conv1d_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 24%] 2023-03-31T05:48:33.6255275Z test_quantization.py::TestStaticQuantizedModule::test_conv1d_relu_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 24%] 2023-03-31T05:48:33.6255899Z test_quantization.py::TestStaticQuantizedModule::test_conv2d_add <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6256477Z test_quantization.py::TestStaticQuantizedModule::test_conv2d_add_relu <- test/quantization/core/test_quantized_module.py PASSED [ 24%] 2023-03-31T05:48:33.6257132Z test_quantization.py::TestStaticQuantizedModule::test_conv2d_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 24%] 2023-03-31T05:48:33.6257847Z test_quantization.py::TestStaticQuantizedModule::test_conv2d_relu_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 25%] 2023-03-31T05:48:33.6258477Z test_quantization.py::TestStaticQuantizedModule::test_conv3d_api <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6259053Z test_quantization.py::TestStaticQuantizedModule::test_conv3d_relu_api <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6259623Z test_quantization.py::TestStaticQuantizedModule::test_dropout <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6260198Z test_quantization.py::TestStaticQuantizedModule::test_dropout_serialization <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6260769Z test_quantization.py::TestStaticQuantizedModule::test_elu <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6261435Z test_quantization.py::TestStaticQuantizedModule::test_embedding_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py PASSED [ 25%] 2023-03-31T05:48:33.6262296Z test_quantization.py::TestStaticQuantizedModule::test_embedding_bag_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantization.py PASSED [ 25%] 2023-03-31T05:48:33.6262921Z test_quantization.py::TestStaticQuantizedModule::test_group_norm <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6263590Z test_quantization.py::TestStaticQuantizedModule::test_hard_swish <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6264170Z test_quantization.py::TestStaticQuantizedModule::test_instance_norm <- test/quantization/core/test_quantized_module.py PASSED [ 25%] 2023-03-31T05:48:33.6264745Z test_quantization.py::TestStaticQuantizedModule::test_layer_norm <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6265310Z test_quantization.py::TestStaticQuantizedModule::test_leaky_relu <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6265968Z test_quantization.py::TestStaticQuantizedModule::test_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 26%] 2023-03-31T05:48:33.6266598Z test_quantization.py::TestStaticQuantizedModule::test_linear_leaky_relu <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6267266Z test_quantization.py::TestStaticQuantizedModule::test_linear_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 26%] 2023-03-31T05:48:33.6267876Z test_quantization.py::TestStaticQuantizedModule::test_linear_tanh <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6268446Z test_quantization.py::TestStaticQuantizedModule::test_pool_api <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6269010Z test_quantization.py::TestStaticQuantizedModule::test_prelu <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6269593Z test_quantization.py::TestStaticQuantizedModule::test_quant_dequant_api <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6270148Z test_quantization.py::TestStaticQuantizedModule::test_relu <- test/quantization/core/test_quantized_module.py PASSED [ 26%] 2023-03-31T05:48:33.6270702Z test_quantization.py::TestStaticQuantizedModule::test_sigmoid <- test/quantization/core/test_quantized_module.py PASSED [ 27%] 2023-03-31T05:48:33.6271367Z test_quantization.py::TestDynamicQuantizedModule::test_cell_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 27%] 2023-03-31T05:48:33.6272339Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6273046Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6273546Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6274049Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6274555Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6275134Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6275701Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6276184Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6276684Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6277183Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6277680Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6278177Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6278652Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6279158Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6279657Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6280152Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6280466Z PASSED [ 27%] 2023-03-31T05:48:33.6281265Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6281955Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6282449Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6282948Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6283567Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6284073Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6284570Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6285059Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6285649Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6286198Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6286701Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6287199Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6287694Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6288179Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6288677Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6289171Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6289497Z PASSED [ 27%] 2023-03-31T05:48:33.6290036Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 27%] 2023-03-31T05:48:33.6291030Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose1d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6291744Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6292243Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6292740Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6293212Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6293708Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6294212Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6294708Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6295200Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6295675Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6296251Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6296797Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6297290Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6297768Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6298261Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6298758Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6299088Z PASSED [ 27%] 2023-03-31T05:48:33.6299886Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6300583Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6301082Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6301573Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6302075Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6302557Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6303042Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6303634Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6304125Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6304622Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6305100Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6305596Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6306094Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6306588Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6307208Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6307685Z [W qconv_dynamic.cpp:82] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function apply_dynamic) 2023-03-31T05:48:33.6308013Z PASSED [ 27%] 2023-03-31T05:48:33.6308580Z test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 27%] 2023-03-31T05:48:33.6309314Z test_quantization.py::TestDynamicQuantizedModule::test_gru_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 27%] 2023-03-31T05:48:33.6310023Z test_quantization.py::TestDynamicQuantizedModule::test_linear_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 27%] 2023-03-31T05:48:33.6310739Z test_quantization.py::TestDynamicQuantizedModule::test_lstm_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 28%] 2023-03-31T05:48:33.6311421Z test_quantization.py::TestReferenceQuantizedModule::test_linear_decomposed_weight_custom_qmin_qmax <- test/quantization/core/test_quantized_module.py PASSED [ 28%] 2023-03-31T05:48:33.6312034Z test_quantization.py::TestReferenceQuantizedModule::test_rnn <- test/quantization/core/test_quantized_module.py PASSED [ 28%] 2023-03-31T05:48:33.6312602Z test_quantization.py::TestReferenceQuantizedModule::test_rnn_cell <- test/quantization/core/test_quantized_module.py PASSED [ 28%] 2023-03-31T05:48:33.6313179Z test_quantization.py::TestReferenceQuantizedModule::test_sparse <- test/quantization/core/test_quantized_module.py PASSED [ 28%] 2023-03-31T05:48:33.6313785Z test_quantization.py::TestRecordHistogramObserver::test_observer_scriptable <- test/quantization/core/test_workflow_module.py PASSED [ 28%] 2023-03-31T05:48:33.6314391Z test_quantization.py::TestRecordHistogramObserver::test_record_observer <- test/quantization/core/test_workflow_module.py PASSED [ 28%] 2023-03-31T05:48:33.6314961Z test_quantization.py::TestHistogramObserver::test_histogram_observer <- test/quantization/core/test_workflow_module.py PASSED [ 28%] 2023-03-31T05:48:33.6315558Z test_quantization.py::TestHistogramObserver::test_histogram_observer_against_reference <- test/quantization/core/test_workflow_module.py PASSED [ 28%] 2023-03-31T05:48:33.6316172Z test_quantization.py::TestHistogramObserver::test_histogram_observer_correct_numel <- test/quantization/core/test_workflow_module.py PASSED [ 28%] 2023-03-31T05:48:33.6316785Z test_quantization.py::TestHistogramObserver::test_histogram_observer_extreme_inputs <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6317377Z test_quantization.py::TestHistogramObserver::test_histogram_observer_one_sided <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6317977Z test_quantization.py::TestHistogramObserver::test_histogram_observer_same_inputs <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6318560Z test_quantization.py::TestHistogramObserver::test_observer_scriptable <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6319251Z test_quantization.py::TestDistributed::test_device_affinity <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py SKIPPED (multi-GPU not supported) [ 29%] 2023-03-31T05:48:33.6319893Z test_quantization.py::TestDistributed::test_fake_quant_preserves_buffers <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6320452Z test_quantization.py::TestDistributed::test_observers_preserve_buffers <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6321182Z test_quantization.py::TestDistributed::test_qat_convbn_fused_syncbn_replacement <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6321787Z test_quantization.py::TestDistributed::test_qat_data_parallel <- test/quantization/core/test_workflow_module.py SKIPPED (multi-GPU not supported) [ 29%] 2023-03-31T05:48:33.6322363Z test_quantization.py::TestDistributed::test_syncbn_preserves_qconfig <- test/quantization/core/test_workflow_module.py PASSED [ 29%] 2023-03-31T05:48:33.6322956Z test_quantization.py::TestFusedObsFakeQuantModule::test_compare_fused_obs_fq_oss_module <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6323838Z test_quantization.py::TestFusedObsFakeQuantModule::test_default_fused_qat_config <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6324465Z test_quantization.py::TestFusedObsFakeQuantModule::test_embedding_bag_qat_config <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6325092Z test_quantization.py::TestFusedObsFakeQuantModule::test_embedding_qat_config <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6325684Z test_quantization.py::TestFusedObsFakeQuantModule::test_fused_mod_per_channel <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6326391Z test_quantization.py::TestFusedObsFakeQuantModule::test_fused_mod_reduce_range <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6327042Z test_quantization.py::TestFusedObsFakeQuantModule::test_fused_obs_fq_module <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6327659Z test_quantization.py::TestFusedObsFakeQuantModule::test_fused_obs_fq_moving_avg_module <- test/quantization/core/test_workflow_module.py PASSED [ 30%] 2023-03-31T05:48:33.6328241Z test_quantization.py::TestBackendConfig::test_backend_config_from_dict <- test/quantization/core/test_backend_config.py PASSED [ 30%] 2023-03-31T05:48:33.6328841Z test_quantization.py::TestBackendConfig::test_backend_config_set_backend_pattern_config <- test/quantization/core/test_backend_config.py PASSED [ 30%] 2023-03-31T05:48:33.6329428Z test_quantization.py::TestBackendConfig::test_backend_config_set_name <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6329984Z test_quantization.py::TestBackendConfig::test_backend_config_to_dict <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6330547Z test_quantization.py::TestBackendConfig::test_backend_op_config_add_dtype_config <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6331131Z test_quantization.py::TestBackendConfig::test_backend_op_config_from_dict <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6331727Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_extra_inputs_getter <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6332333Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_fused_module <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6332903Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_fuser_method <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6333503Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_input_type_to_index <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6334135Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_num_tensor_args_to_observation_type <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6334754Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_observation_type <- test/quantization/core/test_backend_config.py PASSED [ 31%] 2023-03-31T05:48:33.6335507Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_qat_module <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6336174Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_reference_quantized_module <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6336912Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_module <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6337690Z test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_node_getter <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6338272Z test_quantization.py::TestBackendConfig::test_backend_op_config_to_dict <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6338819Z test_quantization.py::TestBackendConfig::test_dtype_config_from_dict <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6339377Z test_quantization.py::TestBackendConfig::test_dtype_config_to_dict <- test/quantization/core/test_backend_config.py PASSED [ 32%] 2023-03-31T05:48:33.6339933Z test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_complex_args <- test/quantization/core/test_utils.py PASSED [ 32%] 2023-03-31T05:48:33.6340481Z test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_default_kwargs <- test/quantization/core/test_utils.py PASSED [ 32%] 2023-03-31T05:48:33.6341007Z test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_simple <- test/quantization/core/test_utils.py PASSED [ 32%] 2023-03-31T05:48:33.6341548Z test_quantization.py::TestUtils::test_quantize_weight_clamping_per_channel <- test/quantization/core/test_utils.py PASSED [ 33%] 2023-03-31T05:48:33.6342097Z test_quantization.py::TestUtils::test_quantize_weight_clamping_per_tensor <- test/quantization/core/test_utils.py PASSED [ 33%] 2023-03-31T05:48:33.6342649Z test_quantization.py::TestQuantizationDocs::test_quantization_doc_custom <- test/quantization/core/test_docs.py PASSED [ 33%] 2023-03-31T05:48:33.6343182Z test_quantization.py::TestQuantizationDocs::test_quantization_doc_fx <- test/quantization/core/test_docs.py PASSED [ 33%] 2023-03-31T05:48:33.6343811Z test_quantization.py::TestQuantizationDocs::test_quantization_doc_ptdq <- test/quantization/core/test_docs.py PASSED [ 33%] 2023-03-31T05:48:33.6344357Z test_quantization.py::TestQuantizationDocs::test_quantization_doc_ptsq <- test/quantization/core/test_docs.py PASSED [ 33%] 2023-03-31T05:48:33.6344897Z test_quantization.py::TestQuantizationDocs::test_quantization_doc_qat <- test/quantization/core/test_docs.py PASSED [ 33%] 2023-03-31T05:48:33.6345460Z test_quantization.py::TestQuantizeEagerPTQStatic::test_activations <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 33%] 2023-03-31T05:48:33.6346093Z test_quantization.py::TestQuantizeEagerPTQStatic::test_activations_in_non_leaf_module_list <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 33%] 2023-03-31T05:48:33.6346758Z test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_fails_early <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 33%] 2023-03-31T05:48:33.6347432Z test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_qconfig_none <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 33%] 2023-03-31T05:48:33.6348049Z test_quantization.py::TestQuantizeEagerPTQStatic::test_custom_module_class <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6348645Z test_quantization.py::TestQuantizeEagerPTQStatic::test_dequant_stub <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6349342Z test_quantization.py::TestQuantizeEagerPTQStatic::test_forward_hooks_preserved <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 34%] 2023-03-31T05:48:33.6350081Z test_quantization.py::TestQuantizeEagerPTQStatic::test_manual <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6350755Z test_quantization.py::TestQuantizeEagerPTQStatic::test_mha_batch_first_attr_is_copied_in_prepare <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6351373Z test_quantization.py::TestQuantizeEagerPTQStatic::test_nested1 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6351942Z test_quantization.py::TestQuantizeEagerPTQStatic::test_nested2 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6352506Z test_quantization.py::TestQuantizeEagerPTQStatic::test_nested3 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6353094Z test_quantization.py::TestQuantizeEagerPTQStatic::test_normalization <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6353687Z test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 34%] 2023-03-31T05:48:33.6354317Z test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding_bag <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6354976Z test_quantization.py::TestQuantizeEagerPTQStatic::test_quantwrapper_attaches_qconfig_to_dequant <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6355599Z test_quantization.py::TestQuantizeEagerPTQStatic::test_resnet_base <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6356191Z test_quantization.py::TestQuantizeEagerPTQStatic::test_save_load_state_dict <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6356785Z test_quantization.py::TestQuantizeEagerPTQStatic::test_single_layer <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6357370Z test_quantization.py::TestQuantizeEagerPTQStatic::test_skip_quant <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6357951Z test_quantization.py::TestQuantizeEagerPTQStatic::test_two_layers <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6358534Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_embedding_ops_dynamic <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6359157Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_forward_hooks_preserved <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6359770Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_linear_relu_fusion <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 35%] 2023-03-31T05:48:33.6360358Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested1 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6360925Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested2 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6361500Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested3 <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6362121Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_per_channel_linear_quantize <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6362732Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_quantized_rnn <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6363533Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_quantized_rnn_cell <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6364133Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_single_layer <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6364725Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_two_layers <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6365519Z test_quantization.py::TestQuantizeEagerPTQDynamic::test_type_match_rule <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6366079Z test_quantization.py::TestQuantizeEagerOps::test_conv_1d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 36%] 2023-03-31T05:48:33.6366619Z test_quantization.py::TestQuantizeEagerOps::test_conv_2d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6367152Z test_quantization.py::TestQuantizeEagerOps::test_conv_3d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6367712Z test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_1d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6368268Z test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_2d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6368845Z test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_3d <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6369422Z test_quantization.py::TestQuantizeEagerOps::test_functional_module <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6370090Z test_quantization.py::TestQuantizeEagerOps::test_int16_reference_module <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 37%] 2023-03-31T05:48:33.6370691Z test_quantization.py::TestQuantizeEagerOps::test_leaky_relu <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6371241Z test_quantization.py::TestQuantizeEagerOps::test_linear <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6371780Z test_quantization.py::TestQuantizeEagerOps::test_relu <- test/quantization/eager/test_quantize_eager_ptq.py PASSED [ 37%] 2023-03-31T05:48:33.6372364Z test_quantization.py::TestQuantizeEagerQAT::test_add_scalar_uses_input_qparams <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6372933Z test_quantization.py::TestQuantizeEagerQAT::test_conv_linear <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6373492Z test_quantization.py::TestQuantizeEagerQAT::test_conv_linear_symm <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6374077Z test_quantization.py::TestQuantizeEagerQAT::test_defused_embedding_bag_linear <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6374643Z test_quantization.py::TestQuantizeEagerQAT::test_dropout <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6375197Z test_quantization.py::TestQuantizeEagerQAT::test_dynamic_qat_linear <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6375781Z test_quantization.py::TestQuantizeEagerQAT::test_embedding_bag_linear <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6376375Z test_quantization.py::TestQuantizeEagerQAT::test_embedding_qat_qconfig_equal <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6376960Z test_quantization.py::TestQuantizeEagerQAT::test_eval_only_fake_quant <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 38%] 2023-03-31T05:48:33.6377631Z test_quantization.py::TestQuantizeEagerQAT::test_forward_hooks_preserved <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 38%] 2023-03-31T05:48:33.6378225Z test_quantization.py::TestQuantizeEagerQAT::test_manual <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6378809Z test_quantization.py::TestQuantizeEagerQAT::test_mul_scalar_uses_input_qparams <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6379567Z test_quantization.py::TestQuantizeEagerQAT::test_qat_embedding_bag_errors <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 39%] 2023-03-31T05:48:33.6380248Z test_quantization.py::TestQuantizeEagerQAT::test_train_save_load_eval <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6380843Z test_quantization.py::TestQuantizeEagerQATNumerics::test_conv_bn_folded_vs_unfolded <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6381450Z test_quantization.py::TestQuantizeEagerQATNumerics::test_conv_bn_relu <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6382051Z test_quantization.py::TestQuantizeEagerQATNumerics::test_fixed_qparam_ops <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6382648Z test_quantization.py::TestQuantizeEagerQATNumerics::test_leaky_relu <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 39%] 2023-03-31T05:48:33.6383422Z test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_numerics <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 39%] 2023-03-31T05:48:33.6384183Z test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_symm_numerics <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 39%] 2023-03-31T05:48:33.6384936Z test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_workflow <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 40%] 2023-03-31T05:48:33.6385572Z test_quantization.py::TestQuantizeEagerQATNumerics::test_relu <- test/quantization/eager/test_quantize_eager_qat.py PASSED [ 40%] 2023-03-31T05:48:33.6386136Z test_quantization.py::TestFuseEager::test_forward_hooks_preserved <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6386679Z test_quantization.py::TestFuseEager::test_fuse_function_customization <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6387208Z test_quantization.py::TestFuseEager::test_fuse_module_eval <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6387725Z test_quantization.py::TestFuseEager::test_fuse_module_train <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6388250Z test_quantization.py::TestFuseEager::test_fusion_conv_with_bias <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6388776Z test_quantization.py::TestFuseEager::test_fusion_convtranspose_bn_eval <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6389311Z test_quantization.py::TestFuseEager::test_fusion_linear_bn_eval <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6389850Z test_quantization.py::TestFuseEager::test_fusion_sequential_model_eval <- test/quantization/eager/test_fuse_eager.py PASSED [ 40%] 2023-03-31T05:48:33.6390407Z test_quantization.py::TestFuseEager::test_fusion_sequential_model_train <- test/quantization/eager/test_fuse_eager.py PASSED [ 41%] 2023-03-31T05:48:33.6390983Z test_quantization.py::TestModelNumericsEager::test_fake_quant_true_quant_compare <- test/quantization/eager/test_model_numerics.py PASSED [ 41%] 2023-03-31T05:48:33.6391595Z test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_channel <- test/quantization/eager/test_model_numerics.py PASSED [ 41%] 2023-03-31T05:48:33.6392206Z test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_tensor <- test/quantization/eager/test_model_numerics.py PASSED [ 41%] 2023-03-31T05:48:33.6392832Z test_quantization.py::TestModelNumericsEager::test_weight_only_activation_only_fakequant <- test/quantization/eager/test_model_numerics.py PASSED [ 41%] 2023-03-31T05:48:33.6393613Z test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_conv_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6394428Z test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_functional_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6395195Z test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6395948Z test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6396697Z test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_lstm_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6397433Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_conv_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 41%] 2023-03-31T05:48:33.6398180Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_functional_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6398933Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_linear_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6399673Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_linear_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6400430Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_lstm_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6401149Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_partial <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6401896Z test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_submodule_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6402635Z test_quantization.py::TestNumericSuiteEager::test_compare_weights_conv_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6403649Z test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6404405Z test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_static <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6405123Z test_quantization.py::TestNumericSuiteEager::test_compare_weights_lstm_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 42%] 2023-03-31T05:48:33.6405749Z test_quantization.py::TestNumericSuiteEager::test_mobilenet_v2 <- test/quantization/eager/test_numeric_suite_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6406310Z test_quantization.py::TestNumericSuiteEager::test_mobilenet_v3 <- test/quantization/eager/test_numeric_suite_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6406965Z test_quantization.py::TestNumericSuiteEager::test_output_logger <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 43%] 2023-03-31T05:48:33.6407852Z test_quantization.py::TestNumericSuiteEager::test_shadow_logger <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 43%] 2023-03-31T05:48:33.6408433Z test_quantization.py::TestEqualizeEager::test_converged <- test/quantization/eager/test_equalize_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6408995Z test_quantization.py::TestEqualizeEager::test_cross_layer_equalization <- test/quantization/eager/test_equalize_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6409541Z test_quantization.py::TestEqualizeEager::test_equalize <- test/quantization/eager/test_equalize_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6410090Z test_quantization.py::TestEqualizeEager::test_equalize_fused_convrelu <- test/quantization/eager/test_equalize_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6410657Z test_quantization.py::TestEqualizeEager::test_equalize_fused_linearrelu <- test/quantization/eager/test_equalize_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6411232Z test_quantization.py::TestBiasCorrectionEager::test_conv_chain <- test/quantization/eager/test_bias_correction_eager.py PASSED [ 43%] 2023-03-31T05:48:33.6411810Z test_quantization.py::TestBiasCorrectionEager::test_linear_chain <- test/quantization/eager/test_bias_correction_eager.py PASSED [ 44%] 2023-03-31T05:48:33.6412517Z test_quantization.py::TestFuseFx::test_fuse_addtional_fuser_method <- test/quantization/fx/test_quantize_fx.py SKIPPED (Temporarily skipping the test case, will enable after the simplepattern format is supported) [ 44%] 2023-03-31T05:48:33.6413143Z test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_by_default <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6413680Z test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6414214Z test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_onednn <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6414728Z test_quantization.py::TestFuseFx::test_fuse_conv_bn_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6415230Z test_quantization.py::TestFuseFx::test_fuse_convtranspose_bn_eval <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6415744Z test_quantization.py::TestFuseFx::test_fuse_custom_pattern <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6416248Z test_quantization.py::TestFuseFx::test_fuse_linear_bn_eval <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6416773Z test_quantization.py::TestFuseFx::test_fuse_linear_bn_leaky_relu_onednn <- test/quantization/fx/test_quantize_fx.py PASSED [ 44%] 2023-03-31T05:48:33.6417306Z test_quantization.py::TestFuseFx::test_fuse_linear_tanh_for_onednn_backend <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6417823Z test_quantization.py::TestFuseFx::test_fuse_module_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6418349Z test_quantization.py::TestFuseFx::test_fusion_pattern_with_matchallnode <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6418903Z test_quantization.py::TestFuseFx::test_fusion_pattern_with_multiple_inputs <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6419451Z test_quantization.py::TestFuseFx::test_linear_bn_leaky_relu_not_fused_by_default <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6419997Z test_quantization.py::TestFuseFx::test_linear_tanh_not_fused_by_default <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6420523Z test_quantization.py::TestFuseFx::test_problematic_fuse_example <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6421119Z test_quantization.py::TestFuseFx::test_qconfig_fused_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6421707Z test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6422306Z test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_dynamic_quant <- test/quantization/fx/test_quantize_fx.py PASSED [ 45%] 2023-03-31T05:48:33.6422924Z test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_per_channel_quant <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6423603Z test_quantization.py::TestQuantizeFx::test_assert_on_size_after_quant_layer <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6424219Z test_quantization.py::TestQuantizeFx::test_attention <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 46%] 2023-03-31T05:48:33.6424836Z test_quantization.py::TestQuantizeFx::test_backend_config_check_for_weight_and_bias <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6425407Z test_quantization.py::TestQuantizeFx::test_backend_config_quantization_range <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6425953Z test_quantization.py::TestQuantizeFx::test_backend_config_scale_min <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6426506Z test_quantization.py::TestQuantizeFx::test_change_backend_config_for_fixed_qparam_ops <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6427062Z test_quantization.py::TestQuantizeFx::test_channel_shuffle_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6427678Z test_quantization.py::TestQuantizeFx::test_conv_bn_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6428247Z test_quantization.py::TestQuantizeFx::test_conv_linear_not_reference <- test/quantization/fx/test_quantize_fx.py PASSED [ 46%] 2023-03-31T05:48:33.6428767Z test_quantization.py::TestQuantizeFx::test_conv_linear_reference <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6429287Z test_quantization.py::TestQuantizeFx::test_conv_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6429819Z test_quantization.py::TestQuantizeFx::test_conv_transpose_not_reference <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6430358Z test_quantization.py::TestQuantizeFx::test_conv_transpose_reference <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6430891Z test_quantization.py::TestQuantizeFx::test_convert_custom_config_from_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6431490Z test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_observed_to_quantized_mapping <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6432101Z test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_preserved_attributes <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6432670Z test_quantization.py::TestQuantizeFx::test_convert_custom_config_to_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6433195Z test_quantization.py::TestQuantizeFx::test_convert_qconfig_mapping <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6433751Z test_quantization.py::TestQuantizeFx::test_convtranspose_per_channel_fails_early <- test/quantization/fx/test_quantize_fx.py PASSED [ 47%] 2023-03-31T05:48:33.6434322Z test_quantization.py::TestQuantizeFx::test_copy_node_has_shared_actpp_instance <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6434865Z test_quantization.py::TestQuantizeFx::test_custom_module_class <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6435497Z test_quantization.py::TestQuantizeFx::test_custom_module_class_input_has_duplicate_nodes <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6436143Z test_quantization.py::TestQuantizeFx::test_custom_module_class_input_has_multiple_users <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6436709Z test_quantization.py::TestQuantizeFx::test_deepcopy_preserve_attributes <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6437283Z test_quantization.py::TestQuantizeFx::test_default_qconfig_mapping_override_global <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6437835Z test_quantization.py::TestQuantizeFx::test_default_quant_after_none_qconfig <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6438357Z test_quantization.py::TestQuantizeFx::test_dequantize <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6438859Z test_quantization.py::TestQuantizeFx::test_dict_output <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6439400Z test_quantization.py::TestQuantizeFx::test_dynamic_linear_input_multiple_use <- test/quantization/fx/test_quantize_fx.py PASSED [ 48%] 2023-03-31T05:48:33.6439931Z test_quantization.py::TestQuantizeFx::test_dynamic_quant_fp16 <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6440469Z test_quantization.py::TestQuantizeFx::test_dynamic_quant_weight_observer <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6440997Z test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6441537Z test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion_multiple_uses <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6442057Z test_quantization.py::TestQuantizeFx::test_fold_quant_dequant <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6442585Z test_quantization.py::TestQuantizeFx::test_fp32_input_fp32_output <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6443330Z test_quantization.py::TestQuantizeFx::test_fp32_input_quantized_output <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6443852Z test_quantization.py::TestQuantizeFx::test_fp32_sum <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6444364Z test_quantization.py::TestQuantizeFx::test_fuse_custom_config_from_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6444939Z test_quantization.py::TestQuantizeFx::test_fuse_custom_config_set_preserved_attributes <- test/quantization/fx/test_quantize_fx.py PASSED [ 49%] 2023-03-31T05:48:33.6445498Z test_quantization.py::TestQuantizeFx::test_fuse_custom_config_to_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6446022Z test_quantization.py::TestQuantizeFx::test_fused_module_qat_swap <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6446553Z test_quantization.py::TestQuantizeFx::test_fusion_pattern_unquantized <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6447115Z test_quantization.py::TestQuantizeFx::test_get_default_qconfig_valid_backend <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6447671Z test_quantization.py::TestQuantizeFx::test_get_executorch_backend_config <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6448217Z test_quantization.py::TestQuantizeFx::test_getattr_with_nontensor_result <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6448718Z test_quantization.py::TestQuantizeFx::test_linear_bn <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6449241Z test_quantization.py::TestQuantizeFx::test_linear_leaky_relu_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6449880Z test_quantization.py::TestQuantizeFx::test_linear_qint8_activation <- test/quantization/fx/test_quantize_fx.py PASSED [ 50%] 2023-03-31T05:48:33.6450580Z test_quantization.py::TestQuantizeFx::test_linear_shape_view <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 50%] 2023-03-31T05:48:33.6451249Z test_quantization.py::TestQuantizeFx::test_linear_size_view <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 50%] 2023-03-31T05:48:33.6451826Z test_quantization.py::TestQuantizeFx::test_linear_tanh_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6452374Z test_quantization.py::TestQuantizeFx::test_lowering_functional_conv_with_kwargs <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6452948Z test_quantization.py::TestQuantizeFx::test_lowering_functional_linear_with_kwargs <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6453514Z test_quantization.py::TestQuantizeFx::test_masked_fill_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6454077Z test_quantization.py::TestQuantizeFx::test_match_pattern_with_multiple_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6454603Z test_quantization.py::TestQuantizeFx::test_mixed_dtypes <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6455116Z test_quantization.py::TestQuantizeFx::test_mul_add_fp16_config <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6455663Z test_quantization.py::TestQuantizeFx::test_no_obs_between_unmatched_node_and_copy_node <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6456213Z test_quantization.py::TestQuantizeFx::test_non_traceable_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6456719Z test_quantization.py::TestQuantizeFx::test_not_used <- test/quantization/fx/test_quantize_fx.py PASSED [ 51%] 2023-03-31T05:48:33.6457220Z test_quantization.py::TestQuantizeFx::test_observer_fqn <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6457727Z test_quantization.py::TestQuantizeFx::test_output_lists_and_dicts <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6458263Z test_quantization.py::TestQuantizeFx::test_packed_weight_fused_op <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6458777Z test_quantization.py::TestQuantizeFx::test_pattern_match <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6459294Z test_quantization.py::TestQuantizeFx::test_pattern_match_constant <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6459836Z test_quantization.py::TestQuantizeFx::test_permute_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6460401Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_from_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6460997Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_float_to_observed_mapping <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6461620Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_input_quantized_indexes <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6462221Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_classes <- test/quantization/fx/test_quantize_fx.py PASSED [ 52%] 2023-03-31T05:48:33.6462830Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_names <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6463522Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_output_quantized_indexes <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6464209Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_preserved_attributes <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6464871Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_standalone_module_class <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6465474Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_standalone_module_name <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6466043Z test_quantization.py::TestQuantizeFx::test_prepare_custom_config_to_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6466564Z test_quantization.py::TestQuantizeFx::test_prepare_mode <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6467089Z test_quantization.py::TestQuantizeFx::test_prepared_model_deepcopy <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6467604Z test_quantization.py::TestQuantizeFx::test_preserve_attributes <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6468124Z test_quantization.py::TestQuantizeFx::test_preserve_qconfig <- test/quantization/fx/test_quantize_fx.py PASSED [ 53%] 2023-03-31T05:48:33.6468635Z test_quantization.py::TestQuantizeFx::test_preserve_tuple <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6469180Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6469786Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_split_tuple_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6470393Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_tuple_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6470984Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_list_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6471580Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_split_list_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6472167Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_split_tuple_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6472758Z test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_tuple_args <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6473295Z test_quantization.py::TestQuantizeFx::test_qat_and_script <- test/quantization/fx/test_quantize_fx.py PASSED [ 54%] 2023-03-31T05:48:33.6473982Z test_quantization.py::TestQuantizeFx::test_qat_prepare_device_affinity <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py SKIPPED (multi-GPU not supported) [ 54%] 2023-03-31T05:48:33.6474590Z test_quantization.py::TestQuantizeFx::test_qat_skip_untraced <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6475117Z test_quantization.py::TestQuantizeFx::test_qconfig_dict_setup <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6475761Z test_quantization.py::TestQuantizeFx::test_qconfig_dict_with_fused_modules <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 55%] 2023-03-31T05:48:33.6476363Z test_quantization.py::TestQuantizeFx::test_qconfig_for_call_func <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6476882Z test_quantization.py::TestQuantizeFx::test_qconfig_for_call_method <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6477404Z test_quantization.py::TestQuantizeFx::test_qconfig_function <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6477994Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_from_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6478572Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_repr <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6479100Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_global <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6479655Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name <- test/quantization/fx/test_quantize_fx.py PASSED [ 55%] 2023-03-31T05:48:33.6480236Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_object_type_order <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6480817Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_regex <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6481361Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_object_type <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6481906Z test_quantization.py::TestQuantizeFx::test_qconfig_mapping_to_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6482467Z test_quantization.py::TestQuantizeFx::test_qconfig_module_name_object_type_order <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6483012Z test_quantization.py::TestQuantizeFx::test_qconfig_module_name_regex <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6483819Z test_quantization.py::TestQuantizeFx::test_qconfig_module_type <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6484326Z test_quantization.py::TestQuantizeFx::test_qconfig_none <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6484847Z test_quantization.py::TestQuantizeFx::test_qconfig_precedence <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6485378Z test_quantization.py::TestQuantizeFx::test_qconfig_qat_module_type <- test/quantization/fx/test_quantize_fx.py PASSED [ 56%] 2023-03-31T05:48:33.6485903Z test_quantization.py::TestQuantizeFx::test_qnnpack_backend_config <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6486422Z test_quantization.py::TestQuantizeFx::test_qparams_buffers <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6486925Z test_quantization.py::TestQuantizeFx::test_qparams_fqn <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6487459Z test_quantization.py::TestQuantizeFx::test_quant_output_always_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6487991Z test_quantization.py::TestQuantizeFx::test_quantized_input_fp32_output <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6488542Z test_quantization.py::TestQuantizeFx::test_quantized_input_quantized_output <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6489084Z test_quantization.py::TestQuantizeFx::test_quantized_model_type <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6489606Z test_quantization.py::TestQuantizeFx::test_ref_conv_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6490106Z test_quantization.py::TestQuantizeFx::test_ref_linear_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6490615Z test_quantization.py::TestQuantizeFx::test_register_patterns <- test/quantization/fx/test_quantize_fx.py PASSED [ 57%] 2023-03-31T05:48:33.6491126Z test_quantization.py::TestQuantizeFx::test_relu_lowering <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6491634Z test_quantization.py::TestQuantizeFx::test_remove_qconfig <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6492167Z test_quantization.py::TestQuantizeFx::test_repeat_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6492847Z test_quantization.py::TestQuantizeFx::test_reroute_tuple_getitem_patterns <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6493473Z test_quantization.py::TestQuantizeFx::test_reshape_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6493998Z test_quantization.py::TestQuantizeFx::test_return_none <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6494496Z test_quantization.py::TestQuantizeFx::test_reuse_input_qconfig <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6495027Z test_quantization.py::TestQuantizeFx::test_save_observer_state_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6495537Z test_quantization.py::TestQuantizeFx::test_sequential <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6496072Z test_quantization.py::TestQuantizeFx::test_shape_followed_by_quantized_op <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6496622Z test_quantization.py::TestQuantizeFx::test_size_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 58%] 2023-03-31T05:48:33.6497174Z test_quantization.py::TestQuantizeFx::test_stack_trace_preserved_linear <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6497732Z test_quantization.py::TestQuantizeFx::test_standalone_module_float_interface <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6498299Z test_quantization.py::TestQuantizeFx::test_standalone_module_quantized_interface <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6498807Z test_quantization.py::TestQuantizeFx::test_state_dict <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6499303Z test_quantization.py::TestQuantizeFx::test_static_lstm <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6499836Z test_quantization.py::TestQuantizeFx::test_static_lstm_consume_tuple <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6500400Z test_quantization.py::TestQuantizeFx::test_static_lstm_with_custom_fixed_qparams <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6500907Z test_quantization.py::TestQuantizeFx::test_sub_scalar <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6501447Z test_quantization.py::TestQuantizeFx::test_symmetric_qnnpack_qconfig_mapping <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6502026Z test_quantization.py::TestQuantizeFx::test_torch_transpose_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 59%] 2023-03-31T05:48:33.6502619Z test_quantization.py::TestQuantizeFx::test_torch_unsqueeze_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6503167Z test_quantization.py::TestQuantizeFx::test_trace_quantize_per_tensor <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6503808Z test_quantization.py::TestQuantizeFx::test_transpose_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6504383Z test_quantization.py::TestQuantizeFx::test_unsqueeze__nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6504954Z test_quantization.py::TestQuantizeFx::test_unsqueeze_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6505496Z test_quantization.py::TestQuantizeFx::test_view_nontensor_args_not_observed <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6506016Z test_quantization.py::TestQuantizeFxOps::test_add <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6506516Z test_quantization.py::TestQuantizeFxOps::test_add_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6507168Z test_quantization.py::TestQuantizeFxOps::test_add_relu_multiple_uses_of_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6507765Z test_quantization.py::TestQuantizeFxOps::test_ave_pool_with_custom_cfg <- test/quantization/fx/test_quantize_fx.py PASSED [ 60%] 2023-03-31T05:48:33.6508393Z test_quantization.py::TestQuantizeFxOps::test_bmm <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 61%] 2023-03-31T05:48:33.6508974Z test_quantization.py::TestQuantizeFxOps::test_bmm_int_reference <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6509499Z test_quantization.py::TestQuantizeFxOps::test_boolean_tensor <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6509993Z test_quantization.py::TestQuantizeFxOps::test_cat <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6510493Z test_quantization.py::TestQuantizeFxOps::test_chunk <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6510997Z test_quantization.py::TestQuantizeFxOps::test_clamp <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6511502Z test_quantization.py::TestQuantizeFxOps::test_conv_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6512009Z test_quantization.py::TestQuantizeFxOps::test_conv_transpose_1d <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6512539Z test_quantization.py::TestQuantizeFxOps::test_conv_transpose_2d <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6513069Z test_quantization.py::TestQuantizeFxOps::test_copy_node_fp32_input <- test/quantization/fx/test_quantize_fx.py PASSED [ 61%] 2023-03-31T05:48:33.6513675Z test_quantization.py::TestQuantizeFxOps::test_div <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 62%] 2023-03-31T05:48:33.6514238Z test_quantization.py::TestQuantizeFxOps::test_elu <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6514746Z test_quantization.py::TestQuantizeFxOps::test_embedding <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6515262Z test_quantization.py::TestQuantizeFxOps::test_embedding_bag <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6515774Z test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6516312Z test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_fp16 <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6516855Z test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_qint8 <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6517414Z test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_wrong_qconfig <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6517950Z test_quantization.py::TestQuantizeFxOps::test_float_functional <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6518479Z test_quantization.py::TestQuantizeFxOps::test_functional_conv <- test/quantization/fx/test_quantize_fx.py PASSED [ 62%] 2023-03-31T05:48:33.6519008Z test_quantization.py::TestQuantizeFxOps::test_functional_linear <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6519598Z test_quantization.py::TestQuantizeFxOps::test_gelu_normal <- test/quantization/fx/test_quantize_fx.py SKIPPED (TODO: reenable with backend_config api) [ 63%] 2023-03-31T05:48:33.6520256Z test_quantization.py::TestQuantizeFxOps::test_gelu_reference <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 63%] 2023-03-31T05:48:33.6520848Z test_quantization.py::TestQuantizeFxOps::test_general_shape_ops <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6521452Z test_quantization.py::TestQuantizeFxOps::test_general_value_ops <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6522019Z test_quantization.py::TestQuantizeFxOps::test_getitem <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6522513Z test_quantization.py::TestQuantizeFxOps::test_hardswish <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6523146Z test_quantization.py::TestQuantizeFxOps::test_instance_norm <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6523818Z test_quantization.py::TestQuantizeFxOps::test_int8_input_no_unnecessary_fq <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6524344Z test_quantization.py::TestQuantizeFxOps::test_layer_norm <- test/quantization/fx/test_quantize_fx.py PASSED [ 63%] 2023-03-31T05:48:33.6524840Z test_quantization.py::TestQuantizeFxOps::test_leaky_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6525367Z test_quantization.py::TestQuantizeFxOps::test_linear_dynamic_fp16 <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6525891Z test_quantization.py::TestQuantizeFxOps::test_linear_module <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6526414Z test_quantization.py::TestQuantizeFxOps::test_linear_static_fp16 <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6527039Z test_quantization.py::TestQuantizeFxOps::test_mish_reference <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 64%] 2023-03-31T05:48:33.6527602Z test_quantization.py::TestQuantizeFxOps::test_mul <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6528106Z test_quantization.py::TestQuantizeFxOps::test_mul_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6528660Z test_quantization.py::TestQuantizeFxOps::test_multiple_qconfigs_for_single_value <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6529178Z test_quantization.py::TestQuantizeFxOps::test_narrow <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6529696Z test_quantization.py::TestQuantizeFxOps::test_norm_weight_bias <- test/quantization/fx/test_quantize_fx.py PASSED [ 64%] 2023-03-31T05:48:33.6530214Z test_quantization.py::TestQuantizeFxOps::test_pixel_shuffle <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6530732Z test_quantization.py::TestQuantizeFxOps::test_pixel_unshuffle <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6531228Z test_quantization.py::TestQuantizeFxOps::test_prelu <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6531735Z test_quantization.py::TestQuantizeFxOps::test_qbatch_norm <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6532259Z test_quantization.py::TestQuantizeFxOps::test_qbatch_norm_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6532771Z test_quantization.py::TestQuantizeFxOps::test_qmatmul <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6533276Z test_quantization.py::TestQuantizeFxOps::test_quantized_add_qat <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6533805Z test_quantization.py::TestQuantizeFxOps::test_quantized_conv_relu <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6534336Z test_quantization.py::TestQuantizeFxOps::test_quantized_mul_qat <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6534868Z test_quantization.py::TestQuantizeFxOps::test_ref_pattern_multi_use <- test/quantization/fx/test_quantize_fx.py PASSED [ 65%] 2023-03-31T05:48:33.6535374Z test_quantization.py::TestQuantizeFxOps::test_reshape_fp16 <- test/quantization/fx/test_quantize_fx.py PASSED [ 66%] 2023-03-31T05:48:33.6536149Z test_quantization.py::TestQuantizeFxOps::test_rnn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 66%] 2023-03-31T05:48:33.6536884Z test_quantization.py::TestQuantizeFxOps::test_rnn_cell <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 66%] 2023-03-31T05:48:33.6537575Z test_quantization.py::TestQuantizeFxOps::test_silu_reference <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 66%] 2023-03-31T05:48:33.6538224Z test_quantization.py::TestQuantizeFxOps::test_softmax_normal <- test/quantization/fx/test_quantize_fx.py SKIPPED (TODO: reenable with backend_config api) [ 66%] 2023-03-31T05:48:33.6538902Z test_quantization.py::TestQuantizeFxOps::test_softmax_reference <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 66%] 2023-03-31T05:48:33.6539586Z test_quantization.py::TestQuantizeFxOps::test_sub <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 66%] 2023-03-31T05:48:33.6540251Z test_quantization.py::TestQuantizeFxOps::test_sum <- test/quantization/fx/test_quantize_fx.py SKIPPED (This is no longer needed right now, can enable later with new api) [ 66%] 2023-03-31T05:48:33.6540831Z test_quantization.py::TestQuantizeFxModels::test_model_dropout <- test/quantization/fx/test_quantize_fx.py PASSED [ 66%] 2023-03-31T05:48:33.6541441Z test_quantization.py::TestQuantizeFxModels::test_prepare_serialize_switch_device_convert <- test/quantization/fx/test_quantize_fx.py SKIPPED (gpu is not available.) [ 66%] 2023-03-31T05:48:33.6542143Z test_quantization.py::TestQuantizeFxModels::test_qat_embedding_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 66%] 2023-03-31T05:48:33.6542872Z test_quantization.py::TestQuantizeFxModels::test_qat_embeddingbag_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 67%] 2023-03-31T05:48:33.6543675Z test_quantization.py::TestQuantizeFxModels::test_qat_functional_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 67%] 2023-03-31T05:48:33.6544438Z test_quantization.py::TestQuantizeFxModels::test_resnet18_ddp <- test/quantization/fx/test_quantize_fx.py SKIPPED (TODO: Test is always failing - https://github.com/pytorch/pytorch/issues/54979) [ 67%] 2023-03-31T05:48:33.6545142Z test_quantization.py::TestQuantizeFxModels::test_resnet_base <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 67%] 2023-03-31T05:48:33.6545784Z test_quantization.py::TestQuantizeFxModels::test_static_gpu_convert_basic <- test/quantization/fx/test_quantize_fx.py SKIPPED (gpu is not available.) [ 67%] 2023-03-31T05:48:33.6546423Z test_quantization.py::TestQuantizeFxModels::test_switch_device_prepare_convert <- test/quantization/fx/test_quantize_fx.py SKIPPED (gpu is not available.) [ 67%] 2023-03-31T05:48:33.6547053Z test_quantization.py::TestQuantizeFxModels::test_torchvision <- test/quantization/fx/test_quantize_fx.py SKIPPED (skip for now since tbb failed) [ 67%] 2023-03-31T05:48:33.6547652Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_annotations_int <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 67%] 2023-03-31T05:48:33.6548285Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_correct_output_replacement <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 67%] 2023-03-31T05:48:33.6548915Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_graph_argument_order <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 67%] 2023-03-31T05:48:33.6549604Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_internal_pattern_nodes_cannot_have_users_that_are_not_matched <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6550439Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_multiple_pattern_match <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6551060Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_is_entire_graph <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6551751Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_output_pattern_node_can_have_users_that_are_not_matched <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6552412Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_placeholder_matching <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6553031Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_preserves_logic <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6553667Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_replaces_referenced_submodules <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6554308Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_single_pattern_match <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6554922Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_traced_as_callable <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6555537Z test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_with_oneliner_pattern <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 68%] 2023-03-31T05:48:33.6556160Z test_quantization.py::TestSubgraphRewriter::test_subgraph_writer_replace_consecutive_submodules <- test/quantization/fx/test_subgraph_rewriter.py PASSED [ 69%] 2023-03-31T05:48:33.6556749Z test_quantization.py::TestQuantizePT2E::test_qconfig_module_type <- test/quantization/fx/test_quantize_pt2e.py PASSED [ 69%] 2023-03-31T05:48:33.6557280Z test_quantization.py::TestQuantizePT2E::test_qconfig_none <- test/quantization/fx/test_quantize_pt2e.py PASSED [ 69%] 2023-03-31T05:48:33.6557865Z test_quantization.py::TestQuantizePT2E::test_rearrange_weight_observer_for_decomposed_linear <- test/quantization/fx/test_quantize_pt2e.py PASSED [ 69%] 2023-03-31T05:48:33.6558425Z test_quantization.py::TestQuantizePT2EModels::test_resnet18 <- test/quantization/fx/test_quantize_pt2e.py PASSED [ 69%] 2023-03-31T05:48:33.6559008Z test_quantization.py::TestQuantizePT2EX86Inductor::test_inductor_backend_config_conv <- test/quantization/fx/test_quantize_pt2e.py PASSED [ 69%] 2023-03-31T05:48:33.6559584Z test_quantization.py::TestFXGraphMatcher::test_dict_return_type <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 69%] 2023-03-31T05:48:33.6560154Z test_quantization.py::TestFXGraphMatcher::test_matching_failure_node_count <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 69%] 2023-03-31T05:48:33.6560725Z test_quantization.py::TestFXGraphMatcher::test_matching_failure_node_type <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 69%] 2023-03-31T05:48:33.6561276Z test_quantization.py::TestFXGraphMatcher::test_methods <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 69%] 2023-03-31T05:48:33.6561814Z test_quantization.py::TestFXGraphMatcher::test_nodes_before_cat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6562393Z test_quantization.py::TestFXGraphMatcher::test_nodes_with_equal_types_get_matched <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6562964Z test_quantization.py::TestFXGraphMatcher::test_op_relationship_mapping <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6563729Z test_quantization.py::TestFXGraphMatcher::test_results_order <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6564367Z test_quantization.py::TestFXGraphMatcher::test_simple_fun <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6564984Z test_quantization.py::TestFXGraphMatcher::test_simple_fusion <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6565511Z test_quantization.py::TestFXGraphMatcher::test_simple_mod <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6566053Z test_quantization.py::TestFXGraphMatcher::test_simple_mod_multi <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6566603Z test_quantization.py::TestFXGraphMatcher::test_simple_tensor_ops <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6567162Z test_quantization.py::TestFXGraphMatcher::test_user_defined_function <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 70%] 2023-03-31T05:48:33.6567713Z test_quantization.py::TestFXGraphMatcherModels::test_mobilenet_v2 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6568287Z test_quantization.py::TestFXGraphMatcherModels::test_mobilenet_v2_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6568905Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_loggers_cuda <- test/quantization/fx/test_numeric_suite_fx.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T05:48:33.6569533Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_mul_inputs_activations <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6570173Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_cuda <- test/quantization/fx/test_numeric_suite_fx.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T05:48:33.6570793Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_fun_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6571405Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_fun_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6572023Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_meth_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6572630Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_mod_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6573225Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_mod_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 71%] 2023-03-31T05:48:33.6573844Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extend_logger_results_with_comparison <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6574470Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_conv_fun_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6575083Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_conv_fun_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6575714Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_cuda <- test/quantization/fx/test_numeric_suite_fx.py SKIPPED (CUDA unavailable) [ 72%] 2023-03-31T05:48:33.6576334Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_dynamic <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6576933Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_fqn <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6577538Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_linear_fun_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6578272Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_linear_fun_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6579046Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_mod_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6579713Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_mod_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 72%] 2023-03-31T05:48:33.6580308Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_fp16_shadows_fp32 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6580898Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_fp32_coverage <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6581506Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_fp32_simple <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6582104Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_int8_fun <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6582702Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_int8_mod <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6583339Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_layer_names <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6583939Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_activations <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6584556Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_shadow_activations <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6585202Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_vs_linear_fp16_shadow_activations <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6585820Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_weights <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 73%] 2023-03-31T05:48:33.6586407Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_kwargs_shadow <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6587023Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_loggers_preserve_qat_numerics <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6587618Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_logging_inputs <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6588208Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fqn <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6588979Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fun_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6589598Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fun_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6590213Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_meth_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6590826Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_mod_ptq <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6591421Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_mod_qat <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6592038Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_mul_add_cat_stack_skips_shadowing <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 74%] 2023-03-31T05:48:33.6592651Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_io_dtype_coverage <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6593263Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_either_fp32_or_int8_input <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6593958Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_only_kwargs_skips_shadowing <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6594657Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_ops_with_same_fp32_and_int8_signature <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6595280Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_shadow_activations_fqn <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6595904Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_shadow_loggers_preserve_qat_numerics <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6596534Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_unsupported_op_copy_skips_shadowing <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6597147Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_defined_function <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6597733Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_module <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6598319Z test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_module_scriptable <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6598938Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_conv_bn_relu_fusion_fp32 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 75%] 2023-03-31T05:48:33.6599553Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_conv_bn_relu_fusion_quant <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6600165Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_functions <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6600782Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_fp32_fp32 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6601416Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_fp32_quant <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6602025Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_fp32 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6602645Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_quant <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6603468Z test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_mobilenet_v2 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6604068Z test_quantization.py::TestFXNumericSuiteNShadows::test_conv_bn_relu_mod <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6604660Z test_quantization.py::TestFXNumericSuiteNShadows::test_custom_functions_and_tracer <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6605269Z test_quantization.py::TestFXNumericSuiteNShadows::test_extract_weights_linear <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 76%] 2023-03-31T05:48:33.6605844Z test_quantization.py::TestFXNumericSuiteNShadows::test_functions <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6606411Z test_quantization.py::TestFXNumericSuiteNShadows::test_linear_mod <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6607223Z test_quantization.py::TestFXNumericSuiteNShadows::test_linear_relu_mod <- test/quantization/fx/test_numeric_suite_fx.py [W qlinear_dynamic.cpp:247] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function operator()) 2023-03-31T05:48:33.6607718Z PASSED [ 77%] 2023-03-31T05:48:33.6608315Z test_quantization.py::TestFXNumericSuiteNShadows::test_logger_enabled_and_save_activations_flags <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6608989Z test_quantization.py::TestFXNumericSuiteNShadows::test_mobilenet_v2 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6609573Z test_quantization.py::TestFXNumericSuiteNShadows::test_partial_qconfig_mapping <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6610195Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_deduplication <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6610829Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_end_to_end <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6611451Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_from_list <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6612083Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_insert_padding <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 77%] 2023-03-31T05:48:33.6612703Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_ordering <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6613314Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_repr <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6613952Z test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_retroactive_padding <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6614603Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_conv <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6615234Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_linear <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6615895Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_lstm_dynamic <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6616557Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_conv <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6617218Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_linear <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6617878Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_lstm_dynamic <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6618529Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_conv <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 78%] 2023-03-31T05:48:33.6619167Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_linear <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6619811Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_lstm_dynamic <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6620418Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_mobilenet_v2 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6621014Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_resnet18 <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6621641Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_sparsenn_compare_activations <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6622271Z test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_sparsenn_shadow <- test/quantization/fx/test_numeric_suite_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6622977Z test_quantization.py::TestFxModelReportDetector::test_conv_sub_class_considered <- test/quantization/fx/test_model_report_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6623737Z test_quantization.py::TestFxModelReportDetector::test_fusion_layer_in_sequential <- test/quantization/fx/test_model_report_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6624357Z test_quantization.py::TestFxModelReportDetector::test_multi_linear_model_without_per_channel <- test/quantization/fx/test_model_report_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6624970Z test_quantization.py::TestFxModelReportDetector::test_multiple_q_config_options <- test/quantization/fx/test_model_report_fx.py PASSED [ 79%] 2023-03-31T05:48:33.6625572Z test_quantization.py::TestFxModelReportDetector::test_qat_aware_model_example <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6626151Z test_quantization.py::TestFxModelReportDetector::test_sequential_model_format <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6626730Z test_quantization.py::TestFxModelReportDetector::test_simple_conv <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6627312Z test_quantization.py::TestFxModelReportObserver::test_observer_after_relu <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6627903Z test_quantization.py::TestFxModelReportObserver::test_random_epochs_and_batches <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6628479Z test_quantization.py::TestFxModelReportObserver::test_single_batch_of_ones <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6629057Z test_quantization.py::TestFxModelReportObserver::test_zero_tensor_errors <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6629679Z test_quantization.py::TestFxModelReportDetectDynamicStatic::test_nested_detection_case <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6630270Z test_quantization.py::TestFxModelReportClass::test_constructor <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6630848Z test_quantization.py::TestFxModelReportClass::test_equalization_mapping_generation <- test/quantization/fx/test_model_report_fx.py PASSED [ 80%] 2023-03-31T05:48:33.6631423Z test_quantization.py::TestFxModelReportClass::test_generate_report <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6631984Z test_quantization.py::TestFxModelReportClass::test_generate_visualizer <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6632565Z test_quantization.py::TestFxModelReportClass::test_prepare_model_callibration <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6633145Z test_quantization.py::TestFxModelReportClass::test_qconfig_mapping_generation <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6633803Z test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_determine_points <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6634507Z test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6635200Z test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen_empty <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6635805Z test_quantization.py::TestFxDetectOutliers::test_all_outlier_report_gen <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6636410Z test_quantization.py::TestFxDetectOutliers::test_multiple_run_consistent_spike_outlier_report_gen <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6637005Z test_quantization.py::TestFxDetectOutliers::test_no_outlier_report_gen <- test/quantization/fx/test_model_report_fx.py PASSED [ 81%] 2023-03-31T05:48:33.6637661Z test_quantization.py::TestFxDetectOutliers::test_outlier_detection_determine_points <- test/quantization/fx/test_model_report_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6638325Z test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_match_with_report <- test/quantization/fx/test_model_report_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6651756Z test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_no_match <- test/quantization/fx/test_model_report_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6652619Z test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_single_feat_match <- test/quantization/fx/test_model_report_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6653256Z test_quantization.py::TestFxModelReportVisualizer::test_get_modules_and_features <- test/quantization/fx/test_model_report_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6653829Z test_quantization.py::TestEqualizeFx::test_input_weight_eq_observer <- test/quantization/fx/test_equalize_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6654416Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_activation_values <- test/quantization/fx/test_equalize_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6654986Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_branching <- test/quantization/fx/test_equalize_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6655556Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_convert <- test/quantization/fx/test_equalize_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6656142Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_equalization_scales <- test/quantization/fx/test_equalize_fx.py PASSED [ 82%] 2023-03-31T05:48:33.6656715Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_graphs <- test/quantization/fx/test_equalize_fx.py PASSED [ 83%] 2023-03-31T05:48:33.6657257Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_prepare <- test/quantization/fx/test_equalize_fx.py PASSED [ 83%] 2023-03-31T05:48:33.6657823Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_results <- test/quantization/fx/test_equalize_fx.py PASSED [ 83%] 2023-03-31T05:48:33.6658400Z test_quantization.py::TestEqualizeFx::test_input_weight_equalization_weights_bias <- test/quantization/fx/test_equalize_fx.py PASSED [ 83%] 2023-03-31T05:48:33.6658950Z test_quantization.py::TestEqualizeFx::test_selective_equalization <- test/quantization/fx/test_equalize_fx.py PASSED [ 83%] 2023-03-31T05:48:33.6659554Z test_quantization.py::TestSerialization::test_conv2d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6660228Z test_quantization.py::TestSerialization::test_conv2d_graph <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6660902Z test_quantization.py::TestSerialization::test_conv2d_graph_v2 <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6661582Z test_quantization.py::TestSerialization::test_conv2d_graph_v3 <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6662254Z test_quantization.py::TestSerialization::test_conv2d_nobias <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6662921Z test_quantization.py::TestSerialization::test_conv2d_nobias_graph <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 83%] 2023-03-31T05:48:33.6663743Z test_quantization.py::TestSerialization::test_conv2d_nobias_graph_v2 <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6664436Z test_quantization.py::TestSerialization::test_conv2d_nobias_graph_v3 <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6665451Z test_quantization.py::TestSerialization::test_conv2d_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6666116Z test_quantization.py::TestSerialization::test_conv3d <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6666765Z test_quantization.py::TestSerialization::test_conv3d_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6667374Z test_quantization.py::TestSerialization::test_default_qat_qconfig <- test/quantization/bc/test_backward_compatibility.py PASSED [ 84%] 2023-03-31T05:48:33.6668005Z test_quantization.py::TestSerialization::test_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6668686Z test_quantization.py::TestSerialization::test_linear_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6669342Z test_quantization.py::TestSerialization::test_linear_relu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 84%] 2023-03-31T05:48:33.6669994Z test_quantization.py::TestSerialization::test_linear_relu_package_quantization_transforms <- test/quantization/bc/test_backward_compatibility.py PASSED [ 84%] 2023-03-31T05:48:33.6670650Z test_quantization.py::TestSerialization::test_lstm <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 85%] 2023-03-31T05:48:33.6671247Z test_quantization.py::TestSerialization::test_per_channel_observer <- test/quantization/bc/test_backward_compatibility.py PASSED [ 85%] 2023-03-31T05:48:33.6671812Z test_quantization.py::TestSerialization::test_per_tensor_observer <- test/quantization/bc/test_backward_compatibility.py PASSED [ 85%] 2023-03-31T05:48:33.6672442Z test_quantization.py::TestQuantizeJit::test_conv <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 85%] 2023-03-31T05:48:33.6673087Z test_quantization.py::TestQuantizeJit::test_conv_bn <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 85%] 2023-03-31T05:48:33.6673999Z test_quantization.py::TestQuantizeJit::test_conv_transpose <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W insert_observers.cpp:1586] Warning: prim::Loop is not yet supported in quantization, please make sure nothing needs to be quantized in the loop (function operator()) 2023-03-31T05:48:33.6674521Z PASSED [ 85%] 2023-03-31T05:48:33.6675061Z test_quantization.py::TestQuantizeJit::test_linear_dynamic_fp16 <- test/quantization/jit/test_quantize_jit.py [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6675554Z [W QuantUtils.h:215] Warning: FOUND weight out of range (function HandleWeightsSaturation) 2023-03-31T05:48:33.6675825Z PASSED [ 85%] 2023-03-31T05:48:33.6676314Z test_quantization.py::TestQuantizeJit::test_nested <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 85%] 2023-03-31T05:48:33.6676898Z test_quantization.py::TestQuantizeJit::test_observer_with_ignored_function <- test/quantization/jit/test_quantize_jit.py PASSED [ 85%] 2023-03-31T05:48:33.6677530Z test_quantization.py::TestQuantizeJit::test_single_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 85%] 2023-03-31T05:48:33.6678282Z test_quantization.py::TestQuantizeJit::test_single_linear_dynamic <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 86%] 2023-03-31T05:48:33.6679036Z test_quantization.py::TestQuantizeJit::test_skip_quant <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 86%] 2023-03-31T05:48:33.6679806Z test_quantization.py::TestQuantizeJitPasses::test_conv_trace <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6680364Z test_quantization.py::TestQuantizeJitPasses::test_convtranspose_trace <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6680923Z test_quantization.py::TestQuantizeJitPasses::test_dedup_module_uses <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6681478Z test_quantization.py::TestQuantizeJitPasses::test_finalize_debug <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6682020Z test_quantization.py::TestQuantizeJitPasses::test_finalize_for_linear <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6682584Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_complex_cases <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6683319Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_in_submodule <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6683960Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_no_fusion <- test/quantization/jit/test_quantize_jit.py PASSED [ 86%] 2023-03-31T05:48:33.6684516Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_shared_classtype <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6685077Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6685642Z test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial_nobias <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6686195Z test_quantization.py::TestQuantizeJitPasses::test_fuse_linear <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6686730Z test_quantization.py::TestQuantizeJitPasses::test_inplace_option <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6687281Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6687860Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_child_qconfig <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6688454Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_general_ops <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6689023Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_if <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6689640Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_if_consistent_observation <- test/quantization/jit/test_quantize_jit.py PASSED [ 87%] 2023-03-31T05:48:33.6690248Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_nested_if <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6690840Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_reused_weight <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6691416Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_interface <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6692022Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_interface_unshare_type <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6692634Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6693469Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed_for_function <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6694113Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed_in_submodule <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6694719Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_shared_class_type <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6695302Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_skip_values <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6695889Z test_quantization.py::TestQuantizeJitPasses::test_insert_observers_weight_dtype <- test/quantization/jit/test_quantize_jit.py PASSED [ 88%] 2023-03-31T05:48:33.6696457Z test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6697042Z test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant_shared_class_type <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6697619Z test_quantization.py::TestQuantizeJitPasses::test_interface_with_fork <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6698165Z test_quantization.py::TestQuantizeJitPasses::test_module_list <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6698877Z test_quantization.py::TestQuantizeJitPasses::test_quantize_fork_wait <- test/quantization/jit/test_quantize_jit.py [W utils.py:317] Warning: must run observer before calling calculate_qparams. Returning default values. (function ) 2023-03-31T05:48:33.6699279Z PASSED [ 89%] 2023-03-31T05:48:33.6699727Z test_quantization.py::TestQuantizeJitPasses::test_replicate_dequant_same_value <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6700303Z test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6700886Z test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize_in_block <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6701621Z test_quantization.py::TestQuantizeJitPasses::test_replicate_quantize_for_if <- test/quantization/jit/test_quantize_jit.py [W utils.py:317] Warning: must run observer before calling calculate_qparams. Returning default values. (function ) 2023-03-31T05:48:33.6702151Z [W utils.py:317] Warning: must run observer before calling calculate_qparams. Returning default values. (function ) 2023-03-31T05:48:33.6702547Z [W utils.py:317] Warning: must run observer before calling calculate_qparams. Returning default values. (function ) 2023-03-31T05:48:33.6702937Z [W utils.py:317] Warning: must run observer before calling calculate_qparams. Returning default values. (function ) 2023-03-31T05:48:33.6703284Z PASSED [ 89%] 2023-03-31T05:48:33.6703742Z test_quantization.py::TestQuantizeJitPasses::test_skip_dequant_constant_prop <- test/quantization/jit/test_quantize_jit.py PASSED [ 89%] 2023-03-31T05:48:33.6704328Z test_quantization.py::TestQuantizeJitPasses::test_swap_functional_linear <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6704871Z test_quantization.py::TestQuantizeJitOps::test_cat_linear <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6705373Z test_quantization.py::TestQuantizeJitOps::test_clamp <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6706017Z test_quantization.py::TestQuantizeJitOps::test_conv_with_benchmark_flag <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 90%] 2023-03-31T05:48:33.6706619Z test_quantization.py::TestQuantizeJitOps::test_dequantize_tuple <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6707084Z test_quantization.py::TestQuantizeJitOps::test_elu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 90%] 2023-03-31T05:48:33.6707449Z test_quantization.py::TestQuantizeJitOps::test_general_shape_ops <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6707759Z test_quantization.py::TestQuantizeJitOps::test_general_value_ops <- test/quantization/jit/test_quantize_jit.py PASSED [ 90%] 2023-03-31T05:48:33.6708436Z test_quantization.py::TestQuantizeJitOps::test_group_norm <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W insert_quant_dequant.cpp:1360] Warning: debug option for add_scalar and mul_scalar is not supported, please don't use debug option for models that uses these ops. (function operator()) 2023-03-31T05:48:33.6708499Z PASSED [ 90%] 2023-03-31T05:48:33.6708900Z test_quantization.py::TestQuantizeJitOps::test_hardswish <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 90%] 2023-03-31T05:48:33.6709528Z test_quantization.py::TestQuantizeJitOps::test_instance_norm <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6709757Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6709974Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6710188Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6710418Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6710630Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6710840Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6711062Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6711276Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6711491Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6711717Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6711918Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6712128Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6712344Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6712611Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6712883Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6713100Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6713312Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6713520Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6713734Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6713954Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6714162Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6714377Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6714577Z [W observer.py:1209] Warning: must run observer before calling calculate_qparams. Returning default scale and zero point (function ) 2023-03-31T05:48:33.6714653Z PASSED [ 91%] 2023-03-31T05:48:33.6715060Z test_quantization.py::TestQuantizeJitOps::test_layer_norm <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 91%] 2023-03-31T05:48:33.6715365Z test_quantization.py::TestQuantizeJitOps::test_linear <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6715663Z test_quantization.py::TestQuantizeJitOps::test_qbatch_norm <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6716007Z test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNFuncInplaceRelu <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6716336Z test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNFuncRelu <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6716659Z test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNRelu <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6716969Z test_quantization.py::TestQuantizeJitOps::test_quantized_add <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6717286Z test_quantization.py::TestQuantizeJitOps::test_quantized_add_alpha <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6717589Z test_quantization.py::TestQuantizeJitOps::test_quantized_add_relu <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6717910Z test_quantization.py::TestQuantizeJitOps::test_quantized_add_relu_alpha <- test/quantization/jit/test_quantize_jit.py PASSED [ 91%] 2023-03-31T05:48:33.6718223Z test_quantization.py::TestQuantizeJitOps::test_quantized_add_scalar <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6718544Z test_quantization.py::TestQuantizeJitOps::test_quantized_add_scalar_relu <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6718845Z test_quantization.py::TestQuantizeJitOps::test_quantized_cat <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6719228Z test_quantization.py::TestQuantizeJitOps::test_quantized_conv <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6719621Z test_quantization.py::TestQuantizeJitOps::test_quantized_conv_relu <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6719919Z test_quantization.py::TestQuantizeJitOps::test_quantized_mul <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6720229Z test_quantization.py::TestQuantizeJitOps::test_quantized_mul_relu <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6720544Z test_quantization.py::TestQuantizeJitOps::test_quantized_mul_scalar <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6720852Z test_quantization.py::TestQuantizeJitOps::test_quantized_mul_scalar_relu <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6721197Z test_quantization.py::TestQuantizeDynamicJitPasses::test_convert_dynamic_fp16 <- test/quantization/jit/test_quantize_jit.py PASSED [ 92%] 2023-03-31T05:48:33.6721643Z test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_multi_op <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 93%] 2023-03-31T05:48:33.6722093Z test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_quant_multi_uses <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 93%] 2023-03-31T05:48:33.6722539Z test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_shared_weights <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 93%] 2023-03-31T05:48:33.6722890Z test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_weight_observer <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6723607Z test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_with_if <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 93%] 2023-03-31T05:48:33.6723995Z test_quantization.py::TestQuantizeDynamicJitPasses::test_insert_quant_dequant_linear_dynamic <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6724332Z test_quantization.py::TestQuantizeDynamicJitPasses::test_prepare_dynamic <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6724693Z test_quantization.py::TestQuantizeDynamicJitPasses::test_prepare_dynamic_child_qconfig <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6725036Z test_quantization.py::TestQuantizeDynamicJitPasses::test_quantize_dynamic_fp16 <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6725360Z test_quantization.py::TestQuantizeDynamicJitOps::test_embedding_bag <- test/quantization/jit/test_quantize_jit.py PASSED [ 93%] 2023-03-31T05:48:33.6725702Z test_quantization.py::TestQuantizeDynamicJitOps::test_embedding_bag_padding_idx_error <- test/quantization/jit/test_quantize_jit.py PASSED [ 94%] 2023-03-31T05:48:33.6726116Z test_quantization.py::TestQuantizeDynamicJitOps::test_linear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 94%] 2023-03-31T05:48:33.6726439Z test_quantization.py::TestFusionPasses::test_quantized_add_relu_fusion <- test/quantization/jit/test_fusion_passes.py PASSED [ 94%] 2023-03-31T05:48:33.6726805Z test_quantization.py::TestDeprecatedJitQuantized::test_erase_class_tensor_shapes <- test/quantization/jit/test_deprecated_jit_quant.py PASSED [ 94%] 2023-03-31T05:48:33.6727163Z test_quantization.py::TestDeprecatedJitQuantized::test_quantization_modules <- test/quantization/jit/test_deprecated_jit_quant.py PASSED [ 94%] 2023-03-31T05:48:33.6727516Z test_quantization.py::TestDeprecatedJitQuantized::test_rnn_cell_quantized <- test/quantization/jit/test_deprecated_jit_quant.py PASSED [ 94%] 2023-03-31T05:48:33.6727961Z test_quantization.py::TestDeprecatedJitQuantized::test_rnn_quantized <- test/quantization/jit/test_deprecated_jit_quant.py PASSED [ 94%] 2023-03-31T05:48:33.6728402Z test_quantization.py::TestAOMigrationQuantization::test_function_import_fake_quantize <- test/quantization/ao_migration/test_quantization.py PASSED [ 94%] 2023-03-31T05:48:33.6728782Z test_quantization.py::TestAOMigrationQuantization::test_function_import_fuse_modules <- test/quantization/ao_migration/test_quantization.py PASSED [ 94%] 2023-03-31T05:48:33.6729170Z test_quantization.py::TestAOMigrationQuantization::test_function_import_fuser_method_mappings <- test/quantization/ao_migration/test_quantization.py PASSED [ 94%] 2023-03-31T05:48:33.6729536Z test_quantization.py::TestAOMigrationQuantization::test_function_import_observer <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6729888Z test_quantization.py::TestAOMigrationQuantization::test_function_import_qconfig <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6730261Z test_quantization.py::TestAOMigrationQuantization::test_function_import_quant_type <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6730653Z test_quantization.py::TestAOMigrationQuantization::test_function_import_quantization_mappings <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6731018Z test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6731384Z test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize_jit <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6731744Z test_quantization.py::TestAOMigrationQuantization::test_function_import_stubs <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6732099Z test_quantization.py::TestAOMigrationQuantization::test_function_import_utils <- test/quantization/ao_migration/test_quantization.py PASSED [ 95%] 2023-03-31T05:48:33.6732456Z test_quantization.py::TestAOMigrationNNQuantized::test_functional_import <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 95%] 2023-03-31T05:48:33.6732803Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_conv <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 95%] 2023-03-31T05:48:33.6733171Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_dynamic_linear <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6733535Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_embedding_ops <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6733873Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_linear <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6734249Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantizable_activation <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6734619Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantizable_rnn <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6734998Z test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_quantized_dynamic_import <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6735348Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_activation <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6735698Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_batchnorm <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6736038Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_conv <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6736448Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_dropout <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 96%] 2023-03-31T05:48:33.6736854Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_embedding_ops <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6737219Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_functional_modules <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6737562Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_import <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6737888Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_linear <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6738239Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_normalization <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6738583Z test_quantization.py::TestAOMigrationNNQuantized::test_modules_utils <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6738953Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6739323Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_qat <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6739702Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_quantized <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6740072Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_conv_fused <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 97%] 2023-03-31T05:48:33.6740445Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_fused <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6740824Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_relu <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6741206Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_bn_relu <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6741584Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_conv_relu <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6741964Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_linear_relu <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6742311Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_nn_intrinsic_fused <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6742718Z test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_no_import_nn_intrinsic_quantized_dynamic <- test/quantization/ao_migration/test_ao_migration.py PASSED [ 98%] 2023-03-31T05:48:33.6743085Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 98%] 2023-03-31T05:48:33.6743548Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_convert <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 98%] 2023-03-31T05:48:33.6743925Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_equalize <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 98%] 2023-03-31T05:48:33.6744294Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fuse <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6744683Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fusion_patterns <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6745206Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_graph_module <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6745588Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_match_utils <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6745976Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_pattern_utils <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6746352Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_prepare <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6746752Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_quantization_patterns <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6747112Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_utils <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6747493Z test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_quantize_fx <- test/quantization/ao_migration/test_quantization_fx.py PASSED [ 99%] 2023-03-31T05:48:33.6747778Z test_quantization.py::TestBits::test_subclass <- test/quantization/core/experimental/test_bits.py PASSED [ 99%] 2023-03-31T05:48:33.6748406Z test_quantization.py::TestBits::test_types <- test/quantization/core/experimental/test_bits.py PASSED [100%]/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/nn/modules/lazy.py:180: UserWarning: Lazy modules are a new feature under heavy development so changes to the API or functionality can happen at any moment. 2023-03-31T05:48:33.6748624Z warnings.warn('Lazy modules are a new feature under heavy development ' 2023-03-31T05:48:33.6748639Z 2023-03-31T05:48:33.6748645Z 2023-03-31T05:48:33.6749001Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_quantization/test_quantization-816f8ae0c160c36a.xml - 2023-03-31T05:48:33.6749131Z ================= 948 passed, 64 skipped in 973.49s (0:16:13) ================== 2023-03-31T05:48:33.6749312Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:48:33.6749319Z 2023-03-31T05:48:33.6751531Z ##[endgroup] 2023-03-31T05:48:33.6751869Z FINISHED PRINTING LOG FILE of test_quantization (/var/lib/jenkins/workspace/test/test-reports/test_quantization_zpcgt5ju.log) 2023-03-31T05:48:33.6751877Z 2023-03-31T05:48:35.6642557Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:48:35.6691593Z Ignoring disabled issues: [] 2023-03-31T05:48:35.6864740Z Running test_jit ... [2023-03-31 05:48:35.686152] 2023-03-31T05:48:35.6866771Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_jit.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:48:35.686455] 2023-03-31T05:50:35.9182843Z 2023-03-31T05:50:35.9183523Z Expand the folded group to see the log file of test_jit 2023-03-31T05:50:35.9185206Z ##[group]PRINTING LOG FILE of test_jit (/var/lib/jenkins/workspace/test/test-reports/test_jit_144gxr3v.log) 2023-03-31T05:50:35.9190451Z Test results will be stored in test-reports/python-pytest/test_jit/test_jit-47d330549a92143e.xml 2023-03-31T05:50:35.9191025Z ============================= test session starts ============================== 2023-03-31T05:50:35.9191449Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:50:35.9193587Z cachedir: .pytest_cache 2023-03-31T05:50:35.9194059Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:50:35.9194773Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:50:35.9195303Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:50:35.9195630Z collecting ... collected 2560 items 2023-03-31T05:50:35.9639266Z Running 2560 items in this shard: test/test_jit.py::TestTracer::test_call_traced_fn_from_traced_module, test/test_jit.py::TestTracer::test_call_traced_module_from_traced_module, test/test_jit.py::TestTracer::test_canonicalize_tensor_iterator, test/test_jit.py::TestTracer::test_constant, test/test_jit.py::TestTracer::test_conv, test/test_jit.py::TestTracer::test_export_no_reorder, test/test_jit.py::TestTracer::test_force_outplace_check_fill, test/test_jit.py::TestTracer::test_force_outplace_check_zero, test/test_jit.py::TestTracer::test_ge, test/test_jit.py::TestTracer::test_ge_cuda, test/test_jit.py::TestTracer::test_ge_optimized, test/test_jit.py::TestTracer::test_ge_unoptimized, test/test_jit.py::TestTracer::test_index_put, test/test_jit.py::TestTracer::test_index_put_trace_with_view, test/test_jit.py::TestTracer::test_index_put_trace_without_view, test/test_jit.py::TestTracer::test_inplace_check, test/test_jit.py::TestTracer::test_inplace_copy, test/test_jit.py::TestTracer::test_inplace_copy_force_outplace, test/test_jit.py::TestTracer::test_inplace_flags, test/test_jit.py::TestTracer::test_inplace_transplant, test/test_jit.py::TestTracer::test_inplace_warn, test/test_jit.py::TestTracer::test_input_dict_checkTrace_mut, test/test_jit.py::TestTracer::test_input_dict_empty, test/test_jit.py::TestTracer::test_input_dict_empty_list, test/test_jit.py::TestTracer::test_input_dict_insertion_order, test/test_jit.py::TestTracer::test_input_dict_of_dicts, test/test_jit.py::TestTracer::test_input_dict_of_lists, test/test_jit.py::TestTracer::test_input_dict_recursive, test/test_jit.py::TestTracer::test_input_dict_remembers_keys, test/test_jit.py::TestTracer::test_input_dict_unify, test/test_jit.py::TestTracer::test_input_flatten, test/test_jit.py::TestTracer::test_input_list_mixed_type, test/test_jit.py::TestTracer::test_input_list_of_tuples, test/test_jit.py::TestTracer::test_input_list_toplevel_flatten, test/test_jit.py::TestTracer::test_input_list_toplevel_flatten_direct, test/test_jit.py::TestTracer::test_input_tuple_of_dicts, test/test_jit.py::TestTracer::test_interpolate_trace, test/test_jit.py::TestTracer::test_large_nbr_kernel_args, test/test_jit.py::TestTracer::test_lhs_index_fails, test/test_jit.py::TestTracer::test_lhs_index_trivial, test/test_jit.py::TestTracer::test_max_pool, test/test_jit.py::TestTracer::test_nested_inplace, test/test_jit.py::TestTracer::test_non_tensor_tracing, test/test_jit.py::TestTracer::test_output_unflatten, test/test_jit.py::TestTracer::test_python_function, test/test_jit.py::TestTracer::test_python_function_tup, test/test_jit.py::TestTracer::test_repeated_input, test/test_jit.py::TestTracer::test_repeated_output, test/test_jit.py::TestTracer::test_shared_param, test/test_jit.py::TestTracer::test_simple, test/test_jit.py::TestTracer::test_tensor_with_grad_as_constant, test/test_jit.py::TestTracer::test_trace_aliased_parameter, test/test_jit.py::TestTracer::test_trace_annotation, test/test_jit.py::TestTracer::test_trace_arange, test/test_jit.py::TestTracer::test_trace_arange_with_grad, test/test_jit.py::TestTracer::test_trace_autograd_function, test/test_jit.py::TestTracer::test_trace_c10_ops, test/test_jit.py::TestTracer::test_trace_casts, test/test_jit.py::TestTracer::test_trace_checker_control_flow, test/test_jit.py::TestTracer::test_trace_checker_dot_data, test/test_jit.py::TestTracer::test_trace_checker_dropout_notrain, test/test_jit.py::TestTracer::test_trace_checker_dropout_train, test/test_jit.py::TestTracer::test_trace_checker_inplace_on_view, test/test_jit.py::TestTracer::test_trace_checker_memoization, test/test_jit.py::TestTracer::test_trace_checker_slice_lhs, test/test_jit.py::TestTracer::test_trace_checking_with_deprecated_name, test/test_jit.py::TestTracer::test_trace_checking_with_global_name, test/test_jit.py::TestTracer::test_trace_contiguous, test/test_jit.py::TestTracer::test_trace_contiguous_short_circuit, test/test_jit.py::TestTracer::test_trace_detach, test/test_jit.py::TestTracer::test_trace_detach_inplace, test/test_jit.py::TestTracer::test_trace_detach_inplace_redispatch, test/test_jit.py::TestTracer::test_trace_detach_redispatch, test/test_jit.py::TestTracer::test_trace_dict_input, test/test_jit.py::TestTracer::test_trace_dict_output, test/test_jit.py::TestTracer::test_trace_export_fns, test/test_jit.py::TestTracer::test_trace_export_fns_recursive, test/test_jit.py::TestTracer::test_trace_fork_join_and_module, test/test_jit.py::TestTracer::test_trace_full_dynamic_shape, test/test_jit.py::TestTracer::test_trace_func_argument_names_captured, test/test_jit.py::TestTracer::test_trace_index, test/test_jit.py::TestTracer::test_trace_index_constant, test/test_jit.py::TestTracer::test_trace_indexed_assignment, test/test_jit.py::TestTracer::test_trace_inline_shape, test/test_jit.py::TestTracer::test_trace_inverse, test/test_jit.py::TestTracer::test_trace_invert_module_hierarchy, test/test_jit.py::TestTracer::test_trace_legacy_ctor, test/test_jit.py::TestTracer::test_trace_module_argument_names_captured, test/test_jit.py::TestTracer::test_trace_modulelist, test/test_jit.py::TestTracer::test_trace_multi_output_function, test/test_jit.py::TestTracer::test_trace_namedtuple, test/test_jit.py::TestTracer::test_trace_nested_datatypes, test/test_jit.py::TestTracer::test_trace_nested_fn, test/test_jit.py::TestTracer::test_trace_numel, test/test_jit.py::TestTracer::test_trace_optioanl_dtype, test/test_jit.py::TestTracer::test_trace_optional, test/test_jit.py::TestTracer::test_trace_partial_func_argument_names_captured, test/test_jit.py::TestTracer::test_trace_random, test/test_jit.py::TestTracer::test_trace_records_names, test/test_jit.py::TestTracer::test_trace_save, test/test_jit.py::TestTracer::test_trace_save_load_copy, test/test_jit.py::TestTracer::test_trace_single_tuple, test/test_jit.py::TestTracer::test_trace_size, test/test_jit.py::TestTracer::test_trace_size_with_grad, test/test_jit.py::TestTracer::test_trace_skip_none_submodule, test/test_jit.py::TestTracer::test_trace_slice, test/test_jit.py::TestTracer::test_trace_slice_expr_complete_type, test/test_jit.py::TestTracer::test_trace_slice_full_dim, test/test_jit.py::TestTracer::test_trace_slice_setitem_dynamic_shape, test/test_jit.py::TestTracer::test_trace_slice_with_grad, test/test_jit.py::TestTracer::test_trace_tensor_factory, test/test_jit.py::TestTracer::test_trace_topk, test/test_jit.py::TestTracer::test_trace_tuple, test/test_jit.py::TestTracer::test_trace_variable_instantiation, test/test_jit.py::TestTracer::test_trace_warn, test/test_jit.py::TestTracer::test_trace_with_conditional_property, test/test_jit.py::TestTracer::test_trace_with_nested_tensor_list_output, test/test_jit.py::TestTracer::test_trace_with_number_list_output, test/test_jit.py::TestTracer::test_trace_with_tensor_list_output, test/test_jit.py::TestTracer::test_traced_module_cuda, test/test_jit.py::TestTracer::test_tracing_backward_hook_error, test/test_jit.py::TestTracer::test_tracing_hooks, test/test_jit.py::TestTracer::test_tracing_multiple_methods, test/test_jit.py::TestTracer::test_typeas_trace_check, test/test_jit.py::TestTracer::test_wrapped_number, test/test_jit.py::TestMixTracingScripting::test_call_script_fn_from_traced_module, test/test_jit.py::TestMixTracingScripting::test_call_script_module_from_traced_module, test/test_jit.py::TestMixTracingScripting::test_call_traced_fn_from_script_fn, test/test_jit.py::TestMixTracingScripting::test_call_traced_mod_from_script_fn, test/test_jit.py::TestMixTracingScripting::test_call_tracing_fn_from_script_module, test/test_jit.py::TestMixTracingScripting::test_call_tracing_mod_from_script_module, test/test_jit.py::TestMixTracingScripting::test_jit_trace_callfunction_return_shapes, test/test_jit.py::TestMixTracingScripting::test_script_inline_trace_multiple_args, test/test_jit.py::TestMixTracingScripting::test_trace_dict_mix_script, test/test_jit.py::TestMixTracingScripting::test_trace_hierarchy, test/test_jit.py::TestMixTracingScripting::test_trace_linear, test/test_jit.py::TestMixTracingScripting::test_trace_mixed_by_script_with_dict_output, test/test_jit.py::TestMixTracingScripting::test_trace_of_script, test/test_jit.py::TestMixTracingScripting::test_trace_parameter, test/test_jit.py::TestMixTracingScripting::test_trace_returning_dict_with_tensor_tuples, test/test_jit.py::TestMixTracingScripting::test_trace_script, test/test_jit.py::TestMixTracingScripting::test_trace_script_returning_complex_dict, test/test_jit.py::TestMixTracingScripting::test_trace_with_size, test/test_jit.py::TestMixTracingScripting::test_traced_module_contains_scripted_interface_types, test/test_jit.py::TestMixTracingScripting::test_traced_module_implements_interface, test/test_jit.py::TestMixTracingScripting::test_tracing_indexing, test/test_jit.py::TestMixTracingScripting::test_tracing_slicing, test/test_jit.py::TestRecursiveScript::test_attributes, test/test_jit.py::TestRecursiveScript::test_class_compile, test/test_jit.py::TestRecursiveScript::test_constants_with_final, test/test_jit.py::TestRecursiveScript::test_dir, test/test_jit.py::TestRecursiveScript::test_error_stack, test/test_jit.py::TestRecursiveScript::test_error_stack_annotation, test/test_jit.py::TestRecursiveScript::test_error_stack_class, test/test_jit.py::TestRecursiveScript::test_error_stack_module, test/test_jit.py::TestRecursiveScript::test_failed_function_compilation, test/test_jit.py::TestRecursiveScript::test_function_attribute_in_submodule, test/test_jit.py::TestRecursiveScript::test_ignore_class, test/test_jit.py::TestRecursiveScript::test_inferred_nonetype, test/test_jit.py::TestRecursiveScript::test_init_error, test/test_jit.py::TestRecursiveScript::test_inner_traced_module, test/test_jit.py::TestRecursiveScript::test_iterable_modules, test/test_jit.py::TestRecursiveScript::test_method_call, test/test_jit.py::TestRecursiveScript::test_module_basic, test/test_jit.py::TestRecursiveScript::test_module_function_export, test/test_jit.py::TestRecursiveScript::test_module_name, test/test_jit.py::TestRecursiveScript::test_module_repr, test/test_jit.py::TestRecursiveScript::test_optional_module, test/test_jit.py::TestRecursiveScript::test_override_instance_method_ignore, test/test_jit.py::TestRecursiveScript::test_prepare_scriptable_basic, test/test_jit.py::TestRecursiveScript::test_prepare_scriptable_cycle, test/test_jit.py::TestRecursiveScript::test_prepare_scriptable_iterable_modules, test/test_jit.py::TestRecursiveScript::test_python_function_attribute, test/test_jit.py::TestRecursiveScript::test_repeated_error_stack, test/test_jit.py::TestRecursiveScript::test_script_after_eval, test/test_jit.py::TestRecursiveScript::test_script_basic, test/test_jit.py::TestRecursiveScript::test_script_function_attribute, test/test_jit.py::TestRecursiveScript::test_script_loaded_module, test/test_jit.py::TestTypeSharing::test_assign_python_attr, test/test_jit.py::TestTypeSharing::test_basic, test/test_jit.py::TestTypeSharing::test_builtin_function_different, test/test_jit.py::TestTypeSharing::test_builtin_function_same, test/test_jit.py::TestTypeSharing::test_constants, test/test_jit.py::TestTypeSharing::test_diff_attr_values, test/test_jit.py::TestTypeSharing::test_failed_attribute_compilation, test/test_jit.py::TestTypeSharing::test_ignored_fns, test/test_jit.py::TestTypeSharing::test_linear, test/test_jit.py::TestTypeSharing::test_loaded_modules_work, test/test_jit.py::TestTypeSharing::test_module_dict_same_type_different_name, test/test_jit.py::TestTypeSharing::test_mutate_attr_value, test/test_jit.py::TestTypeSharing::test_param_vs_attribute, test/test_jit.py::TestTypeSharing::test_python_function_attribute_different, test/test_jit.py::TestTypeSharing::test_python_function_attribute_same, test/test_jit.py::TestTypeSharing::test_same_but_different_classes, test/test_jit.py::TestTypeSharing::test_script_function_attribute_different, test/test_jit.py::TestTypeSharing::test_script_function_attribute_same, test/test_jit.py::TestTypeSharing::test_script_module_containing_traced_module, test/test_jit.py::TestTypeSharing::test_submodules, test/test_jit.py::TestTypeSharing::test_tracing_gives_different_types, test/test_jit.py::TestTypeSharing::test_type_not_shared_ignored_attributes, test/test_jit.py::TestTypeSharing::test_type_shared_ignored_attributes, test/test_jit.py::TestTypeSharing::test_type_sharing_define_in_init, test/test_jit.py::TestTypeSharing::test_type_sharing_disabled, test/test_jit.py::TestLogging::test_bump_numeric_counter, test/test_jit.py::TestLogging::test_counter_aggregation, test/test_jit.py::TestLogging::test_logging_levels_set, test/test_jit.py::TestLogging::test_time_measurement_counter, test/test_jit.py::TestLogging::test_time_measurement_counter_script, test/test_jit.py::TestLogging::test_trace_numeric_counter, test/test_jit.py::TestBackends::test_errors, test/test_jit.py::TestBackends::test_execution, test/test_jit.py::TestBackends::test_save_load, test/test_jit.py::TestBackendsWithCompiler::test_errors, test/test_jit.py::TestBackendsWithCompiler::test_execution, test/test_jit.py::TestNnapiBackend::test_adaptive_avg_pool2d, test/test_jit.py::TestNnapiBackend::test_avg_pool2d, test/test_jit.py::TestNnapiBackend::test_cat, test/test_jit.py::TestNnapiBackend::test_compile_spec_santiy, test/test_jit.py::TestNnapiBackend::test_conv2d, test/test_jit.py::TestNnapiBackend::test_conv2d_transpose, test/test_jit.py::TestNnapiBackend::test_dequantize, test/test_jit.py::TestNnapiBackend::test_detach, test/test_jit.py::TestNnapiBackend::test_flatten, test/test_jit.py::TestNnapiBackend::test_hardtanh, test/test_jit.py::TestNnapiBackend::test_linear, test/test_jit.py::TestNnapiBackend::test_log_softmax, test/test_jit.py::TestNnapiBackend::test_max_pool2d, test/test_jit.py::TestNnapiBackend::test_mean, test/test_jit.py::TestNnapiBackend::test_multi_output, test/test_jit.py::TestNnapiBackend::test_pointwise_binary, test/test_jit.py::TestNnapiBackend::test_pointwise_binary_const, test/test_jit.py::TestNnapiBackend::test_pointwise_unary, test/test_jit.py::TestNnapiBackend::test_prelu, test/test_jit.py::TestNnapiBackend::test_qadd, test/test_jit.py::TestNnapiBackend::test_qlinear, test/test_jit.py::TestNnapiBackend::test_quantize, test/test_jit.py::TestNnapiBackend::test_reshape, test/test_jit.py::TestNnapiBackend::test_seblock_mul, test/test_jit.py::TestNnapiBackend::test_slice, test/test_jit.py::TestNnapiBackend::test_softmax, test/test_jit.py::TestNnapiBackend::test_tensor_input, test/test_jit.py::TestNnapiBackend::test_to, test/test_jit.py::TestNnapiBackend::test_unsqueeze, test/test_jit.py::TestNnapiBackend::test_upsample_nearest2d, test/test_jit.py::TestList::test_comprehension_iterable, test/test_jit.py::TestList::test_comprehension_out_type_not_in_type, test/test_jit.py::TestList::test_comprehensions_basic, test/test_jit.py::TestList::test_comprehensions_basic_float, test/test_jit.py::TestList::test_comprehensions_two_comps, test/test_jit.py::TestList::test_copy_list_immutable, test/test_jit.py::TestList::test_copy_list_mutable, test/test_jit.py::TestList::test_del, test/test_jit.py::TestList::test_dict_keyword_is_correctly_typed, test/test_jit.py::TestList::test_dict_keyword_with_dict_comprehension, test/test_jit.py::TestList::test_dict_keyword_with_dict_comprehension_and_kwargs, test/test_jit.py::TestList::test_dict_keyword_with_empty_dict_comprehension, test/test_jit.py::TestList::test_dict_keyword_with_empty_iterable, test/test_jit.py::TestList::test_dict_keyword_with_internal_aggregate_function, test/test_jit.py::TestList::test_dict_keyword_with_iterable, test/test_jit.py::TestList::test_dict_keyword_with_kwargs, test/test_jit.py::TestList::test_dict_keyword_with_kwargs_using_container_values, test/test_jit.py::TestList::test_dict_keyword_with_mapping, test/test_jit.py::TestList::test_dict_keyword_with_mapping_and_kwargs, test/test_jit.py::TestList::test_dict_keyword_with_mismatched_annotations, test/test_jit.py::TestList::test_dict_keyword_with_nested_call, test/test_jit.py::TestList::test_dict_keyword_with_previously_declared_variable, test/test_jit.py::TestList::test_dict_keyword_with_previously_declared_variable_and_kwargs, test/test_jit.py::TestList::test_extend_list_immutable, test/test_jit.py::TestList::test_extend_list_mutable, test/test_jit.py::TestList::test_in_check, test/test_jit.py::TestList::test_list_bool_conversion, test/test_jit.py::TestList::test_list_count, test/test_jit.py::TestList::test_list_count_not_existing, test/test_jit.py::TestList::test_list_gather, test/test_jit.py::TestList::test_list_index, test/test_jit.py::TestList::test_list_index_not_existing, test/test_jit.py::TestList::test_list_keyword, test/test_jit.py::TestList::test_list_len, test/test_jit.py::TestList::test_list_literal, test/test_jit.py::TestList::test_list_none, test/test_jit.py::TestList::test_list_ops, test/test_jit.py::TestList::test_list_slice, test/test_jit.py::TestList::test_list_sort, test/test_jit.py::TestList::test_list_unification_hint, test/test_jit.py::TestList::test_list_variance, test/test_jit.py::TestList::test_min_bool_list, test/test_jit.py::TestList::test_min_max_list, test/test_jit.py::TestList::test_min_max_single_list, test/test_jit.py::TestList::test_mutable_list_append, test/test_jit.py::TestList::test_mutable_list_append_2, test/test_jit.py::TestList::test_mutable_list_append_if, test/test_jit.py::TestList::test_mutable_list_append_if_else, test/test_jit.py::TestList::test_mutable_list_append_loop, test/test_jit.py::TestList::test_mutable_list_append_loop_if, test/test_jit.py::TestList::test_mutable_list_clear, test/test_jit.py::TestList::test_mutable_list_clear_empty, test/test_jit.py::TestList::test_mutable_list_function_inline, test/test_jit.py::TestList::test_mutable_list_insert, test/test_jit.py::TestList::test_mutable_list_insert_neg_out_of_bounds, test/test_jit.py::TestList::test_mutable_list_insert_negative, test/test_jit.py::TestList::test_mutable_list_insert_out_of_bounds, test/test_jit.py::TestList::test_mutable_list_nested_loop, test/test_jit.py::TestList::test_mutable_list_pop, test/test_jit.py::TestList::test_mutable_list_pop2, test/test_jit.py::TestList::test_mutable_list_pop_at, test/test_jit.py::TestList::test_mutable_list_pop_at2, test/test_jit.py::TestList::test_mutable_list_pop_at_negative, test/test_jit.py::TestList::test_mutable_list_pop_at_negative2, test/test_jit.py::TestList::test_mutable_list_pop_empty, test/test_jit.py::TestList::test_mutable_list_pop_slice, test/test_jit.py::TestList::test_mutable_list_remove, test/test_jit.py::TestList::test_mutable_list_remove2, test/test_jit.py::TestList::test_mutable_list_remove_not_existing, test/test_jit.py::TestList::test_mutable_list_remove_tensor, test/test_jit.py::TestList::test_mutable_list_reverse, test/test_jit.py::TestList::test_mutable_list_reverse_empty, test/test_jit.py::TestList::test_mutable_tensor_list_reverse, test/test_jit.py::TestList::test_no_element_type_annotation, test/test_jit.py::TestList::test_slice_index, test/test_jit.py::TestList::test_tensor_list_count, test/test_jit.py::TestList::test_tensor_list_count_not_existing, test/test_jit.py::TestList::test_tensor_list_index, test/test_jit.py::TestList::test_tensor_list_index_not_existing, test/test_jit.py::TestList::test_to_list, test/test_jit.py::TestList::test_to_list_gpu, test/test_jit.py::TestDict::test_aug_assign, test/test_jit.py::TestDict::test_basic, test/test_jit.py::TestDict::test_clear, test/test_jit.py::TestDict::test_copy, test/test_jit.py::TestDict::test_del, test/test_jit.py::TestDict::test_dict_bool_conversion, test/test_jit.py::TestDict::test_dict_preserves_order, test/test_jit.py::TestDict::test_dict_to_python, test/test_jit.py::TestDict::test_dict_variance, test/test_jit.py::TestDict::test_get, test/test_jit.py::TestDict::test_get_boolkey, test/test_jit.py::TestDict::test_items, test/test_jit.py::TestDict::test_key_type, test/test_jit.py::TestDict::test_keys, test/test_jit.py::TestDict::test_len, test/test_jit.py::TestDict::test_loop, test/test_jit.py::TestDict::test_membership, test/test_jit.py::TestDict::test_mutability, test/test_jit.py::TestDict::test_optional_dict_construct, test/test_jit.py::TestDict::test_ordered_dict, test/test_jit.py::TestDict::test_pop, test/test_jit.py::TestDict::test_popitem, test/test_jit.py::TestDict::test_setdefault, test/test_jit.py::TestDict::test_type_annotation_missing_contained_type, test/test_jit.py::TestDict::test_update, test/test_jit.py::TestDict::test_update_existing_key, test/test_jit.py::TestDict::test_values, test/test_jit.py::TestDict::test_view, test/test_jit.py::TestNamedTuple::test_namedtuple, test/test_jit.py::TestNamedTuple::test_namedtuple_as_attr, test/test_jit.py::TestNamedTuple::test_namedtuple_constant, test/test_jit.py::TestNamedTuple::test_namedtuple_input_forwardref, test/test_jit.py::TestNamedTuple::test_namedtuple_inside_forwardref, test/test_jit.py::TestNamedTuple::test_namedtuple_kwarg_construct, test/test_jit.py::TestNamedTuple::test_namedtuple_lower, test/test_jit.py::TestNamedTuple::test_namedtuple_resolution, test/test_jit.py::TestNamedTuple::test_namedtuple_resolution_forwardref, test/test_jit.py::TestNamedTuple::test_namedtuple_serialization, test/test_jit.py::TestNamedTuple::test_namedtuple_slice_unpack, test/test_jit.py::TestNamedTuple::test_namedtuple_type_annotation, test/test_jit.py::TestNamedTuple::test_namedtuple_wrong_types, test/test_jit.py::TestNamedTuple::test_return_named_tuple, test/test_jit.py::TestScriptDict::test_bool, test/test_jit.py::TestScriptDict::test_contains, test/test_jit.py::TestScriptDict::test_delitem, test/test_jit.py::TestScriptDict::test_getitem, test/test_jit.py::TestScriptDict::test_items, test/test_jit.py::TestScriptDict::test_iter, test/test_jit.py::TestScriptDict::test_len, test/test_jit.py::TestScriptDict::test_nested, test/test_jit.py::TestScriptDict::test_reference_semantics, test/test_jit.py::TestScriptDict::test_repr, test/test_jit.py::TestScriptDict::test_setitem, test/test_jit.py::TestScriptList::test_append, test/test_jit.py::TestScriptList::test_bool, test/test_jit.py::TestScriptList::test_clear, test/test_jit.py::TestScriptList::test_contains, test/test_jit.py::TestScriptList::test_count, test/test_jit.py::TestScriptList::test_delitem, test/test_jit.py::TestScriptList::test_extend, test/test_jit.py::TestScriptList::test_getitem, test/test_jit.py::TestScriptList::test_insert, test/test_jit.py::TestScriptList::test_iter, test/test_jit.py::TestScriptList::test_len, test/test_jit.py::TestScriptList::test_nested, test/test_jit.py::TestScriptList::test_pop, test/test_jit.py::TestScriptList::test_reference_semantics, test/test_jit.py::TestScriptList::test_remove, test/test_jit.py::TestScriptList::test_repr, test/test_jit.py::TestScriptList::test_setitem, test/test_jit.py::TestAsync::test_async_future_type_python, test/test_jit.py::TestAsync::test_async_grad_guard_no_grad, test/test_jit.py::TestAsync::test_async_grad_guard_with_grad, test/test_jit.py::TestAsync::test_async_kwargs, test/test_jit.py::TestAsync::test_async_parsing, test/test_jit.py::TestAsync::test_async_python, test/test_jit.py::TestAsync::test_async_script, test/test_jit.py::TestAsync::test_async_script_capture, test/test_jit.py::TestAsync::test_async_script_error, test/test_jit.py::TestAsync::test_async_script_multi_forks, test/test_jit.py::TestAsync::test_async_script_multi_waits, test/test_jit.py::TestAsync::test_async_script_nested, test/test_jit.py::TestAsync::test_async_script_no_script_mod, test/test_jit.py::TestAsync::test_async_script_trace, test/test_jit.py::TestAsync::test_future_subtyping, test/test_jit.py::TestAsync::test_no_future_subtype_message, test/test_jit.py::TestAsync::test_trace_fork_wait, test/test_jit.py::TestAsync::test_trace_fork_wait_inline, test/test_jit.py::TestAsync::test_trace_fork_wait_leaking, test/test_jit.py::TestAsync::test_trace_fork_wait_list_modulecalls, test/test_jit.py::TestAsync::test_trace_modulecalls_with_different_output_types, test/test_jit.py::TestAwait::test_await_class_arg, test/test_jit.py::TestAwait::test_await_class_return, test/test_jit.py::TestAwait::test_await_eager_lazy, test/test_jit.py::TestAwait::test_await_func_arg, test/test_jit.py::TestAwait::test_await_getattr_implicit_convertion, test/test_jit.py::TestAwait::test_await_isinstance, test/test_jit.py::TestAwait::test_await_multiout_save, test/test_jit.py::TestAwait::test_await_nested, test/test_jit.py::TestAwait::test_await_out_of_interpreter, test/test_jit.py::TestAwait::test_await_python, test/test_jit.py::TestAwait::test_await_type_python, test/test_jit.py::TestAwait::test_awaitable_to_await, test/test_jit.py::TestAwait::test_eager_await_non_scriptable, test/test_jit.py::TestAwait::test_jit_trace, test/test_jit.py::TestAwait::test_nowait, test/test_jit.py::TestAwait::test_nowait_class, test/test_jit.py::TestAwait::test_script, test/test_jit.py::TestDataParallel::test_python_submodule_script, test/test_jit.py::TestDataParallel::test_shared_module, test/test_jit.py::TestDataParallel::test_tensor_sharing, test/test_jit.py::TestDataParallel::test_tensor_sharing_with_forward, test/test_jit.py::TestDataParallel::test_traced_module, test/test_jit.py::TestModels::test_alexnet, test/test_jit.py::TestModels::test_dcgan_models, test/test_jit.py::TestModels::test_dcgan_models_cuda, test/test_jit.py::TestModels::test_mnist, test/test_jit.py::TestModels::test_mnist_cuda, test/test_jit.py::TestModels::test_mnist_training_leaks_no_memory_cuda, test/test_jit.py::TestModels::test_neural_style, test/test_jit.py::TestModels::test_neural_style_cuda, test/test_jit.py::TestModels::test_reinforcement_learning, test/test_jit.py::TestModels::test_reinforcement_learning_cuda, test/test_jit.py::TestModels::test_script_module_script_resnet, test/test_jit.py::TestModels::test_script_module_trace_resnet18, test/test_jit.py::TestModels::test_snli, test/test_jit.py::TestModels::test_snli_cuda, test/test_jit.py::TestModels::test_snli_quantized, test/test_jit.py::TestModels::test_super_resolution, test/test_jit.py::TestModels::test_super_resolution_cuda, test/test_jit.py::TestModels::test_time_sequence_prediction, test/test_jit.py::TestModels::test_vae, test/test_jit.py::TestModels::test_vae_cuda, test/test_jit.py::TestModels::test_vae_quantized, test/test_jit.py::TestModules::test_script_module_with_constants_list, test/test_jit.py::TestAutodiffJit::test_autodiff_requires_grad_nograd, test/test_jit.py::TestAutodiffJit::test_requires_grad_outputs, test/test_jit.py::TestAutodiffJit::test_requires_grad_outputs_profiled_twice, test/test_jit.py::TestAutodiffJit::test_requires_grad_outputs_side_effects, test/test_jit.py::TestAutodiffJit::test_undefined_tensor_lists, test/test_jit.py::TestAutodiffSubgraphSlicing::test_aliased_outputs, test/test_jit.py::TestAutodiffSubgraphSlicing::test_bias_as_arg, test/test_jit.py::TestAutodiffSubgraphSlicing::test_bias_as_module_attr, test/test_jit.py::TestAutodiffSubgraphSlicing::test_chunk_constant_script_ad, test/test_jit.py::TestAutodiffSubgraphSlicing::test_constructed_bias, test/test_jit.py::TestAutodiffSubgraphSlicing::test_diff_graph_inline_threshold, test/test_jit.py::TestAutodiffSubgraphSlicing::test_differentiable_graph_ops_requires_grad, test/test_jit.py::TestAutodiffSubgraphSlicing::test_does_not_create_cycles, test/test_jit.py::TestAutodiffSubgraphSlicing::test_does_not_merge_unrelated, test/test_jit.py::TestAutodiffSubgraphSlicing::test_has_profiled_info_aliasing_outputs, test/test_jit.py::TestAutodiffSubgraphSlicing::test_merge_respects_aliasing, test/test_jit.py::TestAutodiffSubgraphSlicing::test_merges_dense, test/test_jit.py::TestAutodiffSubgraphSlicing::test_merges_down, test/test_jit.py::TestAutodiffSubgraphSlicing::test_merges_up, test/test_jit.py::TestAutodiffSubgraphSlicing::test_merges_without_cycles, test/test_jit.py::TestAutodiffSubgraphSlicing::test_prune_grad, test/test_jit.py::TestAutodiffSubgraphSlicing::test_requires_grad_for_tensor_list, test/test_jit.py::TestAutodiffSubgraphSlicing::test_respects_lexical_scoping, test/test_jit.py::TestAutodiffSubgraphSlicing::test_simple_merge, test/test_jit.py::TestAutodiffSubgraphSlicing::test_simple_no_merge, test/test_jit.py::TestCustomOperators::test_calling_scripted_custom_op, test/test_jit.py::TestCustomOperators::test_calling_traced_custom_op, test/test_jit.py::TestCustomOperators::test_default_arguments_are_used, test/test_jit.py::TestCustomOperators::test_dynamic_op_registry, test/test_jit.py::TestCustomOperators::test_generic_list, test/test_jit.py::TestCustomOperators::test_passing_and_returning_lists, test/test_jit.py::TestCustomOperators::test_passing_one_positional_but_not_the_second, test/test_jit.py::TestCustomOperators::test_passing_too_few_args, test/test_jit.py::TestCustomOperators::test_passing_too_many_args, test/test_jit.py::TestCustomOperators::test_passing_unknown_kwargs, test/test_jit.py::TestCustomOperators::test_script_graph_contains_custom_op, test/test_jit.py::TestCustomOperators::test_script_graph_for_custom_ops_matches_traced_graph, test/test_jit.py::TestCustomOperators::test_simply_calling_an_operator, test/test_jit.py::TestCustomOperators::test_where_no_scalar, test/test_jit.py::TestGraphRewritePasses::test_fuse_linear, test/test_jit.py::TestClassType::test_cast_overloads, test/test_jit.py::TestClassType::test_class_attribute_wrong_type, test/test_jit.py::TestClassType::test_class_constant, test/test_jit.py::TestClassType::test_class_constructs_itself, test/test_jit.py::TestClassType::test_class_inheritance, test/test_jit.py::TestClassType::test_class_inheritance_implicit, test/test_jit.py::TestClassType::test_class_sorting, test/test_jit.py::TestClassType::test_class_specialization, test/test_jit.py::TestClassType::test_class_type_as_param, test/test_jit.py::TestClassType::test_classmethod, test/test_jit.py::TestClassType::test_conditional_set_attr, test/test_jit.py::TestClassType::test_custom_delete, test/test_jit.py::TestClassType::test_default_args, test/test_jit.py::TestClassType::test_get_attr, test/test_jit.py::TestClassType::test_get_attr_not_initialized, test/test_jit.py::TestClassType::test_get_with_method, test/test_jit.py::TestClassType::test_imported_classes, test/test_jit.py::TestClassType::test_in, test/test_jit.py::TestClassType::test_init_compiled_first, test/test_jit.py::TestClassType::test_interface, test/test_jit.py::TestClassType::test_optional_type_promotion, test/test_jit.py::TestClassType::test_out_of_order_methods, test/test_jit.py::TestClassType::test_overloaded_fn, test/test_jit.py::TestClassType::test_properties, test/test_jit.py::TestClassType::test_py_class_to_ivalue_missing_attribute, test/test_jit.py::TestClassType::test_python_interop, test/test_jit.py::TestClassType::test_recursive_class, test/test_jit.py::TestClassType::test_recursive_script_builtin_type_resolution, test/test_jit.py::TestClassType::test_recursive_script_module_builtin_type_resolution, test/test_jit.py::TestClassType::test_recursive_scripting, test/test_jit.py::TestClassType::test_recursive_scripting_failed, test/test_jit.py::TestClassType::test_reference_semantics, test/test_jit.py::TestClassType::test_save_load_with_classes, test/test_jit.py::TestClassType::test_save_load_with_classes_nested, test/test_jit.py::TestClassType::test_save_load_with_classes_returned, test/test_jit.py::TestClassType::test_schema_human_readable, test/test_jit.py::TestClassType::test_self_referential_method, test/test_jit.py::TestClassType::test_set_attr_in_method, test/test_jit.py::TestClassType::test_set_attr_non_initialized, test/test_jit.py::TestClassType::test_set_attr_type_mismatch, test/test_jit.py::TestClassType::test_staticmethod, test/test_jit.py::TestClassType::test_type_annotation, test/test_jit.py::TestClassType::test_type_annotations, test/test_jit.py::TestClassType::test_unresolved_class_attributes, test/test_jit.py::TestClassType::test_unused_method, test/test_jit.py::TestBuiltins::test_del, test/test_jit.py::TestBuiltins::test_del_multiple_operands, test/test_jit.py::TestBuiltins::test_has_attr, test/test_jit.py::TestBuiltins::test_has_attr_invalid_args, test/test_jit.py::TestTensorBuiltins::test_method_on_number, test/test_jit.py::TestTensorBuiltins::test_scalar_to_num_conversions, test/test_jit.py::TestTensorBuiltins::test_tensor_item, test/test_jit.py::TestTensorBuiltins::test_tensor_properties, test/test_jit.py::TestTensorBuiltins::test_tensor_subscript_assign, test/test_jit.py::TestTensorBuiltins::test_tensor_subscript_assign_device, test/test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_inp_out, test/test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_just_inp, test/test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_just_out, test/test_jit.py::TestSymbolicShapeAnalysis::test_adaptive_avg_pool2d, test/test_jit.py::TestSymbolicShapeAnalysis::test_arange_shape, test/test_jit.py::TestSymbolicShapeAnalysis::test_binary_shape_fns_inplace, test/test_jit.py::TestSymbolicShapeAnalysis::test_binary_shape_functions, test/test_jit.py::TestSymbolicShapeAnalysis::test_convolution_backward, test/test_jit.py::TestSymbolicShapeAnalysis::test_if_propagation, test/test_jit.py::TestSymbolicShapeAnalysis::test_partial_eval_graph_conv, test/test_jit.py::TestSymbolicShapeAnalysis::test_partial_eval_stitching, test/test_jit.py::TestSymbolicShapeAnalysis::test_refinement_through_graph_stitching, test/test_jit.py::TestSymbolicShapeAnalysis::test_register_function_error_checking, test/test_jit.py::TestSymbolicShapeAnalysis::test_returning_input_symbolic_shapes, test/test_jit.py::TestSymbolicShapeAnalysis::test_shape_analysis, test/test_jit.py::TestSymbolicShapeAnalysis::test_shape_concat, test/test_jit.py::TestSymbolicShapeAnalysis::test_shape_embedding_bag, test/test_jit.py::TestSymbolicShapeAnalysis::test_shape_function_includes, test/test_jit.py::TestSymbolicShapeAnalysis::test_shared_shape_graph, test/test_jit.py::TestSymbolicShapeAnalysis::test_size_and_sizes, test/test_jit.py::TestSymbolicShapeAnalysis::test_stitching_concat, test/test_jit.py::TestSymbolicShapeAnalysis::test_stitching_multi_output, test/test_jit.py::TestSymbolicShapeAnalysis::test_sym_ir_parsing, test/test_jit.py::TestSymbolicShapeAnalysis::test_unary_shape_fns_inplace, test/test_jit.py::TestSymbolicShapeAnalysis::test_unary_shape_functions, test/test_jit.py::TestSymbolicShapeAnalysis::test_write, test/test_jit.py::TestOpDecompositions::test_op_decomposition, test/test_jit.py::TestOpDecompositions::test_registered_decomposition, test/test_jit.py::TestUnsupportedOps::test_factory_ops_requires_grad_fail, test/test_jit.py::TestUnsupportedOps::test_init_ops, test/test_jit.py::TestFreezing::test_freeze_interface_swapping_two_methods, test/test_jit.py::TestFreezing::test_freeze_interface_within_object, test/test_jit.py::TestFreezing::test_freeze_module, test/test_jit.py::TestFreezing::test_freeze_module_detach_gradient, test/test_jit.py::TestFreezing::test_freeze_module_in_training_mode, test/test_jit.py::TestFreezing::test_freeze_module_inlining, test/test_jit.py::TestFreezing::test_freeze_module_no_forward, test/test_jit.py::TestFreezing::test_freeze_module_return_self, test/test_jit.py::TestFreezing::test_freeze_module_return_sub_module, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr2, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr3, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr2, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr3, test/test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr4, test/test_jit.py::TestFreezing::test_freeze_module_with_call_method, test/test_jit.py::TestFreezing::test_freeze_module_with_fork, test/test_jit.py::TestFreezing::test_freeze_module_with_fork2, test/test_jit.py::TestFreezing::test_freeze_module_with_fork_calling_module_method, test/test_jit.py::TestFreezing::test_freeze_module_with_helperfunction, test/test_jit.py::TestFreezing::test_freeze_module_with_inplace_mutable, test/test_jit.py::TestFreezing::test_freeze_module_with_list, test/test_jit.py::TestFreezing::test_freeze_module_with_mutable_dict, test/test_jit.py::TestFreezing::test_freeze_module_with_mutable_list, test/test_jit.py::TestFreezing::test_freeze_module_with_mutable_tensor, test/test_jit.py::TestFreezing::test_freeze_module_with_nested_fork, test/test_jit.py::TestFreezing::test_freeze_module_with_nestedaliasing, test/test_jit.py::TestFreezing::test_freeze_module_with_nestedaliasingscalar, test/test_jit.py::TestFreezing::test_freeze_module_with_non_static_module_container_index, test/test_jit.py::TestFreezing::test_freeze_module_with_overlapping_attrs, test/test_jit.py::TestFreezing::test_freeze_module_with_preserve_sub_module, test/test_jit.py::TestFreezing::test_freeze_module_with_preserve_sub_module_and_mutation, test/test_jit.py::TestFreezing::test_freeze_module_with_sharedclasstype, test/test_jit.py::TestFreezing::test_freeze_module_with_submodule, test/test_jit.py::TestFreezing::test_freeze_module_with_tensor, test/test_jit.py::TestFreezing::test_freeze_module_with_tuple, test/test_jit.py::TestFreezing::test_freeze_module_with_tupleoutput_submodule, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attr, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attribute_on_submodule, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attribute_on_unused_submodule, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method2, test/test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method_on_submodule, test/test_jit.py::TestFreezing::test_freeze_no_forward, test/test_jit.py::TestFreezing::test_freeze_non_interface_module_swap, test/test_jit.py::TestFreezing::test_freeze_non_module_class_getattr, test/test_jit.py::TestFreezing::test_freeze_recursive_interfaces, test/test_jit.py::TestFreezing::test_freeze_recursive_interfaces_same_name, test/test_jit.py::TestFreezing::test_freeze_recursive_interfaces_with_reassignment, test/test_jit.py::TestFreezing::test_freeze_with_interface_mutable, test/test_jit.py::TestFreezing::test_freeze_with_swapping_interfaces, test/test_jit.py::TestFreezing::test_module_getattr_indirection, test/test_jit.py::TestFreezing::test_module_with_shared_type_instances, test/test_jit.py::TestFrozenOptimizations::test_collapse_adjacent_conversions, test/test_jit.py::TestFrozenOptimizations::test_conv_add_folding, test/test_jit.py::TestFrozenOptimizations::test_conv_bn_folding, test/test_jit.py::TestFrozenOptimizations::test_conv_bn_folding_autocast_scenario_cuda, test/test_jit.py::TestFrozenOptimizations::test_conv_bn_folding_not_forward, test/test_jit.py::TestFrozenOptimizations::test_conv_hardswish, test/test_jit.py::TestFrozenOptimizations::test_conv_mul_add_bn, test/test_jit.py::TestFrozenOptimizations::test_conv_to_mkldnn, test/test_jit.py::TestFrozenOptimizations::test_conv_to_mkldnn_no_mkldnn, test/test_jit.py::TestFrozenOptimizations::test_freeze_conv_relu_fusion, test/test_jit.py::TestFrozenOptimizations::test_freeze_conv_relu_fusion_not_forward, test/test_jit.py::TestFrozenOptimizations::test_freeze_mkdlnn, test/test_jit.py::TestFrozenOptimizations::test_freeze_remove_dropout, test/test_jit.py::TestFrozenOptimizations::test_freeze_remove_feature_dropout, test/test_jit.py::TestFrozenOptimizations::test_hardswish_hardsigmoid, test/test_jit.py::TestFrozenOptimizations::test_incompatible_perf_formats, test/test_jit.py::TestFrozenOptimizations::test_linear_bn_folding, test/test_jit.py::TestFrozenOptimizations::test_linear_bn_folding_autocast_scenario_cuda, test/test_jit.py::TestFrozenOptimizations::test_linear_concat, test/test_jit.py::TestFrozenOptimizations::test_linear_concat_complex, test/test_jit.py::TestFrozenOptimizations::test_linear_concat_different_input, test/test_jit.py::TestFrozenOptimizations::test_linear_multiple_blocks, test/test_jit.py::TestFrozenOptimizations::test_linear_non_constant_weight, test/test_jit.py::TestFrozenOptimizations::test_linear_transpose, test/test_jit.py::TestFrozenOptimizations::test_maxpool_mkldnn, test/test_jit.py::TestFrozenOptimizations::test_mkldnn_fuser_broadcasting, test/test_jit.py::TestFrozenOptimizations::test_mkldnn_inplace_removal, test/test_jit.py::TestFrozenOptimizations::test_numel_less_than_size_with_padding, test/test_jit.py::TestFrozenOptimizations::test_optimize_freeze_module, test/test_jit.py::TestFrozenOptimizations::test_pool2d_batchnorm, test/test_jit.py::TestFrozenOptimizations::test_pool3d_batchnorm, test/test_jit.py::TestFrozenOptimizations::test_remove_detach, test/test_jit.py::TestFrozenOptimizations::test_remove_detach_not_applied, test/test_jit.py::TestFrozenOptimizations::test_scalar_mul, test/test_jit.py::TestMKLDNNReinplacing::test_always_alive_values, test/test_jit.py::TestMKLDNNReinplacing::test_merge_liveness, test/test_jit.py::TestMKLDNNReinplacing::test_successful, test/test_jit.py::TestMKLDNNReinplacing::test_switch_inputs_to_inplace, test/test_jit.py::TestPeephole::test_conv_dim_folding, test/test_jit.py::TestPeephole::test_integer_refinement, test/test_jit.py::TestPeephole::test_noop_peephole, test/test_jit.py::TestPeephole::test_normalized_is_op, test/test_jit.py::TestPeephole::test_normalized_isnot_op, test/test_jit.py::TestPeephole::test_normalized_rsub, test/test_jit.py::TestPeephole::test_optimize_out_comparison_same_value, test/test_jit.py::TestPeephole::test_peephole, test/test_jit.py::TestPeephole::test_peephole_add_zero, test/test_jit.py::TestPeephole::test_peephole_arith, test/test_jit.py::TestPeephole::test_peephole_cuda, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_dict_modified, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_get_input_arg, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_keys_might_overlap, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_missing_key, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_overlapping_keys, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_unsupported_type, test/test_jit.py::TestPeephole::test_peephole_dict_getitem_simple, test/test_jit.py::TestPeephole::test_peephole_dict_len, test/test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_keys_might_overlap, test/test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_overlapping_keys, test/test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_unsupported_type, test/test_jit.py::TestPeephole::test_peephole_dynamic, test/test_jit.py::TestPeephole::test_peephole_int, test/test_jit.py::TestPeephole::test_peephole_len_list, test/test_jit.py::TestPeephole::test_peephole_list_len, test/test_jit.py::TestPeephole::test_peephole_list_ops, test/test_jit.py::TestPeephole::test_peephole_no_output_aliasing, test/test_jit.py::TestPeephole::test_peephole_optional_refine, test/test_jit.py::TestPeephole::test_peephole_slice_all_three_args, test/test_jit.py::TestPeephole::test_peephole_slice_one_empty_arg, test/test_jit.py::TestPeephole::test_peephole_slice_optimization_not_applied_list_modified, test/test_jit.py::TestPeephole::test_peephole_slice_optimization_not_applied_non_const_args, test/test_jit.py::TestPeephole::test_peephole_slice_two_empty_args, test/test_jit.py::TestPeephole::test_peephole_type_refinements, test/test_jit.py::TestPeephole::test_peephole_with_non_output_writes, test/test_jit.py::TestPeephole::test_peephole_with_writes, test/test_jit.py::TestPeephole::test_refine_integer_values, test/test_jit.py::TestPeephole::test_short_circuit_optimization, test/test_jit.py::TestAliasAnalysis::test_becomes_wildcard_annotations, test/test_jit.py::TestAliasAnalysis::test_nested_list_construct_not_wildcard, test/test_jit.py::TestAliasAnalysis::test_recursive_calls, test/test_jit.py::TestSaveLoad::test_different_functions, test/test_jit.py::TestSaveLoad::test_different_interfaces, test/test_jit.py::TestSaveLoad::test_different_modules, test/test_jit.py::TestSaveLoad::test_many_collisions, test/test_jit.py::TestSaveLoad::test_save_load_meta_tensors, test/test_jit.py::TestSaveLoad::test_save_load_params_buffers_submodules, test/test_jit.py::TestSaveLoad::test_save_load_using_pathlib, test/test_jit.py::TestSaveLoad::test_save_load_with_extra_files, test/test_jit.py::TestSaveLoad::test_save_load_with_saved_traced_inputs, test/test_jit.py::TestSaveLoad::test_save_namedtuple_input_only, test/test_jit.py::TestSaveLoad::test_save_namedtuple_input_only_forwardref, test/test_jit.py::TestSaveLoad::test_save_namedtuple_output_only, test/test_jit.py::TestSaveLoad::test_save_nonexit_file, test/test_jit.py::TestSaveLoadFlatbuffer::test_different_functions, test/test_jit.py::TestSaveLoadFlatbuffer::test_different_interfaces, test/test_jit.py::TestSaveLoadFlatbuffer::test_different_modules, test/test_jit.py::TestSaveLoadFlatbuffer::test_many_collisions, test/test_jit.py::TestSaveLoadFlatbuffer::test_module_info_flatbuffer, test/test_jit.py::TestSaveLoadFlatbuffer::test_save_load_params_buffers_submodules, test/test_jit.py::TestSaveLoadFlatbuffer::test_save_load_using_pathlib, test/test_jit.py::TestSaveLoadFlatbuffer::test_save_load_with_extra_files, test/test_jit.py::TestSaveLoadFlatbuffer::test_save_namedtuple_input_only, test/test_jit.py::TestSaveLoadFlatbuffer::test_save_namedtuple_output_only, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_inplace, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_reciprocal, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_scalar, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor_inplace, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor_out, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_linspace, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_linspace_out, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_logspace, test/test_jit.py::TestSaveLoadForOpVersion::test_versioned_logspace_out, test/test_jit.py::TestModuleContainers::test_custom_container_forward, test/test_jit.py::TestModuleContainers::test_empty_dict_override_contains, test/test_jit.py::TestModuleContainers::test_module_inplace_construct, test/test_jit.py::TestModuleContainers::test_module_interface_special_methods, test/test_jit.py::TestModuleContainers::test_module_properties, test/test_jit.py::TestModuleContainers::test_moduledict, test/test_jit.py::TestModuleContainers::test_moduledict_getitem, test/test_jit.py::TestModuleContainers::test_moduledict_keyerror, test/test_jit.py::TestModuleContainers::test_normal_list_attribute_with_modules_error, test/test_jit.py::TestModuleContainers::test_parameterdict_script_getitem, test/test_jit.py::TestModuleContainers::test_parameterlist_script_getitem, test/test_jit.py::TestModuleContainers::test_parameterlist_script_iter, test/test_jit.py::TestModuleContainers::test_script_module_list_sequential, test/test_jit.py::TestModuleContainers::test_script_modulelist_index, test/test_jit.py::TestModuleContainers::test_sequential_intermediary_types, test/test_jit.py::TestModuleContainers::test_special_method_with_override, test/test_jit.py::TestModuleContainers::test_typed_module_dict, test/test_jit.py::TestModuleContainers::test_typed_module_list, test/test_jit.py::TestPythonBindings::test_add_input, test/test_jit.py::TestPythonBindings::test_aliasdb, test/test_jit.py::TestPythonBindings::test_canonicalize, test/test_jit.py::TestPythonBindings::test_cu_create_function, test/test_jit.py::TestPythonBindings::test_cu_get_functions, test/test_jit.py::TestPythonBindings::test_graph_create, test/test_jit.py::TestPythonBindings::test_graph_iterator_keepalive, test/test_jit.py::TestPythonBindings::test_invalidation, test/test_jit.py::TestPythonIr::test_param_strides, test/test_jit.py::TestFunctionalBlocks::test_subgraph_creation, test/test_jit.py::TestRemoveMutation::test_aten_inplace, test/test_jit.py::TestRemoveMutation::test_common_pytorch_list_ops, test/test_jit.py::TestRemoveMutation::test_if_output, test/test_jit.py::TestRemoveMutation::test_if_output_fail, test/test_jit.py::TestRemoveMutation::test_list_indexing_removal, test/test_jit.py::TestRemoveMutation::test_lists_append, test/test_jit.py::TestRemoveMutation::test_lists_insert, test/test_jit.py::TestRemoveMutation::test_special_mapped_op, test/test_jit.py::TestTorchbind::test_default_args, test/test_jit.py::TestTorchbind::test_lambda_as_constructor, test/test_jit.py::TestTorchbind::test_profiler_custom_op, test/test_jit.py::TestTorchbind::test_staticmethod, test/test_jit.py::TestTorchbind::test_torchbind, test/test_jit.py::TestTorchbind::test_torchbind_attr_exception, test/test_jit.py::TestTorchbind::test_torchbind_class_attr_recursive, test/test_jit.py::TestTorchbind::test_torchbind_class_attribute, test/test_jit.py::TestTorchbind::test_torchbind_deepcopy, test/test_jit.py::TestTorchbind::test_torchbind_def_property_getter_setter, test/test_jit.py::TestTorchbind::test_torchbind_def_property_just_getter, test/test_jit.py::TestTorchbind::test_torchbind_def_property_readwrite, test/test_jit.py::TestTorchbind::test_torchbind_getattr, test/test_jit.py::TestTorchbind::test_torchbind_getstate, test/test_jit.py::TestTorchbind::test_torchbind_instantiate_missing_class, test/test_jit.py::TestTorchbind::test_torchbind_lambda_method, test/test_jit.py::TestTorchbind::test_torchbind_no_init, test/test_jit.py::TestTorchbind::test_torchbind_optional_explicit_attr, test/test_jit.py::TestTorchbind::test_torchbind_pass_wrong_type, test/test_jit.py::TestTorchbind::test_torchbind_pickle_serialization, test/test_jit.py::TestTorchbind::test_torchbind_python_deepcopy, test/test_jit.py::TestTorchbind::test_torchbind_return_instance, test/test_jit.py::TestTorchbind::test_torchbind_return_instance_from_method, test/test_jit.py::TestTorchbind::test_torchbind_return_tuple, test/test_jit.py::TestTorchbind::test_torchbind_save_load, test/test_jit.py::TestTorchbind::test_torchbind_take_as_arg, test/test_jit.py::TestTorchbind::test_torchbind_take_instance_as_method_arg, test/test_jit.py::TestTorchbind::test_torchbind_tracing, test/test_jit.py::TestTorchbind::test_torchbind_tracing_nested, test/test_jit.py::TestModuleInterface::test_freeze_module_with_inplace_mutation_in_interface, test/test_jit.py::TestModuleInterface::test_freeze_module_with_interface, test/test_jit.py::TestModuleInterface::test_freeze_module_with_interface_and_fork, test/test_jit.py::TestModuleInterface::test_freeze_module_with_mutated_interface, test/test_jit.py::TestModuleInterface::test_freeze_module_with_setattr_in_interface, test/test_jit.py::TestModuleInterface::test_module_apis_interface, test/test_jit.py::TestModuleInterface::test_module_doc_string, test/test_jit.py::TestModuleInterface::test_module_interface, test/test_jit.py::TestModuleInterface::test_module_interface_inheritance, test/test_jit.py::TestModuleInterface::test_module_interface_subtype, test/test_jit.py::TestModuleInterface::test_module_swap, test/test_jit.py::TestModuleInterface::test_module_swap_no_lazy_compile, test/test_jit.py::TestModuleInterface::test_module_swap_no_module_interface, test/test_jit.py::TestModuleInterface::test_module_swap_wrong_module, test/test_jit.py::TestModuleInterface::test_not_submodule_interface_call, test/test_jit.py::TestModuleInterface::test_script_module_as_interface_swap, test/test_jit.py::TestWith::test_with_as, test/test_jit.py::TestWith::test_with_errors, test/test_jit.py::TestWith::test_with_exceptions, test/test_jit.py::TestWith::test_with_no_as, test/test_jit.py::TestWith::test_with_no_grad, test/test_jit.py::TestWith::test_with_record_function, test/test_jit.py::TestEnum::test_closed_over_enum_constant, test/test_jit.py::TestEnum::test_enum_as_const, test/test_jit.py::TestEnum::test_enum_as_module_attribute, test/test_jit.py::TestEnum::test_enum_comp, test/test_jit.py::TestEnum::test_enum_comp_diff_classes, test/test_jit.py::TestEnum::test_enum_explicit_script, test/test_jit.py::TestEnum::test_enum_iterate, test/test_jit.py::TestEnum::test_enum_ivalue_type, test/test_jit.py::TestEnum::test_enum_module_return, test/test_jit.py::TestEnum::test_enum_name, test/test_jit.py::TestEnum::test_enum_return, test/test_jit.py::TestEnum::test_enum_value, test/test_jit.py::TestEnum::test_enum_value_types, test/test_jit.py::TestEnum::test_heterogenous_value_type_enum_error, test/test_jit.py::TestEnum::test_non_existent_enum_value, test/test_jit.py::TestEnum::test_string_enum_as_module_attribute, test/test_jit.py::TestStringFormatting::test_modulo_operator, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_alternate_digit_placeholder, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_capital_exponent_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_char_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_true_string_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_digit_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_digit_placeholder_and_string_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_double_percent_in_string, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_exponent_placeholder_and_string_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_float_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_float_placeholder_and_float_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_lowercase_exponent_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_multiple_placeholders, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_percent_in_string, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_digit_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_format_string_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_string_variable, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_subscript, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_too_few_arguments, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_too_many_arguments, test/test_jit.py::TestStringFormatting::test_string_interpolation_with_unknown_format_specifier, test/test_jit.py::TestProfiler::test_aliasing_merge, test/test_jit.py::TestProfiler::test_autograd_fallback_graph, test/test_jit.py::TestProfiler::test_fallback_graph_not_specialized, test/test_jit.py::TestProfiler::test_iterative_fusion, test/test_jit.py::TestProfiler::test_local_fusion_strategy, test/test_jit.py::TestProfiler::test_not_fusing_scalar_ops, test/test_jit.py::TestProfiler::test_not_optimizing_property, test/test_jit.py::TestProfiler::test_specialize_backward, test/test_jit.py::TestProfiler::test_specialized_types, test/test_jit.py::TestProfiler::test_tensor_constant, test/test_jit.py::TestProfiler::test_tensor_type_not_determined_by_inputs, test/test_jit.py::TestProfiler::test_use_not_profiled, test/test_jit.py::TestSlice::test_module_list_slicing, test/test_jit.py::TestSlice::test_slice_as_variable, test/test_jit.py::TestSlice::test_slice_dynamic_index, test/test_jit.py::TestSlice::test_slice_kwarg, test/test_jit.py::TestSlice::test_slice_one_none, test/test_jit.py::TestSlice::test_slice_start_stop, test/test_jit.py::TestSlice::test_slice_start_stop_step, test/test_jit.py::TestSlice::test_slice_start_stop_with_none, test/test_jit.py::TestSlice::test_slice_stop_clipped, test/test_jit.py::TestSlice::test_slice_stop_only, test/test_jit.py::TestSlice::test_slice_stop_only_with_nones, test/test_jit.py::TestSlice::test_slice_string, test/test_jit.py::TestSlice::test_slice_tensor, test/test_jit.py::TestSlice::test_slice_tensor_multidim, test/test_jit.py::TestSlice::test_slice_tensor_multidim_with_dots, test/test_jit.py::TestSlice::test_slice_three_nones, test/test_jit.py::TestSlice::test_slice_two_nones, test/test_jit.py::TestSlice::test_tuple_slicing, test/test_jit.py::TestIgnorableArgs::test_add_out_ignorable_args, test/test_jit.py::TestIgnorableArgs::test_slice_ignorable_args_for_slice, test/test_jit.py::TestHooks::test_forward_tuple_input, test/test_jit.py::TestHooks::test_hook_compilation_hint, test/test_jit.py::TestHooks::test_hook_hook_name_collision, test/test_jit.py::TestHooks::test_hook_method_name_collision, test/test_jit.py::TestHooks::test_module_direct_forward_invocation, test/test_jit.py::TestHooks::test_module_forward_multiple_inputs, test/test_jit.py::TestHooks::test_module_forward_single_input, test/test_jit.py::TestHooks::test_module_hook_return_nothing, test/test_jit.py::TestHooks::test_module_multiple_hooks_multiple_inputs, test/test_jit.py::TestHooks::test_module_multiple_hooks_single_input, test/test_jit.py::TestHooks::test_module_no_forward_input, test/test_jit.py::TestHooks::test_module_same_hook_repeated, test/test_jit.py::TestHooks::test_submodule_called_directly_with_hooks, test/test_jit.py::TestHooks::test_submodule_direct_forward_invocation, test/test_jit.py::TestHooks::test_submodule_forward_multiple_inputs, test/test_jit.py::TestHooks::test_submodule_forward_single_input, test/test_jit.py::TestHooks::test_submodule_forward_single_input_return_not_tupled, test/test_jit.py::TestHooks::test_submodule_hook_return_nothing, test/test_jit.py::TestHooks::test_submodule_multiple_hooks_multiple_inputs, test/test_jit.py::TestHooks::test_submodule_multiple_hooks_single_input, test/test_jit.py::TestHooks::test_submodule_no_forward_input, test/test_jit.py::TestHooks::test_submodule_same_hook_repeated, test/test_jit.py::TestHooks::test_wrong_hook_signatures, test/test_jit.py::TestHooks::test_wrong_pre_hook_signatures, test/test_jit.py::TestWarn::test_warn, test/test_jit.py::TestWarn::test_warn_multiple_calls_multiple_warnings, test/test_jit.py::TestWarn::test_warn_multiple_calls_same_func_diff_stack, test/test_jit.py::TestWarn::test_warn_once_per_func, test/test_jit.py::TestWarn::test_warn_once_per_func_in_loop, test/test_jit.py::TestWarn::test_warn_only_once, test/test_jit.py::TestWarn::test_warn_only_once_in_loop_func, test/test_jit.py::TestIsinstance::test_bool, test/test_jit.py::TestIsinstance::test_dict, test/test_jit.py::TestIsinstance::test_dict_nested, test/test_jit.py::TestIsinstance::test_dict_no_contained_type, test/test_jit.py::TestIsinstance::test_dict_tensor, test/test_jit.py::TestIsinstance::test_empty_container_special_cases, test/test_jit.py::TestIsinstance::test_empty_container_throws_warning_in_eager, test/test_jit.py::TestIsinstance::test_float, test/test_jit.py::TestIsinstance::test_if_else, test/test_jit.py::TestIsinstance::test_in_if, test/test_jit.py::TestIsinstance::test_in_while_loop, test/test_jit.py::TestIsinstance::test_int, test/test_jit.py::TestIsinstance::test_list, test/test_jit.py::TestIsinstance::test_list_nested, test/test_jit.py::TestIsinstance::test_list_no_contained_type, test/test_jit.py::TestIsinstance::test_list_tensor, test/test_jit.py::TestIsinstance::test_list_tensor_type_true, test/test_jit.py::TestIsinstance::test_nontuple_container_rhs_throws_in_eager, test/test_jit.py::TestIsinstance::test_optional, test/test_jit.py::TestIsinstance::test_optional_nested, test/test_jit.py::TestIsinstance::test_optional_no_contained_type, test/test_jit.py::TestIsinstance::test_optional_none, test/test_jit.py::TestIsinstance::test_tensor_type_false, test/test_jit.py::TestIsinstance::test_tuple, test/test_jit.py::TestIsinstance::test_tuple_nested, test/test_jit.py::TestIsinstance::test_tuple_no_contained_type, test/test_jit.py::TestIsinstance::test_tuple_rhs, test/test_jit.py::TestIsinstance::test_tuple_tensor, test/test_jit.py::TestIsinstance::test_type_refinement, test/test_jit.py::TestPythonBuiltinOP::test_add, test/test_jit.py::TestPythonBuiltinOP::test_adv_indexing_list, test/test_jit.py::TestPythonBuiltinOP::test_advancedindex, test/test_jit.py::TestPythonBuiltinOP::test_gather, test/test_jit.py::TestPythonBuiltinOP::test_index, test/test_jit.py::TestPythonBuiltinOP::test_index_ellipses, test/test_jit.py::TestPythonBuiltinOP::test_inf, test/test_jit.py::TestPythonBuiltinOP::test_matmul_py3, test/test_jit.py::TestPythonBuiltinOP::test_mul, test/test_jit.py::TestPythonBuiltinOP::test_pow, test/test_jit.py::TestPythonBuiltinOP::test_random, test/test_jit.py::TestPythonBuiltinOP::test_slice, test/test_jit.py::TestPythonBuiltinOP::test_stepped_tuple_slicing, test/test_jit.py::TestPythonBuiltinOP::test_str_to_float, test/test_jit.py::TestPythonBuiltinOP::test_triple, test/test_jit.py::TestTyping::test_bool_list_io, test/test_jit.py::TestTyping::test_dict_comprehension, test/test_jit.py::TestTyping::test_dict_comprehension_scope, test/test_jit.py::TestTyping::test_dict_comprehension_with_type_annotation, test/test_jit.py::TestTyping::test_dict_in_not_in, test/test_jit.py::TestTyping::test_dict_invalid_annotations, test/test_jit.py::TestTyping::test_dict_type_refinement_annotation_key_mismatch, test/test_jit.py::TestTyping::test_dict_type_refinement_annotation_value_mismatch, test/test_jit.py::TestTyping::test_for_in_dict, test/test_jit.py::TestTyping::test_for_in_string, test/test_jit.py::TestTyping::test_for_tuple_assign, test/test_jit.py::TestTyping::test_for_tuple_unpack, test/test_jit.py::TestTyping::test_list_io, test/test_jit.py::TestTyping::test_list_iterables, test/test_jit.py::TestTyping::test_list_sum, test/test_jit.py::TestTyping::test_list_type_refinement_annotation_element_mismatch, test/test_jit.py::TestTyping::test_list_unification, test/test_jit.py::TestTyping::test_multiple_assign, test/test_jit.py::TestTyping::test_namedtuple_error_source_attribution, test/test_jit.py::TestTyping::test_namedtuple_good_error, test/test_jit.py::TestTyping::test_namedtuple_py2, test/test_jit.py::TestTyping::test_namedtuple_redefine, test/test_jit.py::TestTyping::test_nested_list, test/test_jit.py::TestTyping::test_opt_opt_refinement, test/test_jit.py::TestTyping::test_optional_conversion, test/test_jit.py::TestTyping::test_optional_refinement, test/test_jit.py::TestTyping::test_optional_tuple, test/test_jit.py::TestTyping::test_singleton_tuple_unpack, test/test_jit.py::TestTyping::test_sum_list_diff_elms, test/test_jit.py::TestTyping::test_sum_list_empty, test/test_jit.py::TestTyping::test_sum_list_literal, test/test_jit.py::TestTyping::test_sum_list_one, test/test_jit.py::TestTyping::test_sum_list_wrong_type, test/test_jit.py::TestTyping::test_tuple_assignments, test/test_jit.py::TestTyping::test_tuple_create_return, test/test_jit.py::TestTyping::test_tuple_io, test/test_jit.py::TestTyping::test_tuple_keyword, test/test_jit.py::TestTyping::test_tuple_specialization, test/test_jit.py::TestHash::test_hash_bool, test/test_jit.py::TestHash::test_hash_device, test/test_jit.py::TestHash::test_hash_float, test/test_jit.py::TestHash::test_hash_int, test/test_jit.py::TestHash::test_hash_none, test/test_jit.py::TestHash::test_hash_string, test/test_jit.py::TestHash::test_hash_tensor, test/test_jit.py::TestHash::test_hash_tuple, test/test_jit.py::TestHash::test_hash_tuple_nested_unhashable_type, test/test_jit.py::TestComplex::test_binary_op_complex_tensor, test/test_jit.py::TestComplex::test_comparison_ops, test/test_jit.py::TestComplex::test_complex_constants_and_ops, test/test_jit.py::TestComplex::test_complex_constructor, test/test_jit.py::TestComplex::test_complex_list_sum, test/test_jit.py::TestComplex::test_complex_parse, test/test_jit.py::TestComplex::test_complexdict, test/test_jit.py::TestComplex::test_complexlist, test/test_jit.py::TestComplex::test_div, test/test_jit.py::TestComplex::test_infj_nanj_pickle, test/test_jit.py::TestComplex::test_pickle, test/test_jit.py::TestComplex::test_script, test/test_jit.py::TestComplex::test_tensor_attributes, test/test_jit.py::TestComplex::test_torch_complex_constructor_with_tensor, test/test_jit.py::TestJitUtils::test_checkscriptassertraisesregex, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_hybrid, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_keyword_only, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_positional_only, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_positional_or_keyword, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_var_keyword, test/test_jit.py::TestJitUtils::test_get_callable_argument_names_var_positional, test/test_jit.py::TestJitUtils::test_no_tracer_warn_context_manager, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_annotation_and_init_annotation, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_annotation_only, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_jit_annotation, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_dict, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_list, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_optional, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_tensor, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_falsy_base_type, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_nonempty_container, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_attribute, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_dict, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_list, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_optional, test/test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_torch_jit_import, test/test_jit.py::TestTypesAndAnnotation::test_annotate_outside_init, test/test_jit.py::TestTypesAndAnnotation::test_bad_types, test/test_jit.py::TestTypesAndAnnotation::test_ignore_with_types, test/test_jit.py::TestTypesAndAnnotation::test_ignoring_fn_with_nonscriptable_types, test/test_jit.py::TestTypesAndAnnotation::test_ignoring_module_attributes, test/test_jit.py::TestTypesAndAnnotation::test_inferred_type_error_message, test/test_jit.py::TestTypesAndAnnotation::test_mismatched_annotation, test/test_jit.py::TestTypesAndAnnotation::test_optional_no_element_type_annotation, test/test_jit.py::TestTypesAndAnnotation::test_parser_bug, test/test_jit.py::TestTypesAndAnnotation::test_pep585_type, test/test_jit.py::TestTypesAndAnnotation::test_python_callable, test/test_jit.py::TestTypesAndAnnotation::test_reannotate, test/test_jit.py::TestTypesAndAnnotation::test_tuple_no_element_type_annotation, test/test_jit.py::TestTypesAndAnnotation::test_type_annotate_py3, test/test_jit.py::TestTypesAndAnnotation::test_types_as_values, test/test_jit.py::TestTypesAndAnnotation::test_unimported_type_resolution, test/test_jit.py::TestMisc::test_broadcasting_list, test/test_jit.py::TestMisc::test_export_opnames_interface, test/test_jit.py::TestMisc::test_future_isinstance, test/test_jit.py::TestMisc::test_hacked_twin, test/test_jit.py::TestMisc::test_if_returning_any, test/test_jit.py::TestMisc::test_joined_str, test/test_jit.py::TestMisc::test_kwarg_support, test/test_jit.py::TestMisc::test_legacy_tensor_constructor, test/test_jit.py::TestMisc::test_list_literal_infer, test/test_jit.py::TestMisc::test_math_inf, test/test_jit.py::TestMisc::test_parse_ir_annotate, test/test_jit.py::TestMisc::test_parse_ir_single_element_tensor_negative, test/test_jit.py::TestMisc::test_parse_ir_single_element_tensor_positive, test/test_jit.py::TestMisc::test_pow_multiple_dtype, test/test_jit.py::TestMisc::test_script_many_decorators, test/test_jit.py::TestMisc::test_str_refine_any, test/test_jit.py::TestMisc::test_subexpression_Dict_int_Future, test/test_jit.py::TestMisc::test_subexpression_Future_annotate, test/test_jit.py::TestMisc::test_subexpression_List_Future, test/test_jit.py::TestMisc::test_subexpression_Optional, test/test_jit.py::TestMisc::test_subexpression_Tuple_int_int_Future, test/test_jit.py::TestMisc::test_tuple_subscripted_assign, test/test_jit.py::TestUpgraders::test_add_value_to_version_map, test/test_jit.py::TestUpgraders::test_aten_div_scalar_at_3, test/test_jit.py::TestUpgraders::test_aten_div_tensor_at_3, test/test_jit.py::TestUpgraders::test_aten_div_tensor_out_at_3, test/test_jit.py::TestUpgraders::test_aten_full_at_4, test/test_jit.py::TestUpgraders::test_aten_full_other_variants, test/test_jit.py::TestUpgraders::test_aten_full_out_at_4, test/test_jit.py::TestUpgraders::test_aten_linspace, test/test_jit.py::TestUpgraders::test_aten_linspace_out, test/test_jit.py::TestUpgraders::test_aten_logspace, test/test_jit.py::TestUpgraders::test_aten_logspace_out, test/test_jit.py::TestUpgraders::test_aten_test_serialization, test/test_jit.py::TestUpgraders::test_populated_test_upgrader_graph, test/test_jit.py::TestUpgraders::test_populated_upgrader_graph, test/test_jit.py::TestTensorCreationOps::test_randperm_default_dtype, test/test_jit.py::TestTensorCreationOps::test_randperm_specifed_dtype, test/test_jit.py::TestTensorCreationOps::test_tril_indices_default_dtype, test/test_jit.py::TestTensorCreationOps::test_tril_indices_specified_dtype, test/test_jit.py::TestTensorCreationOps::test_triu_indices_default_dtype, test/test_jit.py::TestTensorCreationOps::test_triu_indices_specified_dtype, test/test_jit.py::TestModuleAPIs::test_customized_state_dict_methods, test/test_jit.py::TestModuleAPIs::test_default_state_dict_methods, test/test_jit.py::TestModuleAPIs::test_submodule_customized_state_dict_methods, test/test_jit.py::TestScriptProfile::test_basic, test/test_jit.py::TestScriptProfile::test_empty, test/test_jit.py::TestScriptProfile::test_multi, test/test_jit.py::TestScriptProfile::test_script, test/test_jit.py::TestScriptProfile::test_section, test/test_jit.py::TestFunctionalToInplaceActivation::test_check_no_type_promotion, test/test_jit.py::TestFunctionalToInplaceActivation::test_functional_to_inplace_activation, test/test_jit.py::TestFunctionalToInplaceActivation::test_no_functional_to_inplace, test/test_jit.py::TestFunctionalToInplaceActivation::test_resnet18_correctness, test/test_jit.py::TestInplaceToFunctionalActivation::test_inplace_to_functional_activation, test/test_jit.py::TestInplaceToFunctionalActivation::test_resnet18_correctness, test/test_jit.py::TestParametrization::test_scriptable, test/test_jit.py::TestParametrization::test_traceable, test/test_jit.py::TestGetDefaultAttr::test_getattr_with_default, test/test_jit.py::TestAtenPow::test_aten_pow_zero_negative_exponent, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_1, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_2, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_3, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_4, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_insert_pre_packed_linear_before_inline_and_conv_2d_op, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_insert_pre_packed_linear_op, test/test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_replace_conv1d_with_conv2d, test/test_jit.py::TestUnion::test_check_union_annotation, test/test_jit.py::TestUnion::test_union_T_None_is_equivalent_to_optional_T, test/test_jit.py::TestUnion::test_union_argument_order_is_ignored, test/test_jit.py::TestUnion::test_union_argument_order_is_ignored_container, test/test_jit.py::TestUnion::test_union_as_annotation, test/test_jit.py::TestUnion::test_union_as_annotation_in_typed_container, test/test_jit.py::TestUnion::test_union_as_annotation_py2, test/test_jit.py::TestUnion::test_union_as_dict_key, test/test_jit.py::TestUnion::test_union_as_dict_value, test/test_jit.py::TestUnion::test_union_as_internal_tuple_type, test/test_jit.py::TestUnion::test_union_branching_does_not_autoinfer_undeclared_union, test/test_jit.py::TestUnion::test_union_branching_does_not_widen_existing_inferred_type, test/test_jit.py::TestUnion::test_union_branching_with_union_return_and_homogenous_types, test/test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type, test/test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type_empty_container, test/test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type_union, test/test_jit.py::TestUnion::test_union_in_class_constructor, test/test_jit.py::TestUnion::test_union_memory_aliasing, test/test_jit.py::TestUnion::test_union_module_with_union_class_variable, test/test_jit.py::TestUnion::test_union_module_with_union_instance_variable, test/test_jit.py::TestUnion::test_union_optional_of_union_is_flattened, test/test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped, test/test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_container, test/test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_optional, test/test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_subtyping, test/test_jit.py::TestUnion::test_union_return_type, test/test_jit.py::TestUnion::test_union_schema_matching_on_internal_type, test/test_jit.py::TestUnion::test_union_serialization_preserves_type_annotations, test/test_jit.py::TestUnion::test_union_subclasses_larger_union, test/test_jit.py::TestUnion::test_union_subtractive_refinement, test/test_jit.py::TestUnion::test_union_subtractive_refinement_with_container, test/test_jit.py::TestUnion::test_union_type_refinement, test/test_jit.py::TestUnion::test_union_type_refinement_internal_declaration, test/test_jit.py::TestUnion::test_union_type_refinement_partial_static_refinement_tuple_rhs, test/test_jit.py::TestUnion::test_union_type_refinement_partial_static_refinement_union_rhs, test/test_jit.py::TestUnion::test_union_type_refinement_statically_false, test/test_jit.py::TestUnion::test_union_type_refinement_statically_true, test/test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs, test/test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs_noncontained_type, test/test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs_union, test/test_jit.py::TestUnion::test_union_type_refinement_union_rhs, test/test_jit.py::TestUnion::test_union_variable_can_be_reassigned, test/test_jit.py::TestUnion::test_union_with_collections, test/test_jit.py::TestUnion::test_union_with_dict_assignment, test/test_jit.py::TestUnion::test_union_with_enum, test/test_jit.py::TestUnion::test_union_with_list_assignment, test/test_jit.py::TestUnion::test_union_with_scalar_values, test/test_jit.py::TestUnion::test_unions_of_a_single_argument_vanish, test/test_jit.py::TestUnion::test_unions_of_unions_are_flattened, test/test_jit.py::TestBatchMM::test_batch_mm_no_mutation, test/test_jit.py::TestBatchMM::test_batch_mm_permitted_mutation, test/test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation, test/test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation_if_node, test/test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation_multiple_adds, test/test_jit.py::TestBatchMM::test_batch_mm_side_permitted_mutation, test/test_jit.py::TestBatchMM::test_batch_mm_side_prohibited_mutation_common_side, test/test_jit.py::TestBatchMM::test_batch_mm_side_prohibited_mutation_uncommon_side, test/test_jit.py::TestDtypeAnalysis::test_binary_scalar, test/test_jit.py::TestDtypeAnalysis::test_binary_tensors, test/test_jit.py::TestDtypeAnalysis::test_combined, test/test_jit.py::TestDtypeAnalysis::test_conv_no_mixed_args, test/test_jit.py::TestDtypeAnalysis::test_custom_rules, test/test_jit.py::TestDtypeAnalysis::test_unary, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_expected_failure_nn_functional_adaptive_max_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_expected_failure_nn_functional_adaptive_max_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_complex128, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_complex64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_int64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool3d_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool3d_cpu_float64, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_bfloat16, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_float32, test/test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_float64, test/test_jit.py::TestDeviceAnalysis::test_custom_device_op, test/test_jit.py::TestDeviceAnalysis::test_device_apply, test/test_jit.py::TestDeviceAnalysis::test_device_arg, test/test_jit.py::TestDeviceAnalysis::test_device_if_propagation, test/test_jit.py::TestDeviceAnalysis::test_if_loop_mix, test/test_jit.py::TestDeviceAnalysis::test_loop_device_change, test/test_jit.py::TestDeviceAnalysis::test_loop_simple, test/test_jit.py::TestDeviceAnalysis::test_mobilenet, test/test_jit.py::TestDeviceAnalysis::test_nested_loops, test/test_jit.py::TestDeviceAnalysis::test_set_dtype, test/test_jit.py::TestDeviceAnalysis::test_simple, test/test_jit.py::TestDeviceAnalysis::test_tensor_as_fns, test/test_jit.py::TestDeviceAnalysis::test_while_change, test/test_jit.py::TestDeviceAnalysis::test_zerodim_cpu, test/test_jit.py::TestDeviceAnalysis::test_zerodim_gpu, test/test_jit.py::TestDeviceAnalysis::test_zerodim_no_device, test/test_jit.py::TestDCE::test_setattr_no_aliasdb, test/test_jit.py::TestDCE::test_setattr_removed, test/test_jit.py::TestSparse::test_freeze_sparse_coo, test/test_jit.py::TestSparse::test_freeze_sparse_csr, test/test_jit.py::TestSparse::test_serialize_sparse_coo, test/test_jit.py::TestSparse::test_serialize_sparse_csr, test/test_jit.py::TestTensorMethods::test_getitem, test/test_jit.py::TestTensorMethods::test_getitem_invalid, test/test_jit.py::TestDataclasses::test__post_init__, test/test_jit.py::TestDataclasses::test_comparators, test/test_jit.py::TestDataclasses::test_custom__eq__, test/test_jit.py::TestDataclasses::test_default_factories, test/test_jit.py::TestDataclasses::test_init_vars, test/test_jit.py::TestDataclasses::test_no_source, test/test_jit.py::TestDataclasses::test_use_unregistered_dataclass_raises, test/test_jit.py::TestJitProfiler::test_profiler, test/test_jit.py::TestJit::test_ModuleList, test/test_jit.py::TestJit::test_Sequential, test/test_jit.py::TestJit::test_T_mT_H_mH, test/test_jit.py::TestJit::test_add_relu_fusion, test/test_jit.py::TestJit::test_arg_configurations, test/test_jit.py::TestJit::test_attrs, test/test_jit.py::TestJit::test_batchnorm, test/test_jit.py::TestJit::test_big, test/test_jit.py::TestJit::test_conj_transpose, test/test_jit.py::TestJit::test_constant_insertion, test/test_jit.py::TestJit::test_constant_prop_aliasing_type, test/test_jit.py::TestJit::test_constant_prop_exception, test/test_jit.py::TestJit::test_constant_prop_if_constant, test/test_jit.py::TestJit::test_constant_prop_if_inline, test/test_jit.py::TestJit::test_constant_prop_loop_constant, test/test_jit.py::TestJit::test_constant_prop_nested, test/test_jit.py::TestJit::test_constant_prop_none, test/test_jit.py::TestJit::test_constant_prop_print, test/test_jit.py::TestJit::test_constant_prop_rand, test/test_jit.py::TestJit::test_constant_prop_remove_output, test/test_jit.py::TestJit::test_constant_prop_simple, test/test_jit.py::TestJit::test_constants_pkl, test/test_jit.py::TestJit::test_cpp, test/test_jit.py::TestJit::test_cse, test/test_jit.py::TestJit::test_cse_not_introduce_aliasing, test/test_jit.py::TestJit::test_cu_escaped_number, test/test_jit.py::TestJit::test_cuda_export_restore, test/test_jit.py::TestJit::test_debug_flush_compilation_cache, test/test_jit.py::TestJit::test_decompose_addmm, test/test_jit.py::TestJit::test_device_not_equal, test/test_jit.py::TestJit::test_diff_subgraph_clones_constants, test/test_jit.py::TestJit::test_disabled, test/test_jit.py::TestJit::test_dropout, test/test_jit.py::TestJit::test_dropout_cuda, test/test_jit.py::TestJit::test_dropout_func_requires_grad, test/test_jit.py::TestJit::test_dropout_module_requires_grad, test/test_jit.py::TestJit::test_einsum, test/test_jit.py::TestJit::test_element_size, test/test_jit.py::TestJit::test_expand_fold_quant_inputs, test/test_jit.py::TestJit::test_expand_quantlint, test/test_jit.py::TestJit::test_export_batchnorm, test/test_jit.py::TestJit::test_export_dropout, test/test_jit.py::TestJit::test_export_lstm, test/test_jit.py::TestJit::test_export_opnames, test/test_jit.py::TestJit::test_export_rnn, test/test_jit.py::TestJit::test_flags, test/test_jit.py::TestJit::test_function_default_values, test/test_jit.py::TestJit::test_hide_source_ranges_context_manager, test/test_jit.py::TestJit::test_import_method, test/test_jit.py::TestJit::test_inferred_as_tensor, test/test_jit.py::TestJit::test_layout, test/test_jit.py::TestJit::test_matrix_conj_transpose, test/test_jit.py::TestJit::test_matrix_transpose, test/test_jit.py::TestJit::test_module_default_values, test/test_jit.py::TestJit::test_mutable_default_values, test/test_jit.py::TestJit::test_native_dropout_corner_case, test/test_jit.py::TestJit::test_nn_conv, test/test_jit.py::TestJit::test_nn_lp_pool1d, test/test_jit.py::TestJit::test_nn_lp_pool2d, test/test_jit.py::TestJit::test_nn_padding, test/test_jit.py::TestJit::test_nn_padding_functional, test/test_jit.py::TestJit::test_no_erroneous_warnings, test/test_jit.py::TestJit::test_non_ascii_string, test/test_jit.py::TestJit::test_numel, test/test_jit.py::TestJit::test_pattern_based_module_rewrite, test/test_jit.py::TestJit::test_pattern_based_rewrite, test/test_jit.py::TestJit::test_pattern_based_rewrite_with_source_range_preserved, test/test_jit.py::TestJit::test_peephole_optimize_shape_ops, test/test_jit.py::TestJit::test_permute_inputs_binding, test/test_jit.py::TestJit::test_pretty_printer, test/test_jit.py::TestJit::test_print_classes_module, test/test_jit.py::TestJit::test_print_op_module, test/test_jit.py::TestJit::test_print_torch_ops_modules, test/test_jit.py::TestJit::test_python_bindings, test/test_jit.py::TestJit::test_python_ir, test/test_jit.py::TestJit::test_python_ir_utils, test/test_jit.py::TestJit::test_python_ir_utils_graph, test/test_jit.py::TestJit::test_python_ivalue, test/test_jit.py::TestJit::test_pytorch_jit_env_off, test/test_jit.py::TestJit::test_recursive_cse, test/test_jit.py::TestJit::test_repeat_interleave_script, test/test_jit.py::TestJit::test_restore_device, test/test_jit.py::TestJit::test_restore_device_cuda, test/test_jit.py::TestJit::test_restore_shared_storage_on_cuda, test/test_jit.py::TestJit::test_script_autograd_grad, test/test_jit.py::TestJit::test_script_backward, test/test_jit.py::TestJit::test_script_backward_twice, test/test_jit.py::TestJit::test_script_fn_pkl, test/test_jit.py::TestJit::test_script_tensor_type, test/test_jit.py::TestJit::test_shape_analysis_broadcast, test/test_jit.py::TestJit::test_shape_analysis_masked_select, test/test_jit.py::TestJit::test_shape_analysis_unsqueeze_in_loop, test/test_jit.py::TestJit::test_sparse_csr_tensors, test/test_jit.py::TestJit::test_sparse_tensors, test/test_jit.py::TestJit::test_torch_complex, test/test_jit.py::TestJit::test_torch_load_error, test/test_jit.py::TestJit::test_torch_load_zipfile_check, test/test_jit.py::TestJit::test_torch_ops_kwonly, test/test_jit.py::TestJit::test_torch_ops_overloaded, test/test_jit.py::TestJit::test_torch_sum, test/test_jit.py::TestJit::test_trace_retains_train, test/test_jit.py::TestJit::test_train_eval, test/test_jit.py::TestJit::test_transpose, test/test_jit.py::TestJit::test_unchecked_cast, test/test_jit.py::TestJit::test_unique_state_dict, test/test_jit.py::TestJit::test_verify, test/test_jit.py::TestJit::test_warnings, test/test_jit.py::TestFrontend::test_dictionary_as_example_inputs_for_jit_trace, test/test_jit.py::TestFrontend::test_instancing_error, test/test_jit.py::TestScript::test_add_out, test/test_jit.py::TestScript::test_add_tuple_different_types, test/test_jit.py::TestScript::test_add_tuple_non_optional, test/test_jit.py::TestScript::test_add_tuple_optional, test/test_jit.py::TestScript::test_add_tuple_same_types, test/test_jit.py::TestScript::test_addmm_grad, test/test_jit.py::TestScript::test_alias_covariant_type_containers, test/test_jit.py::TestScript::test_all, test/test_jit.py::TestScript::test_annot_ast_mypy_fn, test/test_jit.py::TestScript::test_annot_ast_mypy_method, test/test_jit.py::TestScript::test_annot_ast_py3_fn, test/test_jit.py::TestScript::test_annot_ast_py3_method, test/test_jit.py::TestScript::test_annot_string_mypy_fn, test/test_jit.py::TestScript::test_annot_string_mypy_method, test/test_jit.py::TestScript::test_annot_string_py3_fn, test/test_jit.py::TestScript::test_annot_string_py3_method, test/test_jit.py::TestScript::test_annotated_script_fn, test/test_jit.py::TestScript::test_annotated_script_fn_arg_mismatch, test/test_jit.py::TestScript::test_annotated_script_fn_return_mismatch, test/test_jit.py::TestScript::test_annotated_script_method, test/test_jit.py::TestScript::test_annoying_doubles, test/test_jit.py::TestScript::test_any, test/test_jit.py::TestScript::test_assert_is_scripting_metacompile, test/test_jit.py::TestScript::test_assertion_optional_refinement, test/test_jit.py::TestScript::test_attr_module_constants, test/test_jit.py::TestScript::test_attr_qscheme_script, test/test_jit.py::TestScript::test_attribute_in_init, test/test_jit.py::TestScript::test_attribute_serialization, test/test_jit.py::TestScript::test_attribute_unpickling, test/test_jit.py::TestScript::test_augmented_assign, test/test_jit.py::TestScript::test_autodiff_complex, test/test_jit.py::TestScript::test_backend_cudnn_enabled, test/test_jit.py::TestScript::test_bad_multiline_annotations, test/test_jit.py::TestScript::test_bailout_loop_carried_deps_name_clash, test/test_jit.py::TestScript::test_bailout_loop_counter_transition, test/test_jit.py::TestScript::test_batch_norm_inference_backward_cuda, test/test_jit.py::TestScript::test_batchnorm_fuser_cpu, test/test_jit.py::TestScript::test_big_float_literals, test/test_jit.py::TestScript::test_big_int_literals, test/test_jit.py::TestScript::test_binary_op_shape, test/test_jit.py::TestScript::test_bitwise_ops, test/test_jit.py::TestScript::test_block_input_grad_in_loop, test/test_jit.py::TestScript::test_bool_augassign_bitwise_and, test/test_jit.py::TestScript::test_bool_augassign_bitwise_or, test/test_jit.py::TestScript::test_bool_augassign_bitwise_xor, test/test_jit.py::TestScript::test_bool_dispatch, test/test_jit.py::TestScript::test_boolean_literal_constant_metacompile, test/test_jit.py::TestScript::test_break_continue_error, test/test_jit.py::TestScript::test_breaks_continues, test/test_jit.py::TestScript::test_builtin_args, test/test_jit.py::TestScript::test_builtin_args_fails, test/test_jit.py::TestScript::test_builtin_function_attributes, test/test_jit.py::TestScript::test_builtin_use_as_value, test/test_jit.py::TestScript::test_call_ge, test/test_jit.py::TestScript::test_call_python_fn_from_script_fn, test/test_jit.py::TestScript::test_call_python_fn_from_script_module, test/test_jit.py::TestScript::test_call_python_fn_from_traced_module, test/test_jit.py::TestScript::test_call_python_fn_from_tracing_fn, test/test_jit.py::TestScript::test_call_python_mod_from_script_fn, test/test_jit.py::TestScript::test_call_python_mod_from_script_module, test/test_jit.py::TestScript::test_call_python_mod_from_traced_module, test/test_jit.py::TestScript::test_call_python_mod_from_tracing_fn, test/test_jit.py::TestScript::test_call_script_fn_from_script_fn, test/test_jit.py::TestScript::test_call_script_fn_from_script_module, test/test_jit.py::TestScript::test_call_script_fn_from_tracing_fn, test/test_jit.py::TestScript::test_call_script_mod_from_script_fn, test/test_jit.py::TestScript::test_call_script_mod_from_script_module, test/test_jit.py::TestScript::test_call_script_mod_from_tracing_fn, test/test_jit.py::TestScript::test_call_traced_fn_from_tracing_fn, test/test_jit.py::TestScript::test_call_traced_mod_from_tracing_fn, test/test_jit.py::TestScript::test_calls_in_type_annotations, test/test_jit.py::TestScript::test_canonicalize_control_outputs, test/test_jit.py::TestScript::test_cast, test/test_jit.py::TestScript::test_cat, test/test_jit.py::TestScript::test_cat_lifts, test/test_jit.py::TestScript::test_chr, test/test_jit.py::TestScript::test_circular_dependency, test/test_jit.py::TestScript::test_class_as_attribute, test/test_jit.py::TestScript::test_class_attribute, test/test_jit.py::TestScript::test_class_attribute_in_script, test/test_jit.py::TestScript::test_class_with_comment_at_lower_indentation, test/test_jit.py::TestScript::test_code_with_constants, test/test_jit.py::TestScript::test_code_with_constants_restore, test/test_jit.py::TestScript::test_comment_ignore_indent, test/test_jit.py::TestScript::test_compare_two_bool_inputs, test/test_jit.py::TestScript::test_compile_module_with_constant, test/test_jit.py::TestScript::test_conditional_casting, test/test_jit.py::TestScript::test_constant_as_attr, test/test_jit.py::TestScript::test_constant_pooling_introduce_aliasing, test/test_jit.py::TestScript::test_constant_pooling_none, test/test_jit.py::TestScript::test_constant_pooling_same_identity, test/test_jit.py::TestScript::test_context_manager, test/test_jit.py::TestScript::test_conv_error, test/test_jit.py::TestScript::test_convert_base, test/test_jit.py::TestScript::test_cpp_function_tensor_str, test/test_jit.py::TestScript::test_cpp_module_iterator, test/test_jit.py::TestScript::test_desugar_module, test/test_jit.py::TestScript::test_device_kwarg, test/test_jit.py::TestScript::test_device_type, test/test_jit.py::TestScript::test_device_type_cuda, test/test_jit.py::TestScript::test_dict_str, test/test_jit.py::TestScript::test_dir, test/test_jit.py::TestScript::test_divmod, test/test_jit.py::TestScript::test_dominated_bailout, test/test_jit.py::TestScript::test_dropout_eval, test/test_jit.py::TestScript::test_dtype_attr, test/test_jit.py::TestScript::test_dtype_op_shape, test/test_jit.py::TestScript::test_dtype_op_shape2, test/test_jit.py::TestScript::test_early_return_closure, test/test_jit.py::TestScript::test_early_return_fork_join, test/test_jit.py::TestScript::test_early_return_rewrite, test/test_jit.py::TestScript::test_early_return_type_refinement, test/test_jit.py::TestScript::test_early_returns_loops, test/test_jit.py::TestScript::test_ellipsis_const_end, test/test_jit.py::TestScript::test_ellipsis_const_mid, test/test_jit.py::TestScript::test_ellipsis_const_mid_select, test/test_jit.py::TestScript::test_ellipsis_const_start, test/test_jit.py::TestScript::test_ellipsis_end, test/test_jit.py::TestScript::test_ellipsis_mid, test/test_jit.py::TestScript::test_ellipsis_mid_select, test/test_jit.py::TestScript::test_ellipsis_start, test/test_jit.py::TestScript::test_embedding_renorm_grad_error, test/test_jit.py::TestScript::test_empty_like_memory_format_bc, test/test_jit.py::TestScript::test_empty_tuple_str, test/test_jit.py::TestScript::test_enumerate_modlist_range, test/test_jit.py::TestScript::test_erase_number_types, test/test_jit.py::TestScript::test_error, test/test_jit.py::TestScript::test_error_stacktrace, test/test_jit.py::TestScript::test_error_stacktrace_interface, test/test_jit.py::TestScript::test_eval_python, test/test_jit.py::TestScript::test_exception_exits_closure, test/test_jit.py::TestScript::test_exceptions_with_control_flow, test/test_jit.py::TestScript::test_expand, test/test_jit.py::TestScript::test_fibb, test/test_jit.py::TestScript::test_fibb_totally_better, test/test_jit.py::TestScript::test_file_format_serialization, test/test_jit.py::TestScript::test_file_line_error, test/test_jit.py::TestScript::test_file_line_error_class_defn, test/test_jit.py::TestScript::test_file_line_graph, test/test_jit.py::TestScript::test_file_line_save_load, test/test_jit.py::TestScript::test_file_line_string, test/test_jit.py::TestScript::test_file_line_trace, test/test_jit.py::TestScript::test_filecheck, test/test_jit.py::TestScript::test_filecheck_parse, test/test_jit.py::TestScript::test_first_class_calls, test/test_jit.py::TestScript::test_first_class_module, test/test_jit.py::TestScript::test_floor_div, test/test_jit.py::TestScript::test_floordiv, test/test_jit.py::TestScript::test_for_else, test/test_jit.py::TestScript::test_for_in_dict, test/test_jit.py::TestScript::test_for_in_enumerate, test/test_jit.py::TestScript::test_for_in_range, test/test_jit.py::TestScript::test_for_in_range_ast, test/test_jit.py::TestScript::test_for_in_range_dynamic, test/test_jit.py::TestScript::test_for_in_range_if_ast, test/test_jit.py::TestScript::test_for_in_range_start_end, test/test_jit.py::TestScript::test_for_in_range_start_end_step, test/test_jit.py::TestScript::test_for_in_range_zero_step, test/test_jit.py::TestScript::test_for_in_string, test/test_jit.py::TestScript::test_for_in_tensors, test/test_jit.py::TestScript::test_for_in_tensors_fail_scalar, test/test_jit.py::TestScript::test_for_in_tensors_nested, test/test_jit.py::TestScript::test_for_in_tensors_rank0, test/test_jit.py::TestScript::test_for_in_zip, test/test_jit.py::TestScript::test_for_in_zip_enumerate, test/test_jit.py::TestScript::test_for_tuple_assign, test/test_jit.py::TestScript::test_for_tuple_unpack, test/test_jit.py::TestScript::test_format, test/test_jit.py::TestScript::test_func_call, test/test_jit.py::TestScript::test_function_compilation_caching, test/test_jit.py::TestScript::test_function_overload_misuse, test/test_jit.py::TestScript::test_function_overloading_isinstance, test/test_jit.py::TestScript::test_function_overloads, test/test_jit.py::TestScript::test_fuser_double_float_codegen, test/test_jit.py::TestScript::test_fuser_double_literal_precision, test/test_jit.py::TestScript::test_fuser_multiple_blocks, test/test_jit.py::TestScript::test_gather_dynamic_index, test/test_jit.py::TestScript::test_generic_list_errors, test/test_jit.py::TestScript::test_get_set_state, test/test_jit.py::TestScript::test_get_set_state_with_tensors, test/test_jit.py::TestScript::test_grad_from_script, test/test_jit.py::TestScript::test_hash, test/test_jit.py::TestScript::test_hex_literals, test/test_jit.py::TestScript::test_id, test/test_jit.py::TestScript::test_if, test/test_jit.py::TestScript::test_if_define, test/test_jit.py::TestScript::test_if_different_type, test/test_jit.py::TestScript::test_if_for_in_range, test/test_jit.py::TestScript::test_if_is_none_dispatch, test/test_jit.py::TestScript::test_if_list_cat, test/test_jit.py::TestScript::test_if_nest_while, test/test_jit.py::TestScript::test_if_noelse, test/test_jit.py::TestScript::test_if_not_defined_error, test/test_jit.py::TestScript::test_if_supertype, test/test_jit.py::TestScript::test_ignore_decorator, test/test_jit.py::TestScript::test_ignored_as_value, test/test_jit.py::TestScript::test_ignored_method_binding, test/test_jit.py::TestScript::test_ignored_props, test/test_jit.py::TestScript::test_import_constants_not_specialized, test/test_jit.py::TestScript::test_in_for_and_comp_expr, test/test_jit.py::TestScript::test_in_operator_with_two_strings, test/test_jit.py::TestScript::test_index, test/test_jit.py::TestScript::test_index_select_shape_prop, test/test_jit.py::TestScript::test_index_with_tuple, test/test_jit.py::TestScript::test_indexing_error, test/test_jit.py::TestScript::test_infer_size, test/test_jit.py::TestScript::test_inferred_error_msg, test/test_jit.py::TestScript::test_inherit_method, test/test_jit.py::TestScript::test_inline_and_run_annotated_script_fn, test/test_jit.py::TestScript::test_inlined_graph, test/test_jit.py::TestScript::test_inlining_cleanup, test/test_jit.py::TestScript::test_inplace_add, test/test_jit.py::TestScript::test_inplace_copy_script, test/test_jit.py::TestScript::test_input_keyword_in_schema, test/test_jit.py::TestScript::test_int_cast, test/test_jit.py::TestScript::test_integral_shape_inference, test/test_jit.py::TestScript::test_interpret_graph, test/test_jit.py::TestScript::test_interpreter_fuzz, test/test_jit.py::TestScript::test_intlist_args, test/test_jit.py::TestScript::test_invalid_call_arguments, test/test_jit.py::TestScript::test_invalid_lhs_assignment, test/test_jit.py::TestScript::test_invalid_prefix_annotation, test/test_jit.py::TestScript::test_irparser, test/test_jit.py::TestScript::test_is_after_use, test/test_jit.py::TestScript::test_is_isnot, test/test_jit.py::TestScript::test_is_optional, test/test_jit.py::TestScript::test_is_scripting, test/test_jit.py::TestScript::test_is_scripting_metacompile, test/test_jit.py::TestScript::test_isinstance, test/test_jit.py::TestScript::test_isinstance_dynamic, test/test_jit.py::TestScript::test_isinstance_metacompile, test/test_jit.py::TestScript::test_isinstance_refinement, test/test_jit.py::TestScript::test_jitter_bug, test/test_jit.py::TestScript::test_keyword, test/test_jit.py::TestScript::test_kwarg_expansion_error, test/test_jit.py::TestScript::test_kwargs_error_msg, test/test_jit.py::TestScript::test_lazy_script, test/test_jit.py::TestScript::test_lhs_advanced_indexing_assignment, test/test_jit.py::TestScript::test_lhs_advanced_indexing_augmented_assignment, test/test_jit.py::TestScript::test_lhs_indexing, test/test_jit.py::TestScript::test_lhs_indexing_increment, test/test_jit.py::TestScript::test_lhs_indexing_increment_list, test/test_jit.py::TestScript::test_lhs_indexing_increment_list_prim, test/test_jit.py::TestScript::test_lhs_indexing_list, test/test_jit.py::TestScript::test_lhs_indexing_multi, test/test_jit.py::TestScript::test_linear_grad, test/test_jit.py::TestScript::test_list_comprehension_modulelist, test/test_jit.py::TestScript::test_list_comprehension_variable_write, test/test_jit.py::TestScript::test_list_iterables, test/test_jit.py::TestScript::test_list_python_op, test/test_jit.py::TestScript::test_list_unify, test/test_jit.py::TestScript::test_literal, test/test_jit.py::TestScript::test_literals, test/test_jit.py::TestScript::test_logical_short_circuit, test/test_jit.py::TestScript::test_loop_liveness, test/test_jit.py::TestScript::test_loop_unroll_negative, test/test_jit.py::TestScript::test_loop_unroll_unused_counter, test/test_jit.py::TestScript::test_loop_unrolling, test/test_jit.py::TestScript::test_loop_unrolling_const, test/test_jit.py::TestScript::test_loop_unrolling_nested, test/test_jit.py::TestScript::test_lower_nested_tuples, test/test_jit.py::TestScript::test_math_ops, test/test_jit.py::TestScript::test_maxpool_guard_elimination, test/test_jit.py::TestScript::test_meshgrid, test/test_jit.py::TestScript::test_method_casts_script, test/test_jit.py::TestScript::test_method_no_self, test/test_jit.py::TestScript::test_method_overloading, test/test_jit.py::TestScript::test_missing_getstate, test/test_jit.py::TestScript::test_mm_batching, test/test_jit.py::TestScript::test_module_apis, test/test_jit.py::TestScript::test_module_attrs, test/test_jit.py::TestScript::test_module_copy_with_attributes, test/test_jit.py::TestScript::test_module_copying, test/test_jit.py::TestScript::test_module_error, test/test_jit.py::TestScript::test_module_method_reassignment, test/test_jit.py::TestScript::test_module_none_attrs, test/test_jit.py::TestScript::test_module_parameters_and_buffers, test/test_jit.py::TestScript::test_module_str, test/test_jit.py::TestScript::test_module_with_params_called_fails, test/test_jit.py::TestScript::test_multi_reduction, test/test_jit.py::TestScript::test_multi_starred_expr_lhs, test/test_jit.py::TestScript::test_multiline_annot_ast_py3_fn, test/test_jit.py::TestScript::test_multiline_optional_future_refinement, test/test_jit.py::TestScript::test_multiline_string_dedents, test/test_jit.py::TestScript::test_multiple_assign, test/test_jit.py::TestScript::test_multiple_assignment, test/test_jit.py::TestScript::test_mutable_dce, test/test_jit.py::TestScript::test_mutable_dce_block, test/test_jit.py::TestScript::test_mutable_dce_graph_input, test/test_jit.py::TestScript::test_mutable_dce_indirect_wildcard_write, test/test_jit.py::TestScript::test_mutable_dce_indirect_wildcards, test/test_jit.py::TestScript::test_mutable_dce_list, test/test_jit.py::TestScript::test_mutable_dce_loop, test/test_jit.py::TestScript::test_mutable_dce_wildcards, test/test_jit.py::TestScript::test_mutate_constant, test/test_jit.py::TestScript::test_mypy_type_ignore, test/test_jit.py::TestScript::test_named_buffers_are_iterable, test/test_jit.py::TestScript::test_namedtuple_attr, test/test_jit.py::TestScript::test_namedtuple_default_values_Tensor_type, test/test_jit.py::TestScript::test_namedtuple_default_values_container_type, test/test_jit.py::TestScript::test_namedtuple_default_values_missing, test/test_jit.py::TestScript::test_namedtuple_default_values_simple_type, test/test_jit.py::TestScript::test_namedtuple_default_values_using_factory_constructor, test/test_jit.py::TestScript::test_namedtuple_python, test/test_jit.py::TestScript::test_namedtuple_type_inference, test/test_jit.py::TestScript::test_narrow_copy, test/test_jit.py::TestScript::test_nested_aug_assign, test/test_jit.py::TestScript::test_nested_bailouts, test/test_jit.py::TestScript::test_nested_breaks, test/test_jit.py::TestScript::test_nested_list_construct, test/test_jit.py::TestScript::test_nested_select_assign, test/test_jit.py::TestScript::test_nn_GRU, test/test_jit.py::TestScript::test_nn_LSTM, test/test_jit.py::TestScript::test_nn_LSTM_with_layers, test/test_jit.py::TestScript::test_nn_init, test/test_jit.py::TestScript::test_no_dtype_shape, test/test_jit.py::TestScript::test_no_self_arg_ignore_function, test/test_jit.py::TestScript::test_non_final_return, test/test_jit.py::TestScript::test_none_type_str, test/test_jit.py::TestScript::test_not, test/test_jit.py::TestScript::test_not_initialized_err, test/test_jit.py::TestScript::test_ntuple_builtins, test/test_jit.py::TestScript::test_number_abs, test/test_jit.py::TestScript::test_number_augassign, test/test_jit.py::TestScript::test_number_augassign_bitwise_lshift, test/test_jit.py::TestScript::test_number_augassign_bitwise_pow, test/test_jit.py::TestScript::test_number_augassign_bitwise_rshift, test/test_jit.py::TestScript::test_number_div, test/test_jit.py::TestScript::test_number_math, test/test_jit.py::TestScript::test_number_neg, test/test_jit.py::TestScript::test_old_models_bc, test/test_jit.py::TestScript::test_oneline_func, test/test_jit.py::TestScript::test_op_dtype, test/test_jit.py::TestScript::test_operator_precedence, test/test_jit.py::TestScript::test_optional_list, test/test_jit.py::TestScript::test_optional_tensor, test/test_jit.py::TestScript::test_ord, test/test_jit.py::TestScript::test_override_magic, test/test_jit.py::TestScript::test_pack_tuple_into_non_var, test/test_jit.py::TestScript::test_pack_unpack_nested, test/test_jit.py::TestScript::test_pack_unpack_state, test/test_jit.py::TestScript::test_parameter_order, test/test_jit.py::TestScript::test_parse_empty_tuple_annotation, test/test_jit.py::TestScript::test_parse_empty_tuple_annotation_element_error, test/test_jit.py::TestScript::test_parse_nested_names, test/test_jit.py::TestScript::test_parse_none_type_annotation, test/test_jit.py::TestScript::test_parse_tensor_constants, test/test_jit.py::TestScript::test_parser_kwargonly, test/test_jit.py::TestScript::test_parser_type_annotations, test/test_jit.py::TestScript::test_parser_type_annotations_comment, test/test_jit.py::TestScript::test_parser_type_annotations_incompatible_expression, test/test_jit.py::TestScript::test_parser_type_annotations_subscript_non_ident, test/test_jit.py::TestScript::test_parser_type_annotations_subscript_tensor, test/test_jit.py::TestScript::test_parser_type_annotations_unknown_type, test/test_jit.py::TestScript::test_partial_returns, test/test_jit.py::TestScript::test_pass, test/test_jit.py::TestScript::test_pickle_checkpoint, test/test_jit.py::TestScript::test_pickle_checkpoint_cuda, test/test_jit.py::TestScript::test_pickle_checkpoint_tup, test/test_jit.py::TestScript::test_pow_scalar_backward_cuda, test/test_jit.py::TestScript::test_pretty_print_function, test/test_jit.py::TestScript::test_prim_grad_undefined, test/test_jit.py::TestScript::test_print, test/test_jit.py::TestScript::test_print_kwargs, test/test_jit.py::TestScript::test_profiling_graph_executor, test/test_jit.py::TestScript::test_profiling_merge, test/test_jit.py::TestScript::test_pybind_type_comparisons, test/test_jit.py::TestScript::test_python_call, test/test_jit.py::TestScript::test_python_call_annotation, test/test_jit.py::TestScript::test_python_call_annoytation_failure, test/test_jit.py::TestScript::test_python_call_failure, test/test_jit.py::TestScript::test_python_call_non_tensor, test/test_jit.py::TestScript::test_python_call_non_tensor_wrong, test/test_jit.py::TestScript::test_python_frontend, test/test_jit.py::TestScript::test_python_frontend_py3, test/test_jit.py::TestScript::test_python_frontend_source_range, test/test_jit.py::TestScript::test_python_op_builtins, test/test_jit.py::TestScript::test_python_op_name, test/test_jit.py::TestScript::test_python_val_doesnt_have_attr, test/test_jit.py::TestScript::test_rand, test/test_jit.py::TestScript::test_rand_profiling, test/test_jit.py::TestScript::test_range_args, test/test_jit.py::TestScript::test_reassign_module_lhs, test/test_jit.py::TestScript::test_reassign_module_rhs, test/test_jit.py::TestScript::test_refine_tuple_types, test/test_jit.py::TestScript::test_remove_dropout, test/test_jit.py::TestScript::test_repeated_script_on_function, test/test_jit.py::TestScript::test_request_bailout, test/test_jit.py::TestScript::test_requires_grad_loop, test/test_jit.py::TestScript::test_rescripting_loaded_modules, test/test_jit.py::TestScript::test_resize_input_ops, test/test_jit.py::TestScript::test_return, test/test_jit.py::TestScript::test_return_stmt_not_at_end, test/test_jit.py::TestScript::test_return_tuple, test/test_jit.py::TestScript::test_robust_op_resolution, test/test_jit.py::TestScript::test_round, test/test_jit.py::TestScript::test_save_load_attr_error, test/test_jit.py::TestScript::test_script_annotation, test/test_jit.py::TestScript::test_script_bool_constant, test/test_jit.py::TestScript::test_script_chunk, test/test_jit.py::TestScript::test_script_clamp_none, test/test_jit.py::TestScript::test_script_copy, test/test_jit.py::TestScript::test_script_cu, test/test_jit.py::TestScript::test_script_define_order, test/test_jit.py::TestScript::test_script_define_order_recursive_fail, test/test_jit.py::TestScript::test_script_docstring, test/test_jit.py::TestScript::test_script_forward_method_replacement, test/test_jit.py::TestScript::test_script_get_device_cuda, test/test_jit.py::TestScript::test_script_get_tracing_state, test/test_jit.py::TestScript::test_script_is_tracing, test/test_jit.py::TestScript::test_script_kwargs_fn_call, test/test_jit.py::TestScript::test_script_method_docstring, test/test_jit.py::TestScript::test_script_method_torch_function_overload, test/test_jit.py::TestScript::test_script_module, test/test_jit.py::TestScript::test_script_module_call_noscript, test/test_jit.py::TestScript::test_script_module_const, test/test_jit.py::TestScript::test_script_module_const_submodule_fail, test/test_jit.py::TestScript::test_script_module_export_blocks, test/test_jit.py::TestScript::test_script_module_export_shared_storage, test/test_jit.py::TestScript::test_script_module_export_submodule, test/test_jit.py::TestScript::test_script_module_export_tensor_cuda, test/test_jit.py::TestScript::test_script_module_export_tensor_type, test/test_jit.py::TestScript::test_script_module_fail_exist, test/test_jit.py::TestScript::test_script_module_for, test/test_jit.py::TestScript::test_script_module_for2, test/test_jit.py::TestScript::test_script_module_invalid_consts, test/test_jit.py::TestScript::test_script_module_nochange_submodule, test/test_jit.py::TestScript::test_script_module_none_exist_fail, test/test_jit.py::TestScript::test_script_module_not_tuple, test/test_jit.py::TestScript::test_script_module_param_buffer_mutation, test/test_jit.py::TestScript::test_script_module_star_assign2, test/test_jit.py::TestScript::test_script_module_star_assign2_inplace, test/test_jit.py::TestScript::test_script_module_star_assign_fail_builtin, test/test_jit.py::TestScript::test_script_module_star_assign_fail_pythonop, test/test_jit.py::TestScript::test_script_module_tensor_subclass_argument, test/test_jit.py::TestScript::test_script_nested_mod_list, test/test_jit.py::TestScript::test_script_non_tensor_args_outputs, test/test_jit.py::TestScript::test_script_optional_none, test/test_jit.py::TestScript::test_script_outputs, test/test_jit.py::TestScript::test_script_pack_padded_sequence, test/test_jit.py::TestScript::test_script_pad_sequence_pack_sequence, test/test_jit.py::TestScript::test_script_scope, test/test_jit.py::TestScript::test_script_sequential_for, test/test_jit.py::TestScript::test_script_sequential_in_mod_list, test/test_jit.py::TestScript::test_script_sequential_multi_output_fail, test/test_jit.py::TestScript::test_script_sequential_orderdict, test/test_jit.py::TestScript::test_script_sequential_sliced_iteration, test/test_jit.py::TestScript::test_script_star_assign, test/test_jit.py::TestScript::test_script_star_expr, test/test_jit.py::TestScript::test_script_star_expr_string, test/test_jit.py::TestScript::test_scriptable_fn_as_attr, test/test_jit.py::TestScript::test_scriptmodule_multi_head_attn_cuda, test/test_jit.py::TestScript::test_scriptmodule_releases_tensors_cuda, test/test_jit.py::TestScript::test_scriptmodule_transformer_cuda, test/test_jit.py::TestScript::test_select_after_chunk, test/test_jit.py::TestScript::test_sequence_parsing, test/test_jit.py::TestScript::test_sequential_intermediary_types, test/test_jit.py::TestScript::test_serialization_big_ints, test/test_jit.py::TestScript::test_serialization_sharing, test/test_jit.py::TestScript::test_serialize_long_lines, test/test_jit.py::TestScript::test_serialized_source_ranges, test/test_jit.py::TestScript::test_serialized_source_ranges2, test/test_jit.py::TestScript::test_serialized_source_ranges_dont_jitter, test/test_jit.py::TestScript::test_serialized_source_ranges_graph, test/test_jit.py::TestScript::test_serialized_source_ranges_no_dups, test/test_jit.py::TestScript::test_set_attribute_through_optional, test/test_jit.py::TestScript::test_shape_analysis_grad_property, test/test_jit.py::TestScript::test_shape_analysis_loop, test/test_jit.py::TestScript::test_shape_prop_promote_scalar_arg, test/test_jit.py::TestScript::test_shape_prop_promotion, test/test_jit.py::TestScript::test_signed_float_zero, test/test_jit.py::TestScript::test_single_starred_expr_for_loop, test/test_jit.py::TestScript::test_single_starred_lhs, test/test_jit.py::TestScript::test_singleton_tuple_unpack, test/test_jit.py::TestScript::test_slice_guard_elimination, test/test_jit.py::TestScript::test_split, test/test_jit.py::TestScript::test_stack, test/test_jit.py::TestScript::test_static_if_prop, test/test_jit.py::TestScript::test_static_method_on_module, test/test_jit.py::TestScript::test_static_methods, test/test_jit.py::TestScript::test_str_cast, test/test_jit.py::TestScript::test_string_cu, test/test_jit.py::TestScript::test_string_device_implicit_conversion, test/test_jit.py::TestScript::test_string_frontend_elif, test/test_jit.py::TestScript::test_string_index, test/test_jit.py::TestScript::test_string_len, test/test_jit.py::TestScript::test_string_list, test/test_jit.py::TestScript::test_string_new_line, test/test_jit.py::TestScript::test_string_ops, test/test_jit.py::TestScript::test_string_print, test/test_jit.py::TestScript::test_string_single_escape, test/test_jit.py::TestScript::test_string_slicing, test/test_jit.py::TestScript::test_string_sort, test/test_jit.py::TestScript::test_string_sorted, test/test_jit.py::TestScript::test_submodule_attribute_serialization, test/test_jit.py::TestScript::test_submodule_twice, test/test_jit.py::TestScript::test_sum, test/test_jit.py::TestScript::test_sum_list_diff_elms, test/test_jit.py::TestScript::test_sum_list_empty, test/test_jit.py::TestScript::test_sum_list_literal, test/test_jit.py::TestScript::test_sum_list_one, test/test_jit.py::TestScript::test_sum_list_wrong_type, test/test_jit.py::TestScript::test_sys_stdout_override, test/test_jit.py::TestScript::test_tensor_as_tensor_shape_prop, test/test_jit.py::TestScript::test_tensor_data, test/test_jit.py::TestScript::test_tensor_device, test/test_jit.py::TestScript::test_tensor_dtype, test/test_jit.py::TestScript::test_tensor_grad, test/test_jit.py::TestScript::test_tensor_import_export, test/test_jit.py::TestScript::test_tensor_len, test/test_jit.py::TestScript::test_tensor_number_math, test/test_jit.py::TestScript::test_tensor_number_math_cuda, test/test_jit.py::TestScript::test_tensor_requires_grad, test/test_jit.py::TestScript::test_tensor_shape, test/test_jit.py::TestScript::test_tensor_subclasses, test/test_jit.py::TestScript::test_tensor_to, test/test_jit.py::TestScript::test_tensor_to_cpu, test/test_jit.py::TestScript::test_tensor_to_cuda, test/test_jit.py::TestScript::test_tensor_to_device, test/test_jit.py::TestScript::test_ternary, test/test_jit.py::TestScript::test_ternary_module_type_hint, test/test_jit.py::TestScript::test_ternary_right_associative, test/test_jit.py::TestScript::test_ternary_static_if, test/test_jit.py::TestScript::test_torch_any, test/test_jit.py::TestScript::test_torch_functional, test/test_jit.py::TestScript::test_torch_functional_tensordot_int, test/test_jit.py::TestScript::test_torch_functional_tensordot_list, test/test_jit.py::TestScript::test_torch_functional_tensordot_tensor, test/test_jit.py::TestScript::test_torch_functional_tensordot_tuple, test/test_jit.py::TestScript::test_torch_ignore_conversion_to_none, test/test_jit.py::TestScript::test_torch_manual_seed, test/test_jit.py::TestScript::test_torch_pow, test/test_jit.py::TestScript::test_torch_tensor_as_tensor, test/test_jit.py::TestScript::test_torch_tensor_as_tensor_empty_list, test/test_jit.py::TestScript::test_torch_tensor_bad_input, test/test_jit.py::TestScript::test_torch_tensor_dtype, test/test_jit.py::TestScript::test_torchscript_memoryformat, test/test_jit.py::TestScript::test_torchscript_multi_head_attn, test/test_jit.py::TestScript::test_torchscript_multi_head_attn_fast_path, test/test_jit.py::TestScript::test_training_param, test/test_jit.py::TestScript::test_tuple_assignments, test/test_jit.py::TestScript::test_tuple_error_msg, test/test_jit.py::TestScript::test_tuple_index_to_list, test/test_jit.py::TestScript::test_tuple_indexing, test/test_jit.py::TestScript::test_tuple_len, test/test_jit.py::TestScript::test_tuple_nested_sort, test/test_jit.py::TestScript::test_tuple_sort, test/test_jit.py::TestScript::test_tuple_sort_reverse, test/test_jit.py::TestScript::test_tuple_sorted, test/test_jit.py::TestScript::test_tuple_str, test/test_jit.py::TestScript::test_tuple_to_opt_list, test/test_jit.py::TestScript::test_tuple_unsortable_diff_type, test/test_jit.py::TestScript::test_tuple_unsortable_element_type, test/test_jit.py::TestScript::test_tuple_unsortable_nested_diff_type, test/test_jit.py::TestScript::test_type_annotate, test/test_jit.py::TestScript::test_type_annotation_module, test/test_jit.py::TestScript::test_type_annotation_py3, test/test_jit.py::TestScript::test_type_annotations, test/test_jit.py::TestScript::test_type_annotations_repeated_list, test/test_jit.py::TestScript::test_type_annotations_varargs, test/test_jit.py::TestScript::test_type_call_in_script, test/test_jit.py::TestScript::test_type_cast, test/test_jit.py::TestScript::test_type_comments_in_body, test/test_jit.py::TestScript::test_type_inferred_from_empty_annotation, test/test_jit.py::TestScript::test_unbind, test/test_jit.py::TestScript::test_unfold_zero_dim, test/test_jit.py::TestScript::test_unicode_comments, test/test_jit.py::TestScript::test_uninitialized, test/test_jit.py::TestScript::test_union_to_number, test/test_jit.py::TestScript::test_unknown_builtin, test/test_jit.py::TestScript::test_unmatched_type_annotation, test/test_jit.py::TestScript::test_unspecialized_any_binding, test/test_jit.py::TestScript::test_unsqueeze_guard_elimination, test/test_jit.py::TestScript::test_unsupported_builtin_error, test/test_jit.py::TestScript::test_unused_decorator, test/test_jit.py::TestScript::test_unwrap_optional_builtin, test/test_jit.py::TestScript::test_var_aug_assign, test/test_jit.py::TestScript::test_vararg_zeros, test/test_jit.py::TestScript::test_view_listconstruct_shape_prop, test/test_jit.py::TestScript::test_view_shape_prop, test/test_jit.py::TestScript::test_view_write, test/test_jit.py::TestScript::test_weak_cuda, test/test_jit.py::TestScript::test_where, test/test_jit.py::TestScript::test_where_method, test/test_jit.py::TestScript::test_while, test/test_jit.py::TestScript::test_while_nest_if, test/test_jit.py::TestScript::test_while_nonexistent_cond_value, test/test_jit.py::TestScript::test_while_nonexistent_value, test/test_jit.py::TestScript::test_while_write_outer_then_read, test/test_jit.py::TestScript::test_wrong_attr_lookup, test/test_jit.py::TestScript::test_wrong_implicit_expand, test/test_jit.py::TestScript::test_wrong_method_call_inputs, test/test_jit.py::TestScript::test_wrong_module_attr_lookup, test/test_jit.py::TestScript::test_wrong_return_type, test/test_jit.py::TestScript::test_wrong_use_as_callable, test/test_jit.py::TestScript::test_wrong_use_as_tuple, test/test_jit.py::TestScript::test_zeros, test/test_jit.py::TestScript::test_zip_enumerate_modulelist, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_stride_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor_stride_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_stride_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride1_pad0_gpu_input, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_general_output, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride1_pad0_gpu_input, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_fixedkw_output, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_general_output, test/test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_input_nooverlap, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_scalar_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_legacy_enum, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_scalar_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_3d_input, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_3d_input_not_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_affine_simple_average, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_not_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_not_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_2d_simple_average, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_momentum, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_not_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_not_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_3d_simple_average, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_momentum, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_not_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_not_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_Bilinear, test/test_jit.py::TestJitGeneratedModule::test_nn_CELU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_int_target_lengths_intlists, test/test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_int_target_lengths_tensors, test/test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_lengths_tensors, test/test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_lengths_intlists, test/test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_lengths_tensors, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_circular_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_groups, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad1, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad1size1, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad2size1, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_valid, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_reflect_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_replicate_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_zeros_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_circular_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_padded, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_strided, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_with_multiplier, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_groups, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_groups_thnn, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_same, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_same_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_valid, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_padding, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_reflect_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_replicate_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_strided, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_zeros_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_1x1x1_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_circular_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_dilated_strided, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_groups, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_same, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_same_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_valid, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_replicate_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_stride_padding, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_zero_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_zeros_stride2_pad2, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_groups, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_groups, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose3d, test/test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose3d_dilated, test/test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_margin, test/test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_smoothing, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_4d_prob_target, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_4d_prob_target_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_dim_is_3, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_higher_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_CrossMapLRN2d, test/test_jit.py::TestJitGeneratedModule::test_nn_ELU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Embedding, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_discontiguous, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_max, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_max_padding_idx, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_mean_padding_idx, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sparse, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sum_padding_idx, test/test_jit.py::TestJitGeneratedModule::test_nn_Embedding_discontiguous, test/test_jit.py::TestJitGeneratedModule::test_nn_Embedding_sparse, test/test_jit.py::TestJitGeneratedModule::test_nn_Flatten, test/test_jit.py::TestJitGeneratedModule::test_nn_Flatten_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Fold, test/test_jit.py::TestJitGeneratedModule::test_nn_Fold_int_input, test/test_jit.py::TestJitGeneratedModule::test_nn_Fold_no_batch_dim_input, test/test_jit.py::TestJitGeneratedModule::test_nn_Fold_no_batch_dim_int_input, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_return_indices, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_asymsize, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_return_indices, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples, test/test_jit.py::TestJitGeneratedModule::test_nn_GELU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_GLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_GRUCell, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine_GN, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine_large_batch, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_no_affine_IN, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_no_affine_LN, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_affine_large_feature, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_IN, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_LN, test/test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_large_feature, test/test_jit.py::TestJitGeneratedModule::test_nn_Hardshrink_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Hardsigmoid_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Hardswish_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Hardtanh_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_margin, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_margin_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_scalar_margin, test/test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_delta, test/test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_tracking_stats_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_tracking_stats_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_tracking_stats, test/test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_tracking_stats_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_log_target, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_log_target, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_scalar_log_target, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_scalar_log_target, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_with_log_target_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_with_target_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce_complex, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d, test/test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d_norm, test/test_jit.py::TestJitGeneratedModule::test_nn_LPPool2d, test/test_jit.py::TestJitGeneratedModule::test_nn_LPPool2d_norm, test/test_jit.py::TestJitGeneratedModule::test_nn_LSTMCell, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_elementwise_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_empty_elementwise_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_no_elementwise_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_elementwise_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_no_affine_large_feature, test/test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_no_elementwise_affine, test/test_jit.py::TestJitGeneratedModule::test_nn_LeakyReLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Linear, test/test_jit.py::TestJitGeneratedModule::test_nn_Linear_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Linear_no_bias, test/test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_2d_uneven_pad, test/test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_3d_custom_params, test/test_jit.py::TestJitGeneratedModule::test_nn_LogSigmoid_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_prec, test/test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_margin, test/test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d_return_indices, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_3d_input, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_4d_input, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_return_indices, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_return_indices, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_stride, test/test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_stride_padding, test/test_jit.py::TestJitGeneratedModule::test_nn_Mish_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_0d_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_1d_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_index_neg, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_weights_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_1d_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_margin, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_margin_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_p, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_p_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_weights_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_MultiheadAttention, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_dim_is_3, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_higher_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights_ignore_index_neg, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights_ignore_index, test/test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights_ignore_index_neg, test/test_jit.py::TestJitGeneratedModule::test_nn_PReLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding122112_3dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding1221_2dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding12_1dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding2322_2dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding31_1dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding322112_3dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding332122_3dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding3331_2dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_Padding33_1dcircular, test/test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance, test/test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_broadcast_lhs, test/test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_broadcast_rhs, test/test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_with_non_default_args, test/test_jit.py::TestJitGeneratedModule::test_nn_PixelShuffle, test/test_jit.py::TestJitGeneratedModule::test_nn_PixelUnshuffle, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_full_loss, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_full_loss_no_log_input, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_full_loss, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_full_loss_no_log_input, test/test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_RNNCell, test/test_jit.py::TestJitGeneratedModule::test_nn_RReLU, test/test_jit.py::TestJitGeneratedModule::test_nn_RReLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_RReLU_with_up_down, test/test_jit.py::TestJitGeneratedModule::test_nn_RReLU_with_up_down_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_ReLU6_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_ReLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d, test/test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d_complex, test/test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_SELU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_SiLU_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Sigmoid_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_beta, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_reduce_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_zero_beta, test/test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss, test/test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_Softplus_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Softshrink_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Softsign_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Tanh_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Tanhshrink_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Threshold_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Transformer, test/test_jit.py::TestJitGeneratedModule::test_nn_TransformerDecoderLayer_gelu_activation, test/test_jit.py::TestJitGeneratedModule::test_nn_TransformerDecoderLayer_relu_activation, test/test_jit.py::TestJitGeneratedModule::test_nn_TransformerEncoderLayer_gelu_activation, test/test_jit.py::TestJitGeneratedModule::test_nn_TransformerEncoderLayer_relu_activation, test/test_jit.py::TestJitGeneratedModule::test_nn_Transformer_multilayer_coder, test/test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_mean, test/test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_none, test/test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_sum, test/test_jit.py::TestJitGeneratedModule::test_nn_Unflatten_no_batch_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_Unfold, test/test_jit.py::TestJitGeneratedModule::test_nn_Unfold_int_input, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_2d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_shared_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_skewed_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_skewed_2d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_tuple_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_tuple_2d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_2d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_shared_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_skewed_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_skewed_2d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_tuple_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_tuple_2d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_scale_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_scale_1d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_tuple_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_1d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d_launch_configs, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_3d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_1d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_2d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_3d_zero_dim, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_scale_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_scale_3d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_tuple_3d, test/test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_tuple_3d_align_corners, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_dim0, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_dim3, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_lastdim, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_spatial, test/test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_spatial_special, test/test_jit.py::TestJitGeneratedModule::test_nn_multimarginloss_1d_input_0d_target_no_reduce, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_dim0, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_dim3, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_scalar, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_lastdim, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_lastdim_dtype, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial_dtype, test/test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial_special, test/test_jit.py::TestProducerVersion::test_version 2023-03-31T05:50:35.9917745Z 2023-03-31T05:50:35.9918259Z test_jit.py::TestTracer::test_call_traced_fn_from_traced_module <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9918797Z test_jit.py::TestTracer::test_call_traced_module_from_traced_module <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9919289Z test_jit.py::TestTracer::test_canonicalize_tensor_iterator <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9919726Z test_jit.py::TestTracer::test_constant <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9920108Z test_jit.py::TestTracer::test_conv <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9920524Z test_jit.py::TestTracer::test_export_no_reorder <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9920964Z test_jit.py::TestTracer::test_force_outplace_check_fill <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9921396Z test_jit.py::TestTracer::test_force_outplace_check_zero <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9921804Z test_jit.py::TestTracer::test_ge <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9922238Z test_jit.py::TestTracer::test_ge_cuda <- test/jit/test_tracer.py SKIPPED (requires CUDA) [ 0%] 2023-03-31T05:50:35.9922789Z test_jit.py::TestTracer::test_ge_optimized <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/jit_utils.py PASSED [ 0%] 2023-03-31T05:50:35.9923513Z test_jit.py::TestTracer::test_ge_unoptimized <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9923922Z test_jit.py::TestTracer::test_index_put <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9924350Z test_jit.py::TestTracer::test_index_put_trace_with_view <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9924787Z test_jit.py::TestTracer::test_index_put_trace_without_view <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9925214Z test_jit.py::TestTracer::test_inplace_check <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9925622Z test_jit.py::TestTracer::test_inplace_copy <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9926066Z test_jit.py::TestTracer::test_inplace_copy_force_outplace <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9926478Z test_jit.py::TestTracer::test_inplace_flags <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9926911Z test_jit.py::TestTracer::test_inplace_transplant <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9927326Z test_jit.py::TestTracer::test_inplace_warn <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9927759Z test_jit.py::TestTracer::test_input_dict_checkTrace_mut <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9928174Z test_jit.py::TestTracer::test_input_dict_empty <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9928598Z test_jit.py::TestTracer::test_input_dict_empty_list <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9929043Z test_jit.py::TestTracer::test_input_dict_insertion_order <- test/jit/test_tracer.py PASSED [ 0%] 2023-03-31T05:50:35.9929589Z test_jit.py::TestTracer::test_input_dict_of_dicts <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9930358Z test_jit.py::TestTracer::test_input_dict_of_lists <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9930906Z test_jit.py::TestTracer::test_input_dict_recursive <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9931352Z test_jit.py::TestTracer::test_input_dict_remembers_keys <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9931767Z test_jit.py::TestTracer::test_input_dict_unify <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9932182Z test_jit.py::TestTracer::test_input_flatten <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9932610Z test_jit.py::TestTracer::test_input_list_mixed_type <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9933023Z test_jit.py::TestTracer::test_input_list_of_tuples <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9933460Z test_jit.py::TestTracer::test_input_list_toplevel_flatten <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9933930Z test_jit.py::TestTracer::test_input_list_toplevel_flatten_direct <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9934378Z test_jit.py::TestTracer::test_input_tuple_of_dicts <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9934794Z test_jit.py::TestTracer::test_interpolate_trace <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9935256Z test_jit.py::TestTracer::test_large_nbr_kernel_args <- test/jit/test_tracer.py SKIPPED (requires CUDA) [ 1%] 2023-03-31T05:50:35.9935694Z test_jit.py::TestTracer::test_lhs_index_fails <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9936114Z test_jit.py::TestTracer::test_lhs_index_trivial <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9936502Z test_jit.py::TestTracer::test_max_pool <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9936909Z test_jit.py::TestTracer::test_nested_inplace <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9937321Z test_jit.py::TestTracer::test_non_tensor_tracing <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9937727Z test_jit.py::TestTracer::test_output_unflatten <- test/jit/test_tracer.py XFAIL [ 1%] 2023-03-31T05:50:35.9938144Z test_jit.py::TestTracer::test_python_function <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9938567Z test_jit.py::TestTracer::test_python_function_tup <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9938983Z test_jit.py::TestTracer::test_repeated_input <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9939387Z test_jit.py::TestTracer::test_repeated_output <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9939799Z test_jit.py::TestTracer::test_shared_param <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9940199Z test_jit.py::TestTracer::test_simple <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9940618Z test_jit.py::TestTracer::test_tensor_with_grad_as_constant <- test/jit/test_tracer.py PASSED [ 1%] 2023-03-31T05:50:35.9941068Z test_jit.py::TestTracer::test_trace_aliased_parameter <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9941495Z test_jit.py::TestTracer::test_trace_annotation <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9941911Z test_jit.py::TestTracer::test_trace_arange <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9942323Z test_jit.py::TestTracer::test_trace_arange_with_grad <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9942756Z test_jit.py::TestTracer::test_trace_autograd_function <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9943371Z test_jit.py::TestTracer::test_trace_c10_ops <- test/jit/test_tracer.py SKIPPED (Skip the test since c2 ops are not registered.) [ 2%] 2023-03-31T05:50:35.9943827Z test_jit.py::TestTracer::test_trace_casts <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9944248Z test_jit.py::TestTracer::test_trace_checker_control_flow <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9944687Z test_jit.py::TestTracer::test_trace_checker_dot_data <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9945229Z test_jit.py::TestTracer::test_trace_checker_dropout_notrain <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9945762Z test_jit.py::TestTracer::test_trace_checker_dropout_train <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9946228Z test_jit.py::TestTracer::test_trace_checker_inplace_on_view <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9946681Z test_jit.py::TestTracer::test_trace_checker_memoization <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9947123Z test_jit.py::TestTracer::test_trace_checker_slice_lhs <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9947573Z test_jit.py::TestTracer::test_trace_checking_with_deprecated_name <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9948039Z test_jit.py::TestTracer::test_trace_checking_with_global_name <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9948467Z test_jit.py::TestTracer::test_trace_contiguous <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9948922Z test_jit.py::TestTracer::test_trace_contiguous_short_circuit <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9949339Z test_jit.py::TestTracer::test_trace_detach <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9949759Z test_jit.py::TestTracer::test_trace_detach_inplace <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9950212Z test_jit.py::TestTracer::test_trace_detach_inplace_redispatch <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9950650Z test_jit.py::TestTracer::test_trace_detach_redispatch <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9951073Z test_jit.py::TestTracer::test_trace_dict_input <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9951493Z test_jit.py::TestTracer::test_trace_dict_output <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9951914Z test_jit.py::TestTracer::test_trace_export_fns <- test/jit/test_tracer.py PASSED [ 2%] 2023-03-31T05:50:35.9952343Z test_jit.py::TestTracer::test_trace_export_fns_recursive <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9952789Z test_jit.py::TestTracer::test_trace_fork_join_and_module <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9953229Z test_jit.py::TestTracer::test_trace_full_dynamic_shape <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9953679Z test_jit.py::TestTracer::test_trace_func_argument_names_captured <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9954110Z test_jit.py::TestTracer::test_trace_index <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9954530Z test_jit.py::TestTracer::test_trace_index_constant <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9954969Z test_jit.py::TestTracer::test_trace_indexed_assignment <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9955389Z test_jit.py::TestTracer::test_trace_inline_shape <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9955808Z test_jit.py::TestTracer::test_trace_inverse <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9956249Z test_jit.py::TestTracer::test_trace_invert_module_hierarchy <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9956686Z test_jit.py::TestTracer::test_trace_legacy_ctor <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9957130Z test_jit.py::TestTracer::test_trace_module_argument_names_captured <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9957574Z test_jit.py::TestTracer::test_trace_modulelist <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9958019Z test_jit.py::TestTracer::test_trace_multi_output_function <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9958441Z test_jit.py::TestTracer::test_trace_namedtuple <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9958871Z test_jit.py::TestTracer::test_trace_nested_datatypes <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9959291Z test_jit.py::TestTracer::test_trace_nested_fn <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9959777Z test_jit.py::TestTracer::test_trace_numel <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9960245Z test_jit.py::TestTracer::test_trace_optioanl_dtype <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9960667Z test_jit.py::TestTracer::test_trace_optional <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9961135Z test_jit.py::TestTracer::test_trace_partial_func_argument_names_captured <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9961563Z test_jit.py::TestTracer::test_trace_random <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9961981Z test_jit.py::TestTracer::test_trace_records_names <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9962387Z test_jit.py::TestTracer::test_trace_save <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9962805Z test_jit.py::TestTracer::test_trace_save_load_copy <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9963342Z test_jit.py::TestTracer::test_trace_single_tuple <- test/jit/test_tracer.py PASSED [ 3%] 2023-03-31T05:50:35.9963751Z test_jit.py::TestTracer::test_trace_size <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9964176Z test_jit.py::TestTracer::test_trace_size_with_grad <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9964613Z test_jit.py::TestTracer::test_trace_skip_none_submodule <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9965019Z test_jit.py::TestTracer::test_trace_slice <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9965460Z test_jit.py::TestTracer::test_trace_slice_expr_complete_type <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9965897Z test_jit.py::TestTracer::test_trace_slice_full_dim <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9966337Z test_jit.py::TestTracer::test_trace_slice_setitem_dynamic_shape <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9966781Z test_jit.py::TestTracer::test_trace_slice_with_grad <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9967213Z test_jit.py::TestTracer::test_trace_tensor_factory <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9967626Z test_jit.py::TestTracer::test_trace_topk <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9968018Z test_jit.py::TestTracer::test_trace_tuple <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9968452Z test_jit.py::TestTracer::test_trace_variable_instantiation <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9968875Z test_jit.py::TestTracer::test_trace_warn <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9969303Z test_jit.py::TestTracer::test_trace_with_conditional_property <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9969781Z test_jit.py::TestTracer::test_trace_with_nested_tensor_list_output <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9970241Z test_jit.py::TestTracer::test_trace_with_number_list_output <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9970701Z test_jit.py::TestTracer::test_trace_with_tensor_list_output <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9971159Z test_jit.py::TestTracer::test_traced_module_cuda <- test/jit/test_tracer.py SKIPPED (calls .cuda()) [ 4%] 2023-03-31T05:50:35.9971618Z test_jit.py::TestTracer::test_tracing_backward_hook_error <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9972041Z test_jit.py::TestTracer::test_tracing_hooks <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9972472Z test_jit.py::TestTracer::test_tracing_multiple_methods <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9972890Z test_jit.py::TestTracer::test_typeas_trace_check <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9973311Z test_jit.py::TestTracer::test_wrapped_number <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9973797Z test_jit.py::TestMixTracingScripting::test_call_script_fn_from_traced_module <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9974448Z test_jit.py::TestMixTracingScripting::test_call_script_module_from_traced_module <- test/jit/test_tracer.py PASSED [ 4%] 2023-03-31T05:50:35.9975042Z test_jit.py::TestMixTracingScripting::test_call_traced_fn_from_script_fn <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9975559Z test_jit.py::TestMixTracingScripting::test_call_traced_mod_from_script_fn <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9976080Z test_jit.py::TestMixTracingScripting::test_call_tracing_fn_from_script_module <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9976597Z test_jit.py::TestMixTracingScripting::test_call_tracing_mod_from_script_module <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9977130Z test_jit.py::TestMixTracingScripting::test_jit_trace_callfunction_return_shapes <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9977659Z test_jit.py::TestMixTracingScripting::test_script_inline_trace_multiple_args <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9978166Z test_jit.py::TestMixTracingScripting::test_trace_dict_mix_script <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9978638Z test_jit.py::TestMixTracingScripting::test_trace_hierarchy <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9979107Z test_jit.py::TestMixTracingScripting::test_trace_linear <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9979617Z test_jit.py::TestMixTracingScripting::test_trace_mixed_by_script_with_dict_output <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9980108Z test_jit.py::TestMixTracingScripting::test_trace_of_script <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9980564Z test_jit.py::TestMixTracingScripting::test_trace_parameter <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9981076Z test_jit.py::TestMixTracingScripting::test_trace_returning_dict_with_tensor_tuples <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9981563Z test_jit.py::TestMixTracingScripting::test_trace_script <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9982077Z test_jit.py::TestMixTracingScripting::test_trace_script_returning_complex_dict <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9982565Z test_jit.py::TestMixTracingScripting::test_trace_with_size <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9983184Z test_jit.py::TestMixTracingScripting::test_traced_module_contains_scripted_interface_types <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9983734Z test_jit.py::TestMixTracingScripting::test_traced_module_implements_interface <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9984224Z test_jit.py::TestMixTracingScripting::test_tracing_indexing <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9984702Z test_jit.py::TestMixTracingScripting::test_tracing_slicing <- test/jit/test_tracer.py PASSED [ 5%] 2023-03-31T05:50:35.9985182Z test_jit.py::TestRecursiveScript::test_attributes <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9985664Z test_jit.py::TestRecursiveScript::test_class_compile <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9986147Z test_jit.py::TestRecursiveScript::test_constants_with_final <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9986619Z test_jit.py::TestRecursiveScript::test_dir <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9987085Z test_jit.py::TestRecursiveScript::test_error_stack <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9987580Z test_jit.py::TestRecursiveScript::test_error_stack_annotation <- test/jit/test_recursive_script.py PASSED [ 5%] 2023-03-31T05:50:35.9988063Z test_jit.py::TestRecursiveScript::test_error_stack_class <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9988550Z test_jit.py::TestRecursiveScript::test_error_stack_module <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9989058Z test_jit.py::TestRecursiveScript::test_failed_function_compilation <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9989652Z test_jit.py::TestRecursiveScript::test_function_attribute_in_submodule <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9990179Z test_jit.py::TestRecursiveScript::test_ignore_class <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9990665Z test_jit.py::TestRecursiveScript::test_inferred_nonetype <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9991147Z test_jit.py::TestRecursiveScript::test_init_error <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9991622Z test_jit.py::TestRecursiveScript::test_inner_traced_module <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9992117Z test_jit.py::TestRecursiveScript::test_iterable_modules <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9992595Z test_jit.py::TestRecursiveScript::test_method_call <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9993075Z test_jit.py::TestRecursiveScript::test_module_basic <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9993557Z test_jit.py::TestRecursiveScript::test_module_function_export <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9994039Z test_jit.py::TestRecursiveScript::test_module_name <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9994509Z test_jit.py::TestRecursiveScript::test_module_repr <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9994990Z test_jit.py::TestRecursiveScript::test_optional_module <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9995493Z test_jit.py::TestRecursiveScript::test_override_instance_method_ignore <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9996016Z test_jit.py::TestRecursiveScript::test_prepare_scriptable_basic <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9996535Z test_jit.py::TestRecursiveScript::test_prepare_scriptable_cycle <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9997071Z test_jit.py::TestRecursiveScript::test_prepare_scriptable_iterable_modules <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9997577Z test_jit.py::TestRecursiveScript::test_python_function_attribute <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9998081Z test_jit.py::TestRecursiveScript::test_repeated_error_stack <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9998573Z test_jit.py::TestRecursiveScript::test_script_after_eval <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9999051Z test_jit.py::TestRecursiveScript::test_script_basic <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:35.9999535Z test_jit.py::TestRecursiveScript::test_script_function_attribute <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:36.0000037Z test_jit.py::TestRecursiveScript::test_script_loaded_module <- test/jit/test_recursive_script.py PASSED [ 6%] 2023-03-31T05:50:36.0000518Z test_jit.py::TestTypeSharing::test_assign_python_attr <- test/jit/test_type_sharing.py PASSED [ 6%] 2023-03-31T05:50:36.0000953Z test_jit.py::TestTypeSharing::test_basic <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0001419Z test_jit.py::TestTypeSharing::test_builtin_function_different <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0001896Z test_jit.py::TestTypeSharing::test_builtin_function_same <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0002346Z test_jit.py::TestTypeSharing::test_constants <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0002785Z test_jit.py::TestTypeSharing::test_diff_attr_values <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0003411Z test_jit.py::TestTypeSharing::test_failed_attribute_compilation <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0004016Z test_jit.py::TestTypeSharing::test_ignored_fns <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0004451Z test_jit.py::TestTypeSharing::test_linear <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0004963Z test_jit.py::TestTypeSharing::test_loaded_modules_work <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0005465Z test_jit.py::TestTypeSharing::test_module_dict_same_type_different_name <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0005943Z test_jit.py::TestTypeSharing::test_mutate_attr_value <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0006393Z test_jit.py::TestTypeSharing::test_param_vs_attribute <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0006881Z test_jit.py::TestTypeSharing::test_python_function_attribute_different <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0007387Z test_jit.py::TestTypeSharing::test_python_function_attribute_same <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0007881Z test_jit.py::TestTypeSharing::test_same_but_different_classes <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0008371Z test_jit.py::TestTypeSharing::test_script_function_attribute_different <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0008873Z test_jit.py::TestTypeSharing::test_script_function_attribute_same <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0009382Z test_jit.py::TestTypeSharing::test_script_module_containing_traced_module <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0009854Z test_jit.py::TestTypeSharing::test_submodules <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0010320Z test_jit.py::TestTypeSharing::test_tracing_gives_different_types <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0010826Z test_jit.py::TestTypeSharing::test_type_not_shared_ignored_attributes <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0011333Z test_jit.py::TestTypeSharing::test_type_shared_ignored_attributes <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0011822Z test_jit.py::TestTypeSharing::test_type_sharing_define_in_init <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0012292Z test_jit.py::TestTypeSharing::test_type_sharing_disabled <- test/jit/test_type_sharing.py PASSED [ 7%] 2023-03-31T05:50:36.0012746Z test_jit.py::TestLogging::test_bump_numeric_counter <- test/jit/test_logging.py PASSED [ 7%] 2023-03-31T05:50:36.0013192Z test_jit.py::TestLogging::test_counter_aggregation <- test/jit/test_logging.py PASSED [ 8%] 2023-03-31T05:50:36.0013614Z test_jit.py::TestLogging::test_logging_levels_set <- test/jit/test_logging.py PASSED [ 8%] 2023-03-31T05:50:36.0014057Z test_jit.py::TestLogging::test_time_measurement_counter <- test/jit/test_logging.py PASSED [ 8%] 2023-03-31T05:50:36.0014521Z test_jit.py::TestLogging::test_time_measurement_counter_script <- test/jit/test_logging.py PASSED [ 8%] 2023-03-31T05:50:36.0014979Z test_jit.py::TestLogging::test_trace_numeric_counter <- test/jit/test_logging.py PASSED [ 8%] 2023-03-31T05:50:36.0015386Z test_jit.py::TestBackends::test_errors <- test/jit/test_backends.py PASSED [ 8%] 2023-03-31T05:50:36.0015797Z test_jit.py::TestBackends::test_execution <- test/jit/test_backends.py PASSED [ 8%] 2023-03-31T05:50:36.0016213Z test_jit.py::TestBackends::test_save_load <- test/jit/test_backends.py PASSED [ 8%] 2023-03-31T05:50:36.0016653Z test_jit.py::TestBackendsWithCompiler::test_errors <- test/jit/test_backends.py PASSED [ 8%] 2023-03-31T05:50:36.0017105Z test_jit.py::TestBackendsWithCompiler::test_execution <- test/jit/test_backends.py PASSED [ 8%] 2023-03-31T05:50:36.0017560Z test_jit.py::TestNnapiBackend::test_adaptive_avg_pool2d <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0017983Z test_jit.py::TestNnapiBackend::test_avg_pool2d <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0018374Z test_jit.py::TestNnapiBackend::test_cat <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0018929Z test_jit.py::TestNnapiBackend::test_compile_spec_santiy <- test/jit/test_backend_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0019411Z test_jit.py::TestNnapiBackend::test_conv2d <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0019834Z test_jit.py::TestNnapiBackend::test_conv2d_transpose <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0020242Z test_jit.py::TestNnapiBackend::test_dequantize <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0020651Z test_jit.py::TestNnapiBackend::test_detach <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0021048Z test_jit.py::TestNnapiBackend::test_flatten <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0021441Z test_jit.py::TestNnapiBackend::test_hardtanh <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0021843Z test_jit.py::TestNnapiBackend::test_linear <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0022256Z test_jit.py::TestNnapiBackend::test_log_softmax <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0022675Z test_jit.py::TestNnapiBackend::test_max_pool2d <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0023151Z test_jit.py::TestNnapiBackend::test_mean <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0023581Z test_jit.py::TestNnapiBackend::test_multi_output <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0024004Z test_jit.py::TestNnapiBackend::test_pointwise_binary <- test/test_nnapi.py PASSED [ 8%] 2023-03-31T05:50:36.0024435Z test_jit.py::TestNnapiBackend::test_pointwise_binary_const <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0024865Z test_jit.py::TestNnapiBackend::test_pointwise_unary <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0025276Z test_jit.py::TestNnapiBackend::test_prelu <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0025675Z test_jit.py::TestNnapiBackend::test_qadd <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0026060Z test_jit.py::TestNnapiBackend::test_qlinear <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0026469Z test_jit.py::TestNnapiBackend::test_quantize <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0026876Z test_jit.py::TestNnapiBackend::test_reshape <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0027282Z test_jit.py::TestNnapiBackend::test_seblock_mul <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0027683Z test_jit.py::TestNnapiBackend::test_slice <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0028079Z test_jit.py::TestNnapiBackend::test_softmax <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0028510Z test_jit.py::TestNnapiBackend::test_tensor_input <- test/jit/test_backend_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0028914Z test_jit.py::TestNnapiBackend::test_to <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0029319Z test_jit.py::TestNnapiBackend::test_unsqueeze <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0029746Z test_jit.py::TestNnapiBackend::test_upsample_nearest2d <- test/test_nnapi.py PASSED [ 9%] 2023-03-31T05:50:36.0030180Z test_jit.py::TestList::test_comprehension_iterable <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0030646Z test_jit.py::TestList::test_comprehension_out_type_not_in_type <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0031096Z test_jit.py::TestList::test_comprehensions_basic <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0031542Z test_jit.py::TestList::test_comprehensions_basic_float <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0031976Z test_jit.py::TestList::test_comprehensions_two_comps <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0032413Z test_jit.py::TestList::test_copy_list_immutable <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0032837Z test_jit.py::TestList::test_copy_list_mutable <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0033236Z test_jit.py::TestList::test_del <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0033721Z test_jit.py::TestList::test_dict_keyword_is_correctly_typed <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0034191Z test_jit.py::TestList::test_dict_keyword_with_dict_comprehension <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0034744Z test_jit.py::TestList::test_dict_keyword_with_dict_comprehension_and_kwargs <- test/jit/test_list_dict.py PASSED [ 9%] 2023-03-31T05:50:36.0035239Z test_jit.py::TestList::test_dict_keyword_with_empty_dict_comprehension <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0035711Z test_jit.py::TestList::test_dict_keyword_with_empty_iterable <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0036200Z test_jit.py::TestList::test_dict_keyword_with_internal_aggregate_function <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0036666Z test_jit.py::TestList::test_dict_keyword_with_iterable <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0037093Z test_jit.py::TestList::test_dict_keyword_with_kwargs <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0037578Z test_jit.py::TestList::test_dict_keyword_with_kwargs_using_container_values <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0038049Z test_jit.py::TestList::test_dict_keyword_with_mapping <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0038518Z test_jit.py::TestList::test_dict_keyword_with_mapping_and_kwargs <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0038992Z test_jit.py::TestList::test_dict_keyword_with_mismatched_annotations <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0039459Z test_jit.py::TestList::test_dict_keyword_with_nested_call <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0039943Z test_jit.py::TestList::test_dict_keyword_with_previously_declared_variable <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0040468Z test_jit.py::TestList::test_dict_keyword_with_previously_declared_variable_and_kwargs <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0040927Z test_jit.py::TestList::test_extend_list_immutable <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0041356Z test_jit.py::TestList::test_extend_list_mutable <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0041768Z test_jit.py::TestList::test_in_check <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0042169Z test_jit.py::TestList::test_list_bool_conversion <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0042580Z test_jit.py::TestList::test_list_count <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0043002Z test_jit.py::TestList::test_list_count_not_existing <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0043543Z test_jit.py::TestList::test_list_gather <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0043934Z test_jit.py::TestList::test_list_index <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0044356Z test_jit.py::TestList::test_list_index_not_existing <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0044775Z test_jit.py::TestList::test_list_keyword <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0045165Z test_jit.py::TestList::test_list_len <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0045563Z test_jit.py::TestList::test_list_literal <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0045960Z test_jit.py::TestList::test_list_none <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0046351Z test_jit.py::TestList::test_list_ops <- test/jit/test_list_dict.py PASSED [ 10%] 2023-03-31T05:50:36.0046734Z test_jit.py::TestList::test_list_slice <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0047133Z test_jit.py::TestList::test_list_sort <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0047550Z test_jit.py::TestList::test_list_unification_hint <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0047953Z test_jit.py::TestList::test_list_variance <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0048454Z test_jit.py::TestList::test_min_bool_list <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0048939Z test_jit.py::TestList::test_min_max_list <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0049358Z test_jit.py::TestList::test_min_max_single_list <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0049775Z test_jit.py::TestList::test_mutable_list_append <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0050205Z test_jit.py::TestList::test_mutable_list_append_2 <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0050639Z test_jit.py::TestList::test_mutable_list_append_if <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0051078Z test_jit.py::TestList::test_mutable_list_append_if_else <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0051511Z test_jit.py::TestList::test_mutable_list_append_loop <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0051957Z test_jit.py::TestList::test_mutable_list_append_loop_if <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0052388Z test_jit.py::TestList::test_mutable_list_clear <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0052810Z test_jit.py::TestList::test_mutable_list_clear_empty <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0053260Z test_jit.py::TestList::test_mutable_list_function_inline <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0053693Z test_jit.py::TestList::test_mutable_list_insert <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0054147Z test_jit.py::TestList::test_mutable_list_insert_neg_out_of_bounds <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0054592Z test_jit.py::TestList::test_mutable_list_insert_negative <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0055045Z test_jit.py::TestList::test_mutable_list_insert_out_of_bounds <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0055495Z test_jit.py::TestList::test_mutable_list_nested_loop <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0055921Z test_jit.py::TestList::test_mutable_list_pop <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0056329Z test_jit.py::TestList::test_mutable_list_pop2 <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0056749Z test_jit.py::TestList::test_mutable_list_pop_at <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0057176Z test_jit.py::TestList::test_mutable_list_pop_at2 <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0057604Z test_jit.py::TestList::test_mutable_list_pop_at_negative <- test/jit/test_list_dict.py PASSED [ 11%] 2023-03-31T05:50:36.0058055Z test_jit.py::TestList::test_mutable_list_pop_at_negative2 <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0058492Z test_jit.py::TestList::test_mutable_list_pop_empty <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0058923Z test_jit.py::TestList::test_mutable_list_pop_slice <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0059333Z test_jit.py::TestList::test_mutable_list_remove <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0059762Z test_jit.py::TestList::test_mutable_list_remove2 <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0060210Z test_jit.py::TestList::test_mutable_list_remove_not_existing <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0060646Z test_jit.py::TestList::test_mutable_list_remove_tensor <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0082365Z test_jit.py::TestList::test_mutable_list_reverse <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0083524Z test_jit.py::TestList::test_mutable_list_reverse_empty <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0084198Z test_jit.py::TestList::test_mutable_tensor_list_reverse <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0084791Z test_jit.py::TestList::test_no_element_type_annotation <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0085683Z test_jit.py::TestList::test_slice_index <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0086395Z test_jit.py::TestList::test_tensor_list_count <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0087068Z test_jit.py::TestList::test_tensor_list_count_not_existing <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0087680Z test_jit.py::TestList::test_tensor_list_index <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0088311Z test_jit.py::TestList::test_tensor_list_index_not_existing <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0088893Z test_jit.py::TestList::test_to_list <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0089436Z test_jit.py::TestList::test_to_list_gpu <- test/jit/test_list_dict.py SKIPPED (CUDA is not available) [ 12%] 2023-03-31T05:50:36.0089874Z test_jit.py::TestDict::test_aug_assign <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0090268Z test_jit.py::TestDict::test_basic <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0090662Z test_jit.py::TestDict::test_clear <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0091059Z test_jit.py::TestDict::test_copy <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0091448Z test_jit.py::TestDict::test_del <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0091853Z test_jit.py::TestDict::test_dict_bool_conversion <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0092290Z test_jit.py::TestDict::test_dict_preserves_order <- test/jit/test_list_dict.py PASSED [ 12%] 2023-03-31T05:50:36.0092711Z test_jit.py::TestDict::test_dict_to_python <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0093111Z test_jit.py::TestDict::test_dict_variance <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0093509Z test_jit.py::TestDict::test_get <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0093913Z test_jit.py::TestDict::test_get_boolkey <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0094316Z test_jit.py::TestDict::test_items <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0094699Z test_jit.py::TestDict::test_key_type <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0095090Z test_jit.py::TestDict::test_keys <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0095475Z test_jit.py::TestDict::test_len <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0095879Z test_jit.py::TestDict::test_loop <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0096368Z test_jit.py::TestDict::test_membership <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0096774Z test_jit.py::TestDict::test_mutability <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0097205Z test_jit.py::TestDict::test_optional_dict_construct <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0097620Z test_jit.py::TestDict::test_ordered_dict <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0098017Z test_jit.py::TestDict::test_pop <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0098409Z test_jit.py::TestDict::test_popitem <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0098800Z test_jit.py::TestDict::test_setdefault <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0099261Z test_jit.py::TestDict::test_type_annotation_missing_contained_type <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0099691Z test_jit.py::TestDict::test_update <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0100106Z test_jit.py::TestDict::test_update_existing_key <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0100498Z test_jit.py::TestDict::test_values <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0100887Z test_jit.py::TestDict::test_view <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0101428Z test_jit.py::TestNamedTuple::test_namedtuple <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0101944Z test_jit.py::TestNamedTuple::test_namedtuple_as_attr <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0102407Z test_jit.py::TestNamedTuple::test_namedtuple_constant <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0102881Z test_jit.py::TestNamedTuple::test_namedtuple_input_forwardref <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0103466Z test_jit.py::TestNamedTuple::test_namedtuple_inside_forwardref <- test/jit/test_list_dict.py PASSED [ 13%] 2023-03-31T05:50:36.0103946Z test_jit.py::TestNamedTuple::test_namedtuple_kwarg_construct <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0104408Z test_jit.py::TestNamedTuple::test_namedtuple_lower <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0104868Z test_jit.py::TestNamedTuple::test_namedtuple_resolution <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0105354Z test_jit.py::TestNamedTuple::test_namedtuple_resolution_forwardref <- test/jit/test_list_dict.py XFAIL [ 14%] 2023-03-31T05:50:36.0105897Z test_jit.py::TestNamedTuple::test_namedtuple_serialization <- test/jit/test_list_dict.py SKIPPED (broken while these tests were not in CI) [ 14%] 2023-03-31T05:50:36.0106412Z test_jit.py::TestNamedTuple::test_namedtuple_slice_unpack <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0106884Z test_jit.py::TestNamedTuple::test_namedtuple_type_annotation <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0107357Z test_jit.py::TestNamedTuple::test_namedtuple_wrong_types <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0107801Z test_jit.py::TestNamedTuple::test_return_named_tuple <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0108225Z test_jit.py::TestScriptDict::test_bool <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0108641Z test_jit.py::TestScriptDict::test_contains <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0109050Z test_jit.py::TestScriptDict::test_delitem <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0109471Z test_jit.py::TestScriptDict::test_getitem <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0109884Z test_jit.py::TestScriptDict::test_items <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0110291Z test_jit.py::TestScriptDict::test_iter <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0110684Z test_jit.py::TestScriptDict::test_len <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0111233Z test_jit.py::TestScriptDict::test_nested <- test/jit/test_list_dict.py SKIPPED (Cannot pass until all dicts returned from TorchScript are ScriptDicts) [ 14%] 2023-03-31T05:50:36.0111749Z test_jit.py::TestScriptDict::test_reference_semantics <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0112174Z test_jit.py::TestScriptDict::test_repr <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0112580Z test_jit.py::TestScriptDict::test_setitem <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0113002Z test_jit.py::TestScriptList::test_append <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0113414Z test_jit.py::TestScriptList::test_bool <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0113812Z test_jit.py::TestScriptList::test_clear <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0114229Z test_jit.py::TestScriptList::test_contains <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0114641Z test_jit.py::TestScriptList::test_count <- test/jit/test_list_dict.py PASSED [ 14%] 2023-03-31T05:50:36.0115055Z test_jit.py::TestScriptList::test_delitem <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0115457Z test_jit.py::TestScriptList::test_extend <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0115872Z test_jit.py::TestScriptList::test_getitem <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0116359Z test_jit.py::TestScriptList::test_insert <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0116837Z test_jit.py::TestScriptList::test_iter <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0117250Z test_jit.py::TestScriptList::test_len <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0117794Z test_jit.py::TestScriptList::test_nested <- test/jit/test_list_dict.py SKIPPED (Cannot pass until all list returned from TorchScript are ScriptLists) [ 15%] 2023-03-31T05:50:36.0118283Z test_jit.py::TestScriptList::test_pop <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0118709Z test_jit.py::TestScriptList::test_reference_semantics <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0119144Z test_jit.py::TestScriptList::test_remove <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0119554Z test_jit.py::TestScriptList::test_repr <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0119961Z test_jit.py::TestScriptList::test_setitem <- test/jit/test_list_dict.py PASSED [ 15%] 2023-03-31T05:50:36.0120391Z test_jit.py::TestAsync::test_async_future_type_python <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0120834Z test_jit.py::TestAsync::test_async_grad_guard_no_grad <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0121277Z test_jit.py::TestAsync::test_async_grad_guard_with_grad <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0121679Z test_jit.py::TestAsync::test_async_kwargs <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0122083Z test_jit.py::TestAsync::test_async_parsing <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0122487Z test_jit.py::TestAsync::test_async_python <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0122874Z test_jit.py::TestAsync::test_async_script <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0123501Z test_jit.py::TestAsync::test_async_script_capture <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0123931Z test_jit.py::TestAsync::test_async_script_error <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0124369Z test_jit.py::TestAsync::test_async_script_multi_forks <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0124791Z test_jit.py::TestAsync::test_async_script_multi_waits <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0125217Z test_jit.py::TestAsync::test_async_script_nested <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0125651Z test_jit.py::TestAsync::test_async_script_no_script_mod <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0126077Z test_jit.py::TestAsync::test_async_script_trace <- test/jit/test_async.py PASSED [ 15%] 2023-03-31T05:50:36.0126479Z test_jit.py::TestAsync::test_future_subtyping <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0126911Z test_jit.py::TestAsync::test_no_future_subtype_message <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0127338Z test_jit.py::TestAsync::test_trace_fork_wait <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0127746Z test_jit.py::TestAsync::test_trace_fork_wait_inline <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0128178Z test_jit.py::TestAsync::test_trace_fork_wait_leaking <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0128627Z test_jit.py::TestAsync::test_trace_fork_wait_list_modulecalls <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0129117Z test_jit.py::TestAsync::test_trace_modulecalls_with_different_output_types <- test/jit/test_async.py PASSED [ 16%] 2023-03-31T05:50:36.0129548Z test_jit.py::TestAwait::test_await_class_arg <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0129965Z test_jit.py::TestAwait::test_await_class_return <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0130378Z test_jit.py::TestAwait::test_await_eager_lazy <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0130769Z test_jit.py::TestAwait::test_await_func_arg <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0131340Z test_jit.py::TestAwait::test_await_getattr_implicit_convertion <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0131849Z test_jit.py::TestAwait::test_await_isinstance <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0132272Z test_jit.py::TestAwait::test_await_multiout_save <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0132665Z test_jit.py::TestAwait::test_await_nested <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0133088Z test_jit.py::TestAwait::test_await_out_of_interpreter <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0133503Z test_jit.py::TestAwait::test_await_python <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0133915Z test_jit.py::TestAwait::test_await_type_python <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0134317Z test_jit.py::TestAwait::test_awaitable_to_await <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0134752Z test_jit.py::TestAwait::test_eager_await_non_scriptable <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0135172Z test_jit.py::TestAwait::test_jit_trace <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0135556Z test_jit.py::TestAwait::test_nowait <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0135951Z test_jit.py::TestAwait::test_nowait_class <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0136343Z test_jit.py::TestAwait::test_script <- test/jit/test_await.py PASSED [ 16%] 2023-03-31T05:50:36.0136843Z test_jit.py::TestDataParallel::test_python_submodule_script <- test/jit/test_data_parallel.py SKIPPED (multi-GPU not supported) [ 16%] 2023-03-31T05:50:36.0137393Z test_jit.py::TestDataParallel::test_shared_module <- test/jit/test_data_parallel.py SKIPPED (multi-GPU not supported) [ 16%] 2023-03-31T05:50:36.0137941Z test_jit.py::TestDataParallel::test_tensor_sharing <- test/jit/test_data_parallel.py SKIPPED (multi-GPU not supported) [ 17%] 2023-03-31T05:50:36.0138511Z test_jit.py::TestDataParallel::test_tensor_sharing_with_forward <- test/jit/test_data_parallel.py SKIPPED (multi-GPU not supported) [ 17%] 2023-03-31T05:50:36.0139067Z test_jit.py::TestDataParallel::test_traced_module <- test/jit/test_data_parallel.py SKIPPED (multi-GPU not supported) [ 17%] 2023-03-31T05:50:36.0139506Z test_jit.py::TestModels::test_alexnet <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0139918Z test_jit.py::TestModels::test_dcgan_models <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0140359Z test_jit.py::TestModels::test_dcgan_models_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0140763Z test_jit.py::TestModels::test_mnist <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0141180Z test_jit.py::TestModels::test_mnist_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0141658Z test_jit.py::TestModels::test_mnist_training_leaks_no_memory_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0142212Z test_jit.py::TestModels::test_neural_style <- test/jit/test_models.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 17%] 2023-03-31T05:50:36.0142708Z test_jit.py::TestModels::test_neural_style_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0143238Z test_jit.py::TestModels::test_reinforcement_learning <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0143716Z test_jit.py::TestModels::test_reinforcement_learning_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0144292Z test_jit.py::TestModels::test_script_module_script_resnet <- test/jit/test_models.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 17%] 2023-03-31T05:50:36.0144897Z test_jit.py::TestModels::test_script_module_trace_resnet18 <- test/jit/test_models.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 17%] 2023-03-31T05:50:36.0145477Z test_jit.py::TestModels::test_snli <- test/jit/test_models.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 17%] 2023-03-31T05:50:36.0146063Z test_jit.py::TestModels::test_snli_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0146535Z test_jit.py::TestModels::test_snli_quantized <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0147053Z test_jit.py::TestModels::test_super_resolution <- test/jit/test_models.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 17%] 2023-03-31T05:50:36.0147568Z test_jit.py::TestModels::test_super_resolution_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0148028Z test_jit.py::TestModels::test_time_sequence_prediction <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0148443Z test_jit.py::TestModels::test_vae <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0148845Z test_jit.py::TestModels::test_vae_cuda <- test/jit/test_models.py SKIPPED (no CUDA) [ 17%] 2023-03-31T05:50:36.0149272Z test_jit.py::TestModels::test_vae_quantized <- test/jit/test_models.py PASSED [ 17%] 2023-03-31T05:50:36.0149729Z test_jit.py::TestModules::test_script_module_with_constants_list <- test/jit/test_modules.py PASSED [ 17%] 2023-03-31T05:50:36.0150207Z test_jit.py::TestAutodiffJit::test_autodiff_requires_grad_nograd <- test/jit/test_autodiff.py PASSED [ 18%] 2023-03-31T05:50:36.0150681Z test_jit.py::TestAutodiffJit::test_requires_grad_outputs <- test/jit/test_autodiff.py PASSED [ 18%] 2023-03-31T05:50:36.0151177Z test_jit.py::TestAutodiffJit::test_requires_grad_outputs_profiled_twice <- test/jit/test_autodiff.py PASSED [ 18%] 2023-03-31T05:50:36.0151682Z test_jit.py::TestAutodiffJit::test_requires_grad_outputs_side_effects <- test/jit/test_autodiff.py PASSED [ 18%] 2023-03-31T05:50:36.0152145Z test_jit.py::TestAutodiffJit::test_undefined_tensor_lists <- test/jit/test_autodiff.py PASSED [ 18%] 2023-03-31T05:50:36.0152667Z test_jit.py::TestAutodiffSubgraphSlicing::test_aliased_outputs <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0153217Z test_jit.py::TestAutodiffSubgraphSlicing::test_bias_as_arg <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0153770Z test_jit.py::TestAutodiffSubgraphSlicing::test_bias_as_module_attr <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0154327Z test_jit.py::TestAutodiffSubgraphSlicing::test_chunk_constant_script_ad <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0154890Z test_jit.py::TestAutodiffSubgraphSlicing::test_constructed_bias <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0155464Z test_jit.py::TestAutodiffSubgraphSlicing::test_diff_graph_inline_threshold <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0156206Z test_jit.py::TestAutodiffSubgraphSlicing::test_differentiable_graph_ops_requires_grad <- test/jit/test_autodiff_subgraph_slicing.py SKIPPED (disable until we property handle tensor lists with undefined gradients) [ 18%] 2023-03-31T05:50:36.0156869Z test_jit.py::TestAutodiffSubgraphSlicing::test_does_not_create_cycles <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0157446Z test_jit.py::TestAutodiffSubgraphSlicing::test_does_not_merge_unrelated <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0158041Z test_jit.py::TestAutodiffSubgraphSlicing::test_has_profiled_info_aliasing_outputs <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0158632Z test_jit.py::TestAutodiffSubgraphSlicing::test_merge_respects_aliasing <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0159171Z test_jit.py::TestAutodiffSubgraphSlicing::test_merges_dense <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0159706Z test_jit.py::TestAutodiffSubgraphSlicing::test_merges_down <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0160243Z test_jit.py::TestAutodiffSubgraphSlicing::test_merges_up <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0160913Z test_jit.py::TestAutodiffSubgraphSlicing::test_merges_without_cycles <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0161539Z test_jit.py::TestAutodiffSubgraphSlicing::test_prune_grad <- test/jit/test_autodiff_subgraph_slicing.py SKIPPED (Simple Executor doesn't support gradients) [ 18%] 2023-03-31T05:50:36.0162160Z test_jit.py::TestAutodiffSubgraphSlicing::test_requires_grad_for_tensor_list <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0162743Z test_jit.py::TestAutodiffSubgraphSlicing::test_respects_lexical_scoping <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0163507Z test_jit.py::TestAutodiffSubgraphSlicing::test_simple_merge <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0164037Z test_jit.py::TestAutodiffSubgraphSlicing::test_simple_no_merge <- test/jit/test_autodiff_subgraph_slicing.py PASSED [ 18%] 2023-03-31T05:50:36.0164584Z test_jit.py::TestCustomOperators::test_calling_scripted_custom_op <- test/jit/test_custom_operators.py PASSED [ 18%] 2023-03-31T05:50:36.0165112Z test_jit.py::TestCustomOperators::test_calling_traced_custom_op <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0165634Z test_jit.py::TestCustomOperators::test_default_arguments_are_used <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0166125Z test_jit.py::TestCustomOperators::test_dynamic_op_registry <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0166617Z test_jit.py::TestCustomOperators::test_generic_list <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0167122Z test_jit.py::TestCustomOperators::test_passing_and_returning_lists <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0167669Z test_jit.py::TestCustomOperators::test_passing_one_positional_but_not_the_second <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0168183Z test_jit.py::TestCustomOperators::test_passing_too_few_args <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0168690Z test_jit.py::TestCustomOperators::test_passing_too_many_args <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0169202Z test_jit.py::TestCustomOperators::test_passing_unknown_kwargs <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0169733Z test_jit.py::TestCustomOperators::test_script_graph_contains_custom_op <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0170399Z test_jit.py::TestCustomOperators::test_script_graph_for_custom_ops_matches_traced_graph <- test/jit/test_custom_operators.py SKIPPED (Need to figure out default dtype differences between fbcode and oss) [ 19%] 2023-03-31T05:50:36.0171008Z test_jit.py::TestCustomOperators::test_simply_calling_an_operator <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0171511Z test_jit.py::TestCustomOperators::test_where_no_scalar <- test/jit/test_custom_operators.py PASSED [ 19%] 2023-03-31T05:50:36.0172009Z test_jit.py::TestGraphRewritePasses::test_fuse_linear <- test/jit/test_graph_rewrite_passes.py PASSED [ 19%] 2023-03-31T05:50:36.0172465Z test_jit.py::TestClassType::test_cast_overloads <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0172932Z test_jit.py::TestClassType::test_class_attribute_wrong_type <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0173381Z test_jit.py::TestClassType::test_class_constant <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0173827Z test_jit.py::TestClassType::test_class_constructs_itself <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0174279Z test_jit.py::TestClassType::test_class_inheritance <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0174747Z test_jit.py::TestClassType::test_class_inheritance_implicit <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0175311Z test_jit.py::TestClassType::test_class_sorting <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0175751Z test_jit.py::TestClassType::test_class_specialization <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0176273Z test_jit.py::TestClassType::test_class_type_as_param <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0176719Z test_jit.py::TestClassType::test_classmethod <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0177171Z test_jit.py::TestClassType::test_conditional_set_attr <- test/jit/test_class_type.py PASSED [ 19%] 2023-03-31T05:50:36.0177600Z test_jit.py::TestClassType::test_custom_delete <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0178033Z test_jit.py::TestClassType::test_default_args <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0178464Z test_jit.py::TestClassType::test_get_attr <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0178901Z test_jit.py::TestClassType::test_get_attr_not_initialized <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0179351Z test_jit.py::TestClassType::test_get_with_method <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0179796Z test_jit.py::TestClassType::test_imported_classes <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0180214Z test_jit.py::TestClassType::test_in <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0180636Z test_jit.py::TestClassType::test_init_compiled_first <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0181073Z test_jit.py::TestClassType::test_interface <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0181523Z test_jit.py::TestClassType::test_optional_type_promotion <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0181985Z test_jit.py::TestClassType::test_out_of_order_methods <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0182415Z test_jit.py::TestClassType::test_overloaded_fn <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0182852Z test_jit.py::TestClassType::test_properties <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0183409Z test_jit.py::TestClassType::test_py_class_to_ivalue_missing_attribute <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0183857Z test_jit.py::TestClassType::test_python_interop <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0184306Z test_jit.py::TestClassType::test_recursive_class <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0184793Z test_jit.py::TestClassType::test_recursive_script_builtin_type_resolution <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0185323Z test_jit.py::TestClassType::test_recursive_script_module_builtin_type_resolution <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0185790Z test_jit.py::TestClassType::test_recursive_scripting <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0186265Z test_jit.py::TestClassType::test_recursive_scripting_failed <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0186724Z test_jit.py::TestClassType::test_reference_semantics <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0187186Z test_jit.py::TestClassType::test_save_load_with_classes <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0187649Z test_jit.py::TestClassType::test_save_load_with_classes_nested <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0188130Z test_jit.py::TestClassType::test_save_load_with_classes_returned <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0188601Z test_jit.py::TestClassType::test_schema_human_readable <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0189048Z test_jit.py::TestClassType::test_self_referential_method <- test/jit/test_class_type.py PASSED [ 20%] 2023-03-31T05:50:36.0189499Z test_jit.py::TestClassType::test_set_attr_in_method <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0190034Z test_jit.py::TestClassType::test_set_attr_non_initialized <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0190493Z test_jit.py::TestClassType::test_set_attr_type_mismatch <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0190973Z test_jit.py::TestClassType::test_staticmethod <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0191416Z test_jit.py::TestClassType::test_type_annotation <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0191857Z test_jit.py::TestClassType::test_type_annotations <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0192328Z test_jit.py::TestClassType::test_unresolved_class_attributes <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0192765Z test_jit.py::TestClassType::test_unused_method <- test/jit/test_class_type.py PASSED [ 21%] 2023-03-31T05:50:36.0193176Z test_jit.py::TestBuiltins::test_del <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0193610Z test_jit.py::TestBuiltins::test_del_multiple_operands <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0194027Z test_jit.py::TestBuiltins::test_has_attr <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0194465Z test_jit.py::TestBuiltins::test_has_attr_invalid_args <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0194921Z test_jit.py::TestTensorBuiltins::test_method_on_number <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0195397Z test_jit.py::TestTensorBuiltins::test_scalar_to_num_conversions <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0195837Z test_jit.py::TestTensorBuiltins::test_tensor_item <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0196292Z test_jit.py::TestTensorBuiltins::test_tensor_properties <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0196765Z test_jit.py::TestTensorBuiltins::test_tensor_subscript_assign <- test/jit/test_builtins.py PASSED [ 21%] 2023-03-31T05:50:36.0197283Z test_jit.py::TestTensorBuiltins::test_tensor_subscript_assign_device <- test/jit/test_builtins.py SKIPPED (requires CUDA) [ 21%] 2023-03-31T05:50:36.0197860Z test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_inp_out <- test/jit/test_ignore_context_manager.py PASSED [ 21%] 2023-03-31T05:50:36.0198458Z test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_just_inp <- test/jit/test_ignore_context_manager.py PASSED [ 21%] 2023-03-31T05:50:36.0199052Z test_jit.py::TestIgnoreContextManager::test_with_ignore_context_manager_with_just_out <- test/jit/test_ignore_context_manager.py PASSED [ 21%] 2023-03-31T05:50:36.0199623Z test_jit.py::TestSymbolicShapeAnalysis::test_adaptive_avg_pool2d <- test/jit/test_symbolic_shape_analysis.py PASSED [ 21%] 2023-03-31T05:50:36.0200145Z test_jit.py::TestSymbolicShapeAnalysis::test_arange_shape <- test/jit/test_symbolic_shape_analysis.py PASSED [ 21%] 2023-03-31T05:50:36.0200694Z test_jit.py::TestSymbolicShapeAnalysis::test_binary_shape_fns_inplace <- test/jit/test_symbolic_shape_analysis.py PASSED [ 21%] 2023-03-31T05:50:36.0201253Z test_jit.py::TestSymbolicShapeAnalysis::test_binary_shape_functions <- test/jit/test_symbolic_shape_analysis.py PASSED [ 21%] 2023-03-31T05:50:36.0201807Z test_jit.py::TestSymbolicShapeAnalysis::test_convolution_backward <- test/jit/test_symbolic_shape_analysis.py PASSED [ 21%] 2023-03-31T05:50:36.0202328Z test_jit.py::TestSymbolicShapeAnalysis::test_if_propagation <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0202878Z test_jit.py::TestSymbolicShapeAnalysis::test_partial_eval_graph_conv <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0203631Z test_jit.py::TestSymbolicShapeAnalysis::test_partial_eval_stitching <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0204216Z test_jit.py::TestSymbolicShapeAnalysis::test_refinement_through_graph_stitching <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0204792Z test_jit.py::TestSymbolicShapeAnalysis::test_register_function_error_checking <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0205555Z test_jit.py::TestSymbolicShapeAnalysis::test_returning_input_symbolic_shapes <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0206112Z test_jit.py::TestSymbolicShapeAnalysis::test_shape_analysis <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0206640Z test_jit.py::TestSymbolicShapeAnalysis::test_shape_concat <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0207159Z test_jit.py::TestSymbolicShapeAnalysis::test_shape_embedding_bag <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0207780Z test_jit.py::TestSymbolicShapeAnalysis::test_shape_function_includes <- test/jit/test_symbolic_shape_analysis.py SKIPPED (shape functions not loaded in python) [ 22%] 2023-03-31T05:50:36.0208371Z test_jit.py::TestSymbolicShapeAnalysis::test_shared_shape_graph <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0208906Z test_jit.py::TestSymbolicShapeAnalysis::test_size_and_sizes <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0209434Z test_jit.py::TestSymbolicShapeAnalysis::test_stitching_concat <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0209981Z test_jit.py::TestSymbolicShapeAnalysis::test_stitching_multi_output <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0210516Z test_jit.py::TestSymbolicShapeAnalysis::test_sym_ir_parsing <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0211059Z test_jit.py::TestSymbolicShapeAnalysis::test_unary_shape_fns_inplace <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0211595Z test_jit.py::TestSymbolicShapeAnalysis::test_unary_shape_functions <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0212116Z test_jit.py::TestSymbolicShapeAnalysis::test_write <- test/jit/test_symbolic_shape_analysis.py PASSED [ 22%] 2023-03-31T05:50:36.0212633Z test_jit.py::TestOpDecompositions::test_op_decomposition <- test/jit/test_op_decompositions.py PASSED [ 22%] 2023-03-31T05:50:36.0213150Z test_jit.py::TestOpDecompositions::test_registered_decomposition <- test/jit/test_op_decompositions.py PASSED [ 22%] 2023-03-31T05:50:36.0213683Z test_jit.py::TestUnsupportedOps::test_factory_ops_requires_grad_fail <- test/jit/test_unsupported_ops.py PASSED [ 22%] 2023-03-31T05:50:36.0214166Z test_jit.py::TestUnsupportedOps::test_init_ops <- test/jit/test_unsupported_ops.py PASSED [ 22%] 2023-03-31T05:50:36.0214660Z test_jit.py::TestFreezing::test_freeze_interface_swapping_two_methods <- test/jit/test_freezing.py PASSED [ 22%] 2023-03-31T05:50:36.0215128Z test_jit.py::TestFreezing::test_freeze_interface_within_object <- test/jit/test_freezing.py XFAIL [ 22%] 2023-03-31T05:50:36.0215564Z test_jit.py::TestFreezing::test_freeze_module <- test/jit/test_freezing.py PASSED [ 22%] 2023-03-31T05:50:36.0216027Z test_jit.py::TestFreezing::test_freeze_module_detach_gradient <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0216504Z test_jit.py::TestFreezing::test_freeze_module_in_training_mode <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0216951Z test_jit.py::TestFreezing::test_freeze_module_inlining <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0217404Z test_jit.py::TestFreezing::test_freeze_module_no_forward <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0217861Z test_jit.py::TestFreezing::test_freeze_module_return_self <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0218313Z test_jit.py::TestFreezing::test_freeze_module_return_sub_module <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0218790Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0219271Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr2 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0219811Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_attr3 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0220365Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0220858Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr2 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0221347Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr3 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0221831Z test_jit.py::TestFreezing::test_freeze_module_with_aliased_tensor_attr4 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0222294Z test_jit.py::TestFreezing::test_freeze_module_with_call_method <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0222757Z test_jit.py::TestFreezing::test_freeze_module_with_fork <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0223287Z test_jit.py::TestFreezing::test_freeze_module_with_fork2 <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0223792Z test_jit.py::TestFreezing::test_freeze_module_with_fork_calling_module_method <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0224278Z test_jit.py::TestFreezing::test_freeze_module_with_helperfunction <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0224765Z test_jit.py::TestFreezing::test_freeze_module_with_inplace_mutable <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0225231Z test_jit.py::TestFreezing::test_freeze_module_with_list <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0225696Z test_jit.py::TestFreezing::test_freeze_module_with_mutable_dict <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0226156Z test_jit.py::TestFreezing::test_freeze_module_with_mutable_list <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0226640Z test_jit.py::TestFreezing::test_freeze_module_with_mutable_tensor <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0227124Z test_jit.py::TestFreezing::test_freeze_module_with_nested_fork <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0227591Z test_jit.py::TestFreezing::test_freeze_module_with_nestedaliasing <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0228093Z test_jit.py::TestFreezing::test_freeze_module_with_nestedaliasingscalar <- test/jit/test_freezing.py PASSED [ 23%] 2023-03-31T05:50:36.0228622Z test_jit.py::TestFreezing::test_freeze_module_with_non_static_module_container_index <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0229130Z test_jit.py::TestFreezing::test_freeze_module_with_overlapping_attrs <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0229613Z test_jit.py::TestFreezing::test_freeze_module_with_preserve_sub_module <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0230133Z test_jit.py::TestFreezing::test_freeze_module_with_preserve_sub_module_and_mutation <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0230645Z test_jit.py::TestFreezing::test_freeze_module_with_sharedclasstype <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0231127Z test_jit.py::TestFreezing::test_freeze_module_with_submodule <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0231571Z test_jit.py::TestFreezing::test_freeze_module_with_tensor <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0232025Z test_jit.py::TestFreezing::test_freeze_module_with_tuple <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0232513Z test_jit.py::TestFreezing::test_freeze_module_with_tupleoutput_submodule <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0233012Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attr <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0233528Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attribute_on_submodule <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0234159Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_attribute_on_unused_submodule <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0234738Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0235247Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method2 <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0235758Z test_jit.py::TestFreezing::test_freeze_module_with_user_preserved_method_on_submodule <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0236229Z test_jit.py::TestFreezing::test_freeze_no_forward <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0236695Z test_jit.py::TestFreezing::test_freeze_non_interface_module_swap <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0237161Z test_jit.py::TestFreezing::test_freeze_non_module_class_getattr <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0237638Z test_jit.py::TestFreezing::test_freeze_recursive_interfaces <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0238127Z test_jit.py::TestFreezing::test_freeze_recursive_interfaces_same_name <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0238639Z test_jit.py::TestFreezing::test_freeze_recursive_interfaces_with_reassignment <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0239118Z test_jit.py::TestFreezing::test_freeze_with_interface_mutable <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0239597Z test_jit.py::TestFreezing::test_freeze_with_swapping_interfaces <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0240067Z test_jit.py::TestFreezing::test_module_getattr_indirection <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0240548Z test_jit.py::TestFreezing::test_module_with_shared_type_instances <- test/jit/test_freezing.py PASSED [ 24%] 2023-03-31T05:50:36.0241046Z test_jit.py::TestFrozenOptimizations::test_collapse_adjacent_conversions <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0241550Z test_jit.py::TestFrozenOptimizations::test_conv_add_folding <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0242038Z test_jit.py::TestFrozenOptimizations::test_conv_bn_folding <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0242630Z test_jit.py::TestFrozenOptimizations::test_conv_bn_folding_autocast_scenario_cuda <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0243353Z test_jit.py::TestFrozenOptimizations::test_conv_bn_folding_not_forward <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0243854Z test_jit.py::TestFrozenOptimizations::test_conv_hardswish <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0244336Z test_jit.py::TestFrozenOptimizations::test_conv_mul_add_bn <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0244811Z test_jit.py::TestFrozenOptimizations::test_conv_to_mkldnn <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0245332Z test_jit.py::TestFrozenOptimizations::test_conv_to_mkldnn_no_mkldnn <- test/jit/test_freezing.py SKIPPED (Testing no mkldnn) [ 25%] 2023-03-31T05:50:36.0245892Z test_jit.py::TestFrozenOptimizations::test_freeze_conv_relu_fusion <- test/jit/test_freezing.py SKIPPED (requires CUDNN) [ 25%] 2023-03-31T05:50:36.0246464Z test_jit.py::TestFrozenOptimizations::test_freeze_conv_relu_fusion_not_forward <- test/jit/test_freezing.py SKIPPED (requires CUDNN) [ 25%] 2023-03-31T05:50:36.0246975Z test_jit.py::TestFrozenOptimizations::test_freeze_mkdlnn <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0247450Z test_jit.py::TestFrozenOptimizations::test_freeze_remove_dropout <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0247960Z test_jit.py::TestFrozenOptimizations::test_freeze_remove_feature_dropout <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0248468Z test_jit.py::TestFrozenOptimizations::test_hardswish_hardsigmoid <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0249076Z test_jit.py::TestFrozenOptimizations::test_incompatible_perf_formats <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0249641Z test_jit.py::TestFrozenOptimizations::test_linear_bn_folding <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0250246Z test_jit.py::TestFrozenOptimizations::test_linear_bn_folding_autocast_scenario_cuda <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0250869Z test_jit.py::TestFrozenOptimizations::test_linear_concat <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0251461Z test_jit.py::TestFrozenOptimizations::test_linear_concat_complex <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0252087Z test_jit.py::TestFrozenOptimizations::test_linear_concat_different_input <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0252715Z test_jit.py::TestFrozenOptimizations::test_linear_multiple_blocks <- test/jit/test_freezing.py SKIPPED (Optimization currently only run for GPU) [ 25%] 2023-03-31T05:50:36.0253276Z test_jit.py::TestFrozenOptimizations::test_linear_non_constant_weight <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0253760Z test_jit.py::TestFrozenOptimizations::test_linear_transpose <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0254243Z test_jit.py::TestFrozenOptimizations::test_maxpool_mkldnn <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0254743Z test_jit.py::TestFrozenOptimizations::test_mkldnn_fuser_broadcasting <- test/jit/test_freezing.py PASSED [ 25%] 2023-03-31T05:50:36.0255249Z test_jit.py::TestFrozenOptimizations::test_mkldnn_inplace_removal <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0255751Z test_jit.py::TestFrozenOptimizations::test_numel_less_than_size_with_padding <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0256261Z test_jit.py::TestFrozenOptimizations::test_optimize_freeze_module <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0256751Z test_jit.py::TestFrozenOptimizations::test_pool2d_batchnorm <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0257235Z test_jit.py::TestFrozenOptimizations::test_pool3d_batchnorm <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0257702Z test_jit.py::TestFrozenOptimizations::test_remove_detach <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0258198Z test_jit.py::TestFrozenOptimizations::test_remove_detach_not_applied <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0258678Z test_jit.py::TestFrozenOptimizations::test_scalar_mul <- test/jit/test_freezing.py PASSED [ 26%] 2023-03-31T05:50:36.0259676Z test_jit.py::TestMKLDNNReinplacing::test_always_alive_values <- test/jit/test_freezing.py SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/91486 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2023-03-31T05:50:36.0260967Z test_jit.py::TestMKLDNNReinplacing::test_merge_liveness <- test/jit/test_freezing.py SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/91481 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2023-03-31T05:50:36.0262218Z test_jit.py::TestMKLDNNReinplacing::test_successful <- test/jit/test_freezing.py SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/91489 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2023-03-31T05:50:36.0263743Z test_jit.py::TestMKLDNNReinplacing::test_switch_inputs_to_inplace <- test/jit/test_freezing.py SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/91488 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2023-03-31T05:50:36.0264491Z test_jit.py::TestPeephole::test_conv_dim_folding <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0264945Z test_jit.py::TestPeephole::test_integer_refinement <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0265383Z test_jit.py::TestPeephole::test_noop_peephole <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0265823Z test_jit.py::TestPeephole::test_normalized_is_op <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0266254Z test_jit.py::TestPeephole::test_normalized_isnot_op <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0266705Z test_jit.py::TestPeephole::test_normalized_rsub <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0267188Z test_jit.py::TestPeephole::test_optimize_out_comparison_same_value <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0267641Z test_jit.py::TestPeephole::test_peephole <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0268069Z test_jit.py::TestPeephole::test_peephole_add_zero <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0268511Z test_jit.py::TestPeephole::test_peephole_arith <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0268994Z test_jit.py::TestPeephole::test_peephole_cuda <- test/jit/test_peephole.py SKIPPED (cpp tests require CUDA) [ 26%] 2023-03-31T05:50:36.0269514Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_dict_modified <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0270060Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_get_input_arg <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0270618Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_keys_might_overlap <- test/jit/test_peephole.py PASSED [ 26%] 2023-03-31T05:50:36.0271168Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_missing_key <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0271701Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_overlapping_keys <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0272256Z test_jit.py::TestPeephole::test_peephole_dict_getitem_no_optimization_unsupported_type <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0272765Z test_jit.py::TestPeephole::test_peephole_dict_getitem_simple <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0273222Z test_jit.py::TestPeephole::test_peephole_dict_len <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0273721Z test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_keys_might_overlap <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0274270Z test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_overlapping_keys <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0274813Z test_jit.py::TestPeephole::test_peephole_dict_len_no_optimization_unsupported_type <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0275289Z test_jit.py::TestPeephole::test_peephole_dynamic <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0275709Z test_jit.py::TestPeephole::test_peephole_int <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0276147Z test_jit.py::TestPeephole::test_peephole_len_list <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0276590Z test_jit.py::TestPeephole::test_peephole_list_len <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0277014Z test_jit.py::TestPeephole::test_peephole_list_ops <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0277552Z test_jit.py::TestPeephole::test_peephole_no_output_aliasing <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0278086Z test_jit.py::TestPeephole::test_peephole_optional_refine <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0278562Z test_jit.py::TestPeephole::test_peephole_slice_all_three_args <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0279021Z test_jit.py::TestPeephole::test_peephole_slice_one_empty_arg <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0279545Z test_jit.py::TestPeephole::test_peephole_slice_optimization_not_applied_list_modified <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0280094Z test_jit.py::TestPeephole::test_peephole_slice_optimization_not_applied_non_const_args <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0280598Z test_jit.py::TestPeephole::test_peephole_slice_two_empty_args <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0281062Z test_jit.py::TestPeephole::test_peephole_type_refinements <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0281539Z test_jit.py::TestPeephole::test_peephole_with_non_output_writes <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0282005Z test_jit.py::TestPeephole::test_peephole_with_writes <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0282459Z test_jit.py::TestPeephole::test_refine_integer_values <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0282911Z test_jit.py::TestPeephole::test_short_circuit_optimization <- test/jit/test_peephole.py PASSED [ 27%] 2023-03-31T05:50:36.0283603Z test_jit.py::TestAliasAnalysis::test_becomes_wildcard_annotations <- test/jit/test_alias_analysis.py PASSED [ 28%] 2023-03-31T05:50:36.0284127Z test_jit.py::TestAliasAnalysis::test_nested_list_construct_not_wildcard <- test/jit/test_alias_analysis.py PASSED [ 28%] 2023-03-31T05:50:36.0284607Z test_jit.py::TestAliasAnalysis::test_recursive_calls <- test/jit/test_alias_analysis.py PASSED [ 28%] 2023-03-31T05:50:36.0285076Z test_jit.py::TestSaveLoad::test_different_functions <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0285536Z test_jit.py::TestSaveLoad::test_different_interfaces <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0285983Z test_jit.py::TestSaveLoad::test_different_modules <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0286405Z test_jit.py::TestSaveLoad::test_many_collisions <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0286851Z test_jit.py::TestSaveLoad::test_save_load_meta_tensors <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0287332Z test_jit.py::TestSaveLoad::test_save_load_params_buffers_submodules <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0287803Z test_jit.py::TestSaveLoad::test_save_load_using_pathlib <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0288253Z test_jit.py::TestSaveLoad::test_save_load_with_extra_files <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0288790Z test_jit.py::TestSaveLoad::test_save_load_with_saved_traced_inputs <- test/jit/test_save_load.py You have not run this instance of FileCheck! 2023-03-31T05:50:36.0289109Z FileCheck checks: 2023-03-31T05:50:36.0289400Z [W import.cpp:309] Warning: Cannot restore shapes as no traced inputs were stored (function deserialize) 2023-03-31T05:50:36.0289681Z PASSED [ 28%] 2023-03-31T05:50:36.0290052Z test_jit.py::TestSaveLoad::test_save_namedtuple_input_only <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0290543Z test_jit.py::TestSaveLoad::test_save_namedtuple_input_only_forwardref <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0291016Z test_jit.py::TestSaveLoad::test_save_namedtuple_output_only <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0291468Z test_jit.py::TestSaveLoad::test_save_nonexit_file <- test/jit/test_save_load.py PASSED [ 28%] 2023-03-31T05:50:36.0292024Z test_jit.py::TestSaveLoadFlatbuffer::test_different_functions <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0292824Z test_jit.py::TestSaveLoadFlatbuffer::test_different_interfaces <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0293443Z test_jit.py::TestSaveLoadFlatbuffer::test_different_modules <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0294061Z test_jit.py::TestSaveLoadFlatbuffer::test_many_collisions <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0294685Z test_jit.py::TestSaveLoadFlatbuffer::test_module_info_flatbuffer <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0295336Z test_jit.py::TestSaveLoadFlatbuffer::test_save_load_params_buffers_submodules <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0295967Z test_jit.py::TestSaveLoadFlatbuffer::test_save_load_using_pathlib <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0296601Z test_jit.py::TestSaveLoadFlatbuffer::test_save_load_with_extra_files <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0297238Z test_jit.py::TestSaveLoadFlatbuffer::test_save_namedtuple_input_only <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0297866Z test_jit.py::TestSaveLoadFlatbuffer::test_save_namedtuple_output_only <- test/jit/test_save_load.py SKIPPED (Need to enable flatbuffer to run the below tests) [ 28%] 2023-03-31T05:50:36.0298484Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar <- test/jit/test_save_load_for_op_version.py SKIPPED (Failed to load fixture!) [ 29%] 2023-03-31T05:50:36.0299057Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_inplace <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0299629Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_reciprocal <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0300193Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_scalar_scalar <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0300746Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0301293Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor_inplace <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0301846Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_div_tensor_out <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0302388Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_linspace <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0303057Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_linspace_out <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0303598Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_logspace <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0304145Z test_jit.py::TestSaveLoadForOpVersion::test_versioned_logspace_out <- test/jit/test_save_load_for_op_version.py PASSED [ 29%] 2023-03-31T05:50:36.0304681Z test_jit.py::TestModuleContainers::test_custom_container_forward <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0305217Z test_jit.py::TestModuleContainers::test_empty_dict_override_contains <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0305729Z test_jit.py::TestModuleContainers::test_module_inplace_construct <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0306262Z test_jit.py::TestModuleContainers::test_module_interface_special_methods <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0306845Z test_jit.py::TestModuleContainers::test_module_properties <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0307375Z test_jit.py::TestModuleContainers::test_moduledict <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0307875Z test_jit.py::TestModuleContainers::test_moduledict_getitem <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0308385Z test_jit.py::TestModuleContainers::test_moduledict_keyerror <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0308928Z test_jit.py::TestModuleContainers::test_normal_list_attribute_with_modules_error <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0309462Z test_jit.py::TestModuleContainers::test_parameterdict_script_getitem <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0309995Z test_jit.py::TestModuleContainers::test_parameterlist_script_getitem <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0310530Z test_jit.py::TestModuleContainers::test_parameterlist_script_iter <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0311065Z test_jit.py::TestModuleContainers::test_script_module_list_sequential <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0311577Z test_jit.py::TestModuleContainers::test_script_modulelist_index <- test/jit/test_module_containers.py PASSED [ 29%] 2023-03-31T05:50:36.0312106Z test_jit.py::TestModuleContainers::test_sequential_intermediary_types <- test/jit/test_module_containers.py PASSED [ 30%] 2023-03-31T05:50:36.0312642Z test_jit.py::TestModuleContainers::test_special_method_with_override <- test/jit/test_module_containers.py PASSED [ 30%] 2023-03-31T05:50:36.0313153Z test_jit.py::TestModuleContainers::test_typed_module_dict <- test/jit/test_module_containers.py PASSED [ 30%] 2023-03-31T05:50:36.0313635Z test_jit.py::TestModuleContainers::test_typed_module_list <- test/jit/test_module_containers.py PASSED [ 30%] 2023-03-31T05:50:36.0314113Z test_jit.py::TestPythonBindings::test_add_input <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0314582Z test_jit.py::TestPythonBindings::test_aliasdb <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0315054Z test_jit.py::TestPythonBindings::test_canonicalize <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0315526Z test_jit.py::TestPythonBindings::test_cu_create_function <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0316012Z test_jit.py::TestPythonBindings::test_cu_get_functions <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0316486Z test_jit.py::TestPythonBindings::test_graph_create <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0316967Z test_jit.py::TestPythonBindings::test_graph_iterator_keepalive <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0317458Z test_jit.py::TestPythonBindings::test_invalidation <- test/jit/test_python_bindings.py PASSED [ 30%] 2023-03-31T05:50:36.0317908Z test_jit.py::TestPythonIr::test_param_strides <- test/jit/test_python_ir.py PASSED [ 30%] 2023-03-31T05:50:36.0318387Z test_jit.py::TestFunctionalBlocks::test_subgraph_creation <- test/jit/test_functional_blocks.py PASSED [ 30%] 2023-03-31T05:50:36.0318859Z test_jit.py::TestRemoveMutation::test_aten_inplace <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0319350Z test_jit.py::TestRemoveMutation::test_common_pytorch_list_ops <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0319819Z test_jit.py::TestRemoveMutation::test_if_output <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0320290Z test_jit.py::TestRemoveMutation::test_if_output_fail <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0320760Z test_jit.py::TestRemoveMutation::test_list_indexing_removal <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0321302Z test_jit.py::TestRemoveMutation::test_lists_append <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0321843Z test_jit.py::TestRemoveMutation::test_lists_insert <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0322324Z test_jit.py::TestRemoveMutation::test_special_mapped_op <- test/jit/test_remove_mutation.py PASSED [ 30%] 2023-03-31T05:50:36.0322764Z test_jit.py::TestTorchbind::test_default_args <- test/jit/test_torchbind.py PASSED [ 30%] 2023-03-31T05:50:36.0323435Z test_jit.py::TestTorchbind::test_lambda_as_constructor <- test/jit/test_torchbind.py PASSED [ 30%] 2023-03-31T05:50:36.0324048Z test_jit.py::TestTorchbind::test_profiler_custom_op <- test/jit/test_torchbind.py STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0324586Z STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0325075Z STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0325358Z PASSED [ 30%] 2023-03-31T05:50:36.0325718Z test_jit.py::TestTorchbind::test_staticmethod <- test/jit/test_torchbind.py PASSED [ 30%] 2023-03-31T05:50:36.0326137Z test_jit.py::TestTorchbind::test_torchbind <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0326590Z test_jit.py::TestTorchbind::test_torchbind_attr_exception <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0327072Z test_jit.py::TestTorchbind::test_torchbind_class_attr_recursive <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0327535Z test_jit.py::TestTorchbind::test_torchbind_class_attribute <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0327992Z test_jit.py::TestTorchbind::test_torchbind_deepcopy <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0328477Z test_jit.py::TestTorchbind::test_torchbind_def_property_getter_setter <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0328978Z test_jit.py::TestTorchbind::test_torchbind_def_property_just_getter <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0329465Z test_jit.py::TestTorchbind::test_torchbind_def_property_readwrite <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0329932Z test_jit.py::TestTorchbind::test_torchbind_getattr <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0330381Z test_jit.py::TestTorchbind::test_torchbind_getstate <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0330865Z test_jit.py::TestTorchbind::test_torchbind_instantiate_missing_class <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0331330Z test_jit.py::TestTorchbind::test_torchbind_lambda_method <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0331782Z test_jit.py::TestTorchbind::test_torchbind_no_init <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0332253Z test_jit.py::TestTorchbind::test_torchbind_optional_explicit_attr <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0332736Z test_jit.py::TestTorchbind::test_torchbind_pass_wrong_type <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0333209Z test_jit.py::TestTorchbind::test_torchbind_pickle_serialization <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0333692Z test_jit.py::TestTorchbind::test_torchbind_python_deepcopy <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0334163Z test_jit.py::TestTorchbind::test_torchbind_return_instance <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0334639Z test_jit.py::TestTorchbind::test_torchbind_return_instance_from_method <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0335112Z test_jit.py::TestTorchbind::test_torchbind_return_tuple <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0335567Z test_jit.py::TestTorchbind::test_torchbind_save_load <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0336119Z test_jit.py::TestTorchbind::test_torchbind_take_as_arg <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0336595Z test_jit.py::TestTorchbind::test_torchbind_take_instance_as_method_arg <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0337138Z test_jit.py::TestTorchbind::test_torchbind_tracing <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0337602Z test_jit.py::TestTorchbind::test_torchbind_tracing_nested <- test/jit/test_torchbind.py PASSED [ 31%] 2023-03-31T05:50:36.0338272Z test_jit.py::TestModuleInterface::test_freeze_module_with_inplace_mutation_in_interface <- test/jit/test_module_interface.py PASSED [ 31%] 2023-03-31T05:50:36.0338810Z test_jit.py::TestModuleInterface::test_freeze_module_with_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0339350Z test_jit.py::TestModuleInterface::test_freeze_module_with_interface_and_fork <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0339906Z test_jit.py::TestModuleInterface::test_freeze_module_with_mutated_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0340462Z test_jit.py::TestModuleInterface::test_freeze_module_with_setattr_in_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0340976Z test_jit.py::TestModuleInterface::test_module_apis_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0341475Z test_jit.py::TestModuleInterface::test_module_doc_string <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0341968Z test_jit.py::TestModuleInterface::test_module_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0342484Z test_jit.py::TestModuleInterface::test_module_interface_inheritance <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0343066Z test_jit.py::TestModuleInterface::test_module_interface_subtype <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0343558Z test_jit.py::TestModuleInterface::test_module_swap <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0344065Z test_jit.py::TestModuleInterface::test_module_swap_no_lazy_compile <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0344578Z test_jit.py::TestModuleInterface::test_module_swap_no_module_interface <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0345098Z test_jit.py::TestModuleInterface::test_module_swap_wrong_module <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0345617Z test_jit.py::TestModuleInterface::test_not_submodule_interface_call <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0346144Z test_jit.py::TestModuleInterface::test_script_module_as_interface_swap <- test/jit/test_module_interface.py PASSED [ 32%] 2023-03-31T05:50:36.0346580Z test_jit.py::TestWith::test_with_as <- test/jit/test_with.py PASSED [ 32%] 2023-03-31T05:50:36.0346975Z test_jit.py::TestWith::test_with_errors <- test/jit/test_with.py PASSED [ 32%] 2023-03-31T05:50:36.0347380Z test_jit.py::TestWith::test_with_exceptions <- test/jit/test_with.py PASSED [ 32%] 2023-03-31T05:50:36.0347781Z test_jit.py::TestWith::test_with_no_as <- test/jit/test_with.py PASSED [ 32%] 2023-03-31T05:50:36.0348162Z test_jit.py::TestWith::test_with_no_grad <- test/jit/test_with.py PASSED [ 32%] 2023-03-31T05:50:36.0348716Z test_jit.py::TestWith::test_with_record_function <- test/jit/test_with.py STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0349242Z STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0349713Z STAGE:2023-03-31 05:49:15 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0349977Z PASSED [ 32%] 2023-03-31T05:50:36.0350330Z test_jit.py::TestEnum::test_closed_over_enum_constant <- test/jit/test_enum.py PASSED [ 32%] 2023-03-31T05:50:36.0350740Z test_jit.py::TestEnum::test_enum_as_const <- test/jit/test_enum.py PASSED [ 32%] 2023-03-31T05:50:36.0351228Z test_jit.py::TestEnum::test_enum_as_module_attribute <- test/jit/test_enum.py PASSED [ 32%] 2023-03-31T05:50:36.0351688Z test_jit.py::TestEnum::test_enum_comp <- test/jit/test_enum.py PASSED [ 32%] 2023-03-31T05:50:36.0352104Z test_jit.py::TestEnum::test_enum_comp_diff_classes <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0352527Z test_jit.py::TestEnum::test_enum_explicit_script <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0352919Z test_jit.py::TestEnum::test_enum_iterate <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0353318Z test_jit.py::TestEnum::test_enum_ivalue_type <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0353731Z test_jit.py::TestEnum::test_enum_module_return <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0354115Z test_jit.py::TestEnum::test_enum_name <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0354503Z test_jit.py::TestEnum::test_enum_return <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0354895Z test_jit.py::TestEnum::test_enum_value <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0355294Z test_jit.py::TestEnum::test_enum_value_types <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0355721Z test_jit.py::TestEnum::test_heterogenous_value_type_enum_error <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0356156Z test_jit.py::TestEnum::test_non_existent_enum_value <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0356592Z test_jit.py::TestEnum::test_string_enum_as_module_attribute <- test/jit/test_enum.py PASSED [ 33%] 2023-03-31T05:50:36.0357057Z test_jit.py::TestStringFormatting::test_modulo_operator <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0357631Z test_jit.py::TestStringFormatting::test_string_interpolation_with_alternate_digit_placeholder <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0358275Z test_jit.py::TestStringFormatting::test_string_interpolation_with_capital_exponent_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0358918Z test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_char_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0359531Z test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0360166Z test_jit.py::TestStringFormatting::test_string_interpolation_with_char_placeholder_and_true_string_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0360802Z test_jit.py::TestStringFormatting::test_string_interpolation_with_digit_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0361433Z test_jit.py::TestStringFormatting::test_string_interpolation_with_digit_placeholder_and_string_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0362041Z test_jit.py::TestStringFormatting::test_string_interpolation_with_double_percent_in_string <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0362650Z test_jit.py::TestStringFormatting::test_string_interpolation_with_exponent_placeholder_and_string_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0363405Z test_jit.py::TestStringFormatting::test_string_interpolation_with_float_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0364033Z test_jit.py::TestStringFormatting::test_string_interpolation_with_float_placeholder_and_float_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0364681Z test_jit.py::TestStringFormatting::test_string_interpolation_with_lowercase_exponent_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0365282Z test_jit.py::TestStringFormatting::test_string_interpolation_with_multiple_placeholders <- test/jit/test_string_formatting.py PASSED [ 33%] 2023-03-31T05:50:36.0366066Z test_jit.py::TestStringFormatting::test_string_interpolation_with_percent_in_string <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0366678Z test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_digit_variable <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0367324Z test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_format_string_variable <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0367947Z test_jit.py::TestStringFormatting::test_string_interpolation_with_string_placeholder_and_string_variable <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0368534Z test_jit.py::TestStringFormatting::test_string_interpolation_with_subscript <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0369102Z test_jit.py::TestStringFormatting::test_string_interpolation_with_too_few_arguments <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0369679Z test_jit.py::TestStringFormatting::test_string_interpolation_with_too_many_arguments <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0370252Z test_jit.py::TestStringFormatting::test_string_interpolation_with_unknown_format_specifier <- test/jit/test_string_formatting.py PASSED [ 34%] 2023-03-31T05:50:36.0370746Z test_jit.py::TestProfiler::test_aliasing_merge <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0371201Z test_jit.py::TestProfiler::test_autograd_fallback_graph <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0371684Z test_jit.py::TestProfiler::test_fallback_graph_not_specialized <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0372130Z test_jit.py::TestProfiler::test_iterative_fusion <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0372581Z test_jit.py::TestProfiler::test_local_fusion_strategy <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0373034Z test_jit.py::TestProfiler::test_not_fusing_scalar_ops <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0373495Z test_jit.py::TestProfiler::test_not_optimizing_property <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0373936Z test_jit.py::TestProfiler::test_specialize_backward <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0374379Z test_jit.py::TestProfiler::test_specialized_types <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0374812Z test_jit.py::TestProfiler::test_tensor_constant <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0375268Z test_jit.py::TestProfiler::test_tensor_type_not_determined_by_inputs <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0375729Z test_jit.py::TestProfiler::test_use_not_profiled <- test/jit/test_profiler.py PASSED [ 34%] 2023-03-31T05:50:36.0376158Z test_jit.py::TestSlice::test_module_list_slicing <- test/jit/test_slice.py PASSED [ 34%] 2023-03-31T05:50:36.0376580Z test_jit.py::TestSlice::test_slice_as_variable <- test/jit/test_slice.py PASSED [ 34%] 2023-03-31T05:50:36.0376989Z test_jit.py::TestSlice::test_slice_dynamic_index <- test/jit/test_slice.py PASSED [ 34%] 2023-03-31T05:50:36.0377397Z test_jit.py::TestSlice::test_slice_kwarg <- test/jit/test_slice.py PASSED [ 34%] 2023-03-31T05:50:36.0377799Z test_jit.py::TestSlice::test_slice_one_none <- test/jit/test_slice.py PASSED [ 34%] 2023-03-31T05:50:36.0378206Z test_jit.py::TestSlice::test_slice_start_stop <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0378610Z test_jit.py::TestSlice::test_slice_start_stop_step <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0379044Z test_jit.py::TestSlice::test_slice_start_stop_with_none <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0379469Z test_jit.py::TestSlice::test_slice_stop_clipped <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0379954Z test_jit.py::TestSlice::test_slice_stop_only <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0380379Z test_jit.py::TestSlice::test_slice_stop_only_with_nones <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0380845Z test_jit.py::TestSlice::test_slice_string <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0381247Z test_jit.py::TestSlice::test_slice_tensor <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0381653Z test_jit.py::TestSlice::test_slice_tensor_multidim <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0382101Z test_jit.py::TestSlice::test_slice_tensor_multidim_with_dots <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0382536Z test_jit.py::TestSlice::test_slice_three_nones <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0383014Z test_jit.py::TestSlice::test_slice_two_nones <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0383434Z test_jit.py::TestSlice::test_tuple_slicing <- test/jit/test_slice.py PASSED [ 35%] 2023-03-31T05:50:36.0383903Z test_jit.py::TestIgnorableArgs::test_add_out_ignorable_args <- test/jit/test_ignorable_args.py PASSED [ 35%] 2023-03-31T05:50:36.0384417Z test_jit.py::TestIgnorableArgs::test_slice_ignorable_args_for_slice <- test/jit/test_ignorable_args.py PASSED [ 35%] 2023-03-31T05:50:36.0384860Z test_jit.py::TestHooks::test_forward_tuple_input <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0385300Z test_jit.py::TestHooks::test_hook_compilation_hint <- test/jit/test_hooks.py SKIPPED [ 35%] 2023-03-31T05:50:36.0385737Z test_jit.py::TestHooks::test_hook_hook_name_collision <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0386159Z test_jit.py::TestHooks::test_hook_method_name_collision <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0386613Z test_jit.py::TestHooks::test_module_direct_forward_invocation <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0387067Z test_jit.py::TestHooks::test_module_forward_multiple_inputs <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0387524Z test_jit.py::TestHooks::test_module_forward_single_input <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0387953Z test_jit.py::TestHooks::test_module_hook_return_nothing <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0388419Z test_jit.py::TestHooks::test_module_multiple_hooks_multiple_inputs <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0388887Z test_jit.py::TestHooks::test_module_multiple_hooks_single_input <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0389314Z test_jit.py::TestHooks::test_module_no_forward_input <- test/jit/test_hooks.py PASSED [ 35%] 2023-03-31T05:50:36.0389750Z test_jit.py::TestHooks::test_module_same_hook_repeated <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0390206Z test_jit.py::TestHooks::test_submodule_called_directly_with_hooks <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0390673Z test_jit.py::TestHooks::test_submodule_direct_forward_invocation <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0391130Z test_jit.py::TestHooks::test_submodule_forward_multiple_inputs <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0391592Z test_jit.py::TestHooks::test_submodule_forward_single_input <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0392076Z test_jit.py::TestHooks::test_submodule_forward_single_input_return_not_tupled <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0392545Z test_jit.py::TestHooks::test_submodule_hook_return_nothing <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0393003Z test_jit.py::TestHooks::test_submodule_multiple_hooks_multiple_inputs <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0393476Z test_jit.py::TestHooks::test_submodule_multiple_hooks_single_input <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0393925Z test_jit.py::TestHooks::test_submodule_no_forward_input <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0394359Z test_jit.py::TestHooks::test_submodule_same_hook_repeated <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0394859Z test_jit.py::TestHooks::test_wrong_hook_signatures <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0395345Z test_jit.py::TestHooks::test_wrong_pre_hook_signatures <- test/jit/test_hooks.py PASSED [ 36%] 2023-03-31T05:50:36.0395749Z test_jit.py::TestWarn::test_warn <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0396177Z test_jit.py::TestWarn::test_warn_multiple_calls_multiple_warnings <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0396646Z test_jit.py::TestWarn::test_warn_multiple_calls_same_func_diff_stack <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0397077Z test_jit.py::TestWarn::test_warn_once_per_func <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0397496Z test_jit.py::TestWarn::test_warn_once_per_func_in_loop <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0397887Z test_jit.py::TestWarn::test_warn_only_once <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0398309Z test_jit.py::TestWarn::test_warn_only_once_in_loop_func <- test/jit/test_warn.py PASSED [ 36%] 2023-03-31T05:50:36.0398730Z test_jit.py::TestIsinstance::test_bool <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0399139Z test_jit.py::TestIsinstance::test_dict <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0399562Z test_jit.py::TestIsinstance::test_dict_nested <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0400021Z test_jit.py::TestIsinstance::test_dict_no_contained_type <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0400465Z test_jit.py::TestIsinstance::test_dict_tensor <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0400922Z test_jit.py::TestIsinstance::test_empty_container_special_cases <- test/jit/test_isinstance.py PASSED [ 36%] 2023-03-31T05:50:36.0401435Z test_jit.py::TestIsinstance::test_empty_container_throws_warning_in_eager <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0401891Z test_jit.py::TestIsinstance::test_float <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0402298Z test_jit.py::TestIsinstance::test_if_else <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0402719Z test_jit.py::TestIsinstance::test_in_if <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0403297Z test_jit.py::TestIsinstance::test_in_while_loop <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0403723Z test_jit.py::TestIsinstance::test_int <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0404122Z test_jit.py::TestIsinstance::test_list <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0404546Z test_jit.py::TestIsinstance::test_list_nested <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0404997Z test_jit.py::TestIsinstance::test_list_no_contained_type <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0405441Z test_jit.py::TestIsinstance::test_list_tensor <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0405881Z test_jit.py::TestIsinstance::test_list_tensor_type_true <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0406382Z test_jit.py::TestIsinstance::test_nontuple_container_rhs_throws_in_eager <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0406839Z test_jit.py::TestIsinstance::test_optional <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0407269Z test_jit.py::TestIsinstance::test_optional_nested <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0407740Z test_jit.py::TestIsinstance::test_optional_no_contained_type <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0408193Z test_jit.py::TestIsinstance::test_optional_none <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0408644Z test_jit.py::TestIsinstance::test_tensor_type_false <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0409061Z test_jit.py::TestIsinstance::test_tuple <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0409597Z test_jit.py::TestIsinstance::test_tuple_nested <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0410123Z test_jit.py::TestIsinstance::test_tuple_no_contained_type <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0410571Z test_jit.py::TestIsinstance::test_tuple_rhs <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0410993Z test_jit.py::TestIsinstance::test_tuple_tensor <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0411437Z test_jit.py::TestIsinstance::test_type_refinement <- test/jit/test_isinstance.py PASSED [ 37%] 2023-03-31T05:50:36.0411887Z test_jit.py::TestPythonBuiltinOP::test_add <- test/jit/test_python_builtins.py PASSED [ 37%] 2023-03-31T05:50:36.0412353Z test_jit.py::TestPythonBuiltinOP::test_adv_indexing_list <- test/jit/test_python_builtins.py PASSED [ 37%] 2023-03-31T05:50:36.0412842Z test_jit.py::TestPythonBuiltinOP::test_advancedindex <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0413313Z test_jit.py::TestPythonBuiltinOP::test_gather <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0413768Z test_jit.py::TestPythonBuiltinOP::test_index <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0414231Z test_jit.py::TestPythonBuiltinOP::test_index_ellipses <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0414685Z test_jit.py::TestPythonBuiltinOP::test_inf <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0415148Z test_jit.py::TestPythonBuiltinOP::test_matmul_py3 <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0415606Z test_jit.py::TestPythonBuiltinOP::test_mul <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0416038Z test_jit.py::TestPythonBuiltinOP::test_pow <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0416493Z test_jit.py::TestPythonBuiltinOP::test_random <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0416950Z test_jit.py::TestPythonBuiltinOP::test_slice <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0417418Z test_jit.py::TestPythonBuiltinOP::test_stepped_tuple_slicing <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0417907Z test_jit.py::TestPythonBuiltinOP::test_str_to_float <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0418368Z test_jit.py::TestPythonBuiltinOP::test_triple <- test/jit/test_python_builtins.py PASSED [ 38%] 2023-03-31T05:50:36.0418800Z test_jit.py::TestTyping::test_bool_list_io <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0419210Z test_jit.py::TestTyping::test_dict_comprehension <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0419656Z test_jit.py::TestTyping::test_dict_comprehension_scope <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0420133Z test_jit.py::TestTyping::test_dict_comprehension_with_type_annotation <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0420565Z test_jit.py::TestTyping::test_dict_in_not_in <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0420999Z test_jit.py::TestTyping::test_dict_invalid_annotations <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0421916Z test_jit.py::TestTyping::test_dict_type_refinement_annotation_key_mismatch <- test/jit/test_typing.py [W ir_emitter.cpp:4434] Warning: List consists of heterogeneous types, which means that it has been typed as containing Union[int, str]. To use any of the values in this List, it will be necessary to add an `assert isinstance` statement before first use to trigger type refinement. 2023-03-31T05:50:36.0422537Z File "/var/lib/jenkins/workspace/test/jit/test_typing.py", line 90 2023-03-31T05:50:36.0422772Z def fn(): 2023-03-31T05:50:36.0423026Z l1 = [1, 2, "foo", 3] 2023-03-31T05:50:36.0423275Z ~~~~~~~~~~~~~~~ <--- HERE 2023-03-31T05:50:36.0423484Z l2 = ["foo", "bar", "baz", "qux"] 2023-03-31T05:50:36.0423700Z d: Dict[int, str] = dict(zip(l1, l2)) 2023-03-31T05:50:36.0423996Z (function emitListLiteral) 2023-03-31T05:50:36.0424198Z PASSED [ 38%] 2023-03-31T05:50:36.0425088Z test_jit.py::TestTyping::test_dict_type_refinement_annotation_value_mismatch <- test/jit/test_typing.py [W ir_emitter.cpp:4434] Warning: List consists of heterogeneous types, which means that it has been typed as containing Union[int, str]. To use any of the values in this List, it will be necessary to add an `assert isinstance` statement before first use to trigger type refinement. 2023-03-31T05:50:36.0425691Z File "/var/lib/jenkins/workspace/test/jit/test_typing.py", line 104 2023-03-31T05:50:36.0425928Z def fn(): 2023-03-31T05:50:36.0426125Z l1 = ["foo", "bar", "baz", "qux"] 2023-03-31T05:50:36.0426317Z l2 = [1, 2, "foo", 3] 2023-03-31T05:50:36.0426448Z ~~~~~~~~~~~~~~~ <--- HERE 2023-03-31T05:50:36.0426556Z d: Dict[str, int] = dict(zip(l1, l2)) 2023-03-31T05:50:36.0426633Z return d 2023-03-31T05:50:36.0426725Z (function emitListLiteral) 2023-03-31T05:50:36.0426799Z PASSED [ 38%] 2023-03-31T05:50:36.0427022Z test_jit.py::TestTyping::test_for_in_dict <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0427264Z test_jit.py::TestTyping::test_for_in_string <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0427504Z test_jit.py::TestTyping::test_for_tuple_assign <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0427744Z test_jit.py::TestTyping::test_for_tuple_unpack <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0427967Z test_jit.py::TestTyping::test_list_io <- test/jit/test_typing.py PASSED [ 38%] 2023-03-31T05:50:36.0428202Z test_jit.py::TestTyping::test_list_iterables <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0428420Z test_jit.py::TestTyping::test_list_sum <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0428716Z test_jit.py::TestTyping::test_list_type_refinement_annotation_element_mismatch <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0428945Z test_jit.py::TestTyping::test_list_unification <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0429187Z test_jit.py::TestTyping::test_multiple_assign <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0429465Z test_jit.py::TestTyping::test_namedtuple_error_source_attribution <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0429713Z test_jit.py::TestTyping::test_namedtuple_good_error <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0429948Z test_jit.py::TestTyping::test_namedtuple_py2 <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0430193Z test_jit.py::TestTyping::test_namedtuple_redefine <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0430421Z test_jit.py::TestTyping::test_nested_list <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0430662Z test_jit.py::TestTyping::test_opt_opt_refinement <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0430910Z test_jit.py::TestTyping::test_optional_conversion <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0431136Z test_jit.py::TestTyping::test_optional_refinement <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0431375Z test_jit.py::TestTyping::test_optional_tuple <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0431623Z test_jit.py::TestTyping::test_singleton_tuple_unpack <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0431859Z test_jit.py::TestTyping::test_sum_list_diff_elms <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0432087Z test_jit.py::TestTyping::test_sum_list_empty <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0432322Z test_jit.py::TestTyping::test_sum_list_literal <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0432551Z test_jit.py::TestTyping::test_sum_list_one <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0432787Z test_jit.py::TestTyping::test_sum_list_wrong_type <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0433074Z test_jit.py::TestTyping::test_tuple_assignments <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0433312Z test_jit.py::TestTyping::test_tuple_create_return <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0433593Z test_jit.py::TestTyping::test_tuple_io <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0433829Z test_jit.py::TestTyping::test_tuple_keyword <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0434076Z test_jit.py::TestTyping::test_tuple_specialization <- test/jit/test_typing.py PASSED [ 39%] 2023-03-31T05:50:36.0434292Z test_jit.py::TestHash::test_hash_bool <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0434510Z test_jit.py::TestHash::test_hash_device <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0434727Z test_jit.py::TestHash::test_hash_float <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0434944Z test_jit.py::TestHash::test_hash_int <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0435149Z test_jit.py::TestHash::test_hash_none <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0435366Z test_jit.py::TestHash::test_hash_string <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0435589Z test_jit.py::TestHash::test_hash_tensor <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0435803Z test_jit.py::TestHash::test_hash_tuple <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0436065Z test_jit.py::TestHash::test_hash_tuple_nested_unhashable_type <- test/jit/test_hash.py PASSED [ 40%] 2023-03-31T05:50:36.0436322Z test_jit.py::TestComplex::test_binary_op_complex_tensor <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0436560Z test_jit.py::TestComplex::test_comparison_ops <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0436819Z test_jit.py::TestComplex::test_complex_constants_and_ops <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0437053Z test_jit.py::TestComplex::test_complex_constructor <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0437295Z test_jit.py::TestComplex::test_complex_list_sum <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0437537Z test_jit.py::TestComplex::test_complex_parse <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0437771Z test_jit.py::TestComplex::test_complexdict <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0438005Z test_jit.py::TestComplex::test_complexlist <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0438226Z test_jit.py::TestComplex::test_div <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0438464Z test_jit.py::TestComplex::test_infj_nanj_pickle <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0438686Z test_jit.py::TestComplex::test_pickle <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0438894Z test_jit.py::TestComplex::test_script <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0439137Z test_jit.py::TestComplex::test_tensor_attributes <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0439424Z test_jit.py::TestComplex::test_torch_complex_constructor_with_tensor <- test/jit/test_complex.py PASSED [ 40%] 2023-03-31T05:50:36.0439718Z test_jit.py::TestJitUtils::test_checkscriptassertraisesregex <- test/jit/test_jit_utils.py PASSED [ 40%] 2023-03-31T05:50:36.0439998Z test_jit.py::TestJitUtils::test_get_callable_argument_names_hybrid <- test/jit/test_jit_utils.py PASSED [ 40%] 2023-03-31T05:50:36.0440290Z test_jit.py::TestJitUtils::test_get_callable_argument_names_keyword_only <- test/jit/test_jit_utils.py PASSED [ 40%] 2023-03-31T05:50:36.0440589Z test_jit.py::TestJitUtils::test_get_callable_argument_names_positional_only <- test/jit/test_jit_utils.py PASSED [ 41%] 2023-03-31T05:50:36.0440893Z test_jit.py::TestJitUtils::test_get_callable_argument_names_positional_or_keyword <- test/jit/test_jit_utils.py PASSED [ 41%] 2023-03-31T05:50:36.0441183Z test_jit.py::TestJitUtils::test_get_callable_argument_names_var_keyword <- test/jit/test_jit_utils.py PASSED [ 41%] 2023-03-31T05:50:36.0441514Z test_jit.py::TestJitUtils::test_get_callable_argument_names_var_positional <- test/jit/test_jit_utils.py PASSED [ 41%] 2023-03-31T05:50:36.0441849Z test_jit.py::TestJitUtils::test_no_tracer_warn_context_manager <- test/jit/test_jit_utils.py PASSED [ 41%] 2023-03-31T05:50:36.0442281Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_annotation_and_init_annotation <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0442683Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_annotation_only <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0443219Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_class_level_jit_annotation <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0443590Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_dict <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0443968Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_list <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0444351Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_optional <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0444723Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_empty_tensor <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0445099Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_falsy_base_type <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0445481Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_nonempty_container <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0445850Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_attribute <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0446239Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_dict <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0446621Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_list <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0447012Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_jit_empty_optional <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0447397Z test_jit.py::TestScriptModuleInstanceAttributeTypeAnnotation::test_annotated_with_torch_jit_import <- test/jit/test_scriptmod_ann.py PASSED [ 41%] 2023-03-31T05:50:36.0447674Z test_jit.py::TestTypesAndAnnotation::test_annotate_outside_init <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0447926Z test_jit.py::TestTypesAndAnnotation::test_bad_types <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0448194Z test_jit.py::TestTypesAndAnnotation::test_ignore_with_types <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0448504Z test_jit.py::TestTypesAndAnnotation::test_ignoring_fn_with_nonscriptable_types <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0448796Z test_jit.py::TestTypesAndAnnotation::test_ignoring_module_attributes <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0449068Z test_jit.py::TestTypesAndAnnotation::test_inferred_type_error_message <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0449345Z test_jit.py::TestTypesAndAnnotation::test_mismatched_annotation <- test/jit/test_types.py PASSED [ 41%] 2023-03-31T05:50:36.0449648Z test_jit.py::TestTypesAndAnnotation::test_optional_no_element_type_annotation <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0449903Z test_jit.py::TestTypesAndAnnotation::test_parser_bug <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0450157Z test_jit.py::TestTypesAndAnnotation::test_pep585_type <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0450511Z test_jit.py::TestTypesAndAnnotation::test_python_callable <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0450766Z test_jit.py::TestTypesAndAnnotation::test_reannotate <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0451127Z test_jit.py::TestTypesAndAnnotation::test_tuple_no_element_type_annotation <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0451398Z test_jit.py::TestTypesAndAnnotation::test_type_annotate_py3 <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0451648Z test_jit.py::TestTypesAndAnnotation::test_types_as_values <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0451931Z test_jit.py::TestTypesAndAnnotation::test_unimported_type_resolution <- test/jit/test_types.py PASSED [ 42%] 2023-03-31T05:50:36.0452163Z test_jit.py::TestMisc::test_broadcasting_list <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0452408Z test_jit.py::TestMisc::test_export_opnames_interface <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0452642Z test_jit.py::TestMisc::test_future_isinstance <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0452860Z test_jit.py::TestMisc::test_hacked_twin <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0453095Z test_jit.py::TestMisc::test_if_returning_any <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0453312Z test_jit.py::TestMisc::test_joined_str <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0453522Z test_jit.py::TestMisc::test_kwarg_support <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0453767Z test_jit.py::TestMisc::test_legacy_tensor_constructor <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0453995Z test_jit.py::TestMisc::test_list_literal_infer <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0454215Z test_jit.py::TestMisc::test_math_inf <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0454445Z test_jit.py::TestMisc::test_parse_ir_annotate <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0454719Z test_jit.py::TestMisc::test_parse_ir_single_element_tensor_negative <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0454990Z test_jit.py::TestMisc::test_parse_ir_single_element_tensor_positive <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0455276Z test_jit.py::TestMisc::test_pow_multiple_dtype <- test/jit/test_misc.py SKIPPED (need CUDA half support) [ 42%] 2023-03-31T05:50:36.0455516Z test_jit.py::TestMisc::test_script_many_decorators <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0455726Z test_jit.py::TestMisc::test_str_refine_any <- test/jit/test_misc.py PASSED [ 42%] 2023-03-31T05:50:36.0455981Z test_jit.py::TestMisc::test_subexpression_Dict_int_Future <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0456235Z test_jit.py::TestMisc::test_subexpression_Future_annotate <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0456477Z test_jit.py::TestMisc::test_subexpression_List_Future <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0456723Z test_jit.py::TestMisc::test_subexpression_Optional <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0456985Z test_jit.py::TestMisc::test_subexpression_Tuple_int_int_Future <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0457232Z test_jit.py::TestMisc::test_tuple_subscripted_assign <- test/jit/test_misc.py PASSED [ 43%] 2023-03-31T05:50:36.0457497Z test_jit.py::TestUpgraders::test_add_value_to_version_map <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0457751Z test_jit.py::TestUpgraders::test_aten_div_scalar_at_3 <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0457994Z test_jit.py::TestUpgraders::test_aten_div_tensor_at_3 <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0458256Z test_jit.py::TestUpgraders::test_aten_div_tensor_out_at_3 <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0458500Z test_jit.py::TestUpgraders::test_aten_full_at_4 <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0458763Z test_jit.py::TestUpgraders::test_aten_full_other_variants <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0459085Z test_jit.py::TestUpgraders::test_aten_full_out_at_4 <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0459410Z test_jit.py::TestUpgraders::test_aten_linspace <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0459665Z test_jit.py::TestUpgraders::test_aten_linspace_out <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0459911Z test_jit.py::TestUpgraders::test_aten_logspace <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0460149Z test_jit.py::TestUpgraders::test_aten_logspace_out <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0460413Z test_jit.py::TestUpgraders::test_aten_test_serialization <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0460693Z test_jit.py::TestUpgraders::test_populated_test_upgrader_graph <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0460966Z test_jit.py::TestUpgraders::test_populated_upgrader_graph <- test/jit/test_upgraders.py PASSED [ 43%] 2023-03-31T05:50:36.0461271Z test_jit.py::TestTensorCreationOps::test_randperm_default_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0461581Z test_jit.py::TestTensorCreationOps::test_randperm_specifed_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0461884Z test_jit.py::TestTensorCreationOps::test_tril_indices_default_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0462194Z test_jit.py::TestTensorCreationOps::test_tril_indices_specified_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0462498Z test_jit.py::TestTensorCreationOps::test_triu_indices_default_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0462807Z test_jit.py::TestTensorCreationOps::test_triu_indices_specified_dtype <- test/jit/test_tensor_creation_ops.py PASSED [ 43%] 2023-03-31T05:50:36.0463161Z test_jit.py::TestModuleAPIs::test_customized_state_dict_methods <- test/jit/test_module_apis.py PASSED [ 44%] 2023-03-31T05:50:36.0463449Z test_jit.py::TestModuleAPIs::test_default_state_dict_methods <- test/jit/test_module_apis.py PASSED [ 44%] 2023-03-31T05:50:36.0463752Z test_jit.py::TestModuleAPIs::test_submodule_customized_state_dict_methods <- test/jit/test_module_apis.py PASSED [ 44%] 2023-03-31T05:50:36.0464003Z test_jit.py::TestScriptProfile::test_basic <- test/jit/test_script_profile.py PASSED [ 44%] 2023-03-31T05:50:36.0464253Z test_jit.py::TestScriptProfile::test_empty <- test/jit/test_script_profile.py PASSED [ 44%] 2023-03-31T05:50:36.0464497Z test_jit.py::TestScriptProfile::test_multi <- test/jit/test_script_profile.py PASSED [ 44%] 2023-03-31T05:50:36.0464750Z test_jit.py::TestScriptProfile::test_script <- test/jit/test_script_profile.py PASSED [ 44%] 2023-03-31T05:50:36.0465004Z test_jit.py::TestScriptProfile::test_section <- test/jit/test_script_profile.py PASSED [ 44%] 2023-03-31T05:50:36.0465341Z test_jit.py::TestFunctionalToInplaceActivation::test_check_no_type_promotion <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0465691Z test_jit.py::TestFunctionalToInplaceActivation::test_functional_to_inplace_activation <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0466028Z test_jit.py::TestFunctionalToInplaceActivation::test_no_functional_to_inplace <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0466355Z test_jit.py::TestFunctionalToInplaceActivation::test_resnet18_correctness <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0466707Z test_jit.py::TestInplaceToFunctionalActivation::test_inplace_to_functional_activation <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0467034Z test_jit.py::TestInplaceToFunctionalActivation::test_resnet18_correctness <- test/jit/test_convert_activation.py PASSED [ 44%] 2023-03-31T05:50:36.0467309Z test_jit.py::TestParametrization::test_scriptable <- test/jit/test_parametrization.py PASSED [ 44%] 2023-03-31T05:50:36.0467638Z test_jit.py::TestParametrization::test_traceable <- test/jit/test_parametrization.py PASSED [ 44%] 2023-03-31T05:50:36.0467949Z test_jit.py::TestGetDefaultAttr::test_getattr_with_default <- test/jit/test_attr.py PASSED [ 44%] 2023-03-31T05:50:36.0468219Z test_jit.py::TestAtenPow::test_aten_pow_zero_negative_exponent <- test/jit/test_aten_pow.py PASSED [ 44%] 2023-03-31T05:50:36.0468636Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_1 <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0469058Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_2 <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0469481Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_3 <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0469908Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_fuse_activation_with_pack_ops_linear_conv2d_4 <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0470350Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_insert_pre_packed_linear_before_inline_and_conv_2d_op <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0470738Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_insert_pre_packed_linear_op <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0471136Z test_jit.py::TestOptimizeForMobilePreserveDebugInfo::test_replace_conv1d_with_conv2d <- test/jit/test_optimize_for_mobile_preserve_debug_info.py PASSED [ 44%] 2023-03-31T05:50:36.0471379Z test_jit.py::TestUnion::test_check_union_annotation <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0471652Z test_jit.py::TestUnion::test_union_T_None_is_equivalent_to_optional_T <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0471918Z test_jit.py::TestUnion::test_union_argument_order_is_ignored <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0472200Z test_jit.py::TestUnion::test_union_argument_order_is_ignored_container <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0472425Z test_jit.py::TestUnion::test_union_as_annotation <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0472699Z test_jit.py::TestUnion::test_union_as_annotation_in_typed_container <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0472942Z test_jit.py::TestUnion::test_union_as_annotation_py2 <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0473173Z test_jit.py::TestUnion::test_union_as_dict_key <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0473408Z test_jit.py::TestUnion::test_union_as_dict_value <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0473661Z test_jit.py::TestUnion::test_union_as_internal_tuple_type <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0473962Z test_jit.py::TestUnion::test_union_branching_does_not_autoinfer_undeclared_union <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0474266Z test_jit.py::TestUnion::test_union_branching_does_not_widen_existing_inferred_type <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0474563Z test_jit.py::TestUnion::test_union_branching_with_union_return_and_homogenous_types <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0474836Z test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0475153Z test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type_empty_container <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0475448Z test_jit.py::TestUnion::test_union_does_not_replace_existing_annotated_type_union <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0475696Z test_jit.py::TestUnion::test_union_in_class_constructor <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0475990Z test_jit.py::TestUnion::test_union_memory_aliasing <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0476320Z test_jit.py::TestUnion::test_union_module_with_union_class_variable <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0476599Z test_jit.py::TestUnion::test_union_module_with_union_instance_variable <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0476868Z test_jit.py::TestUnion::test_union_optional_of_union_is_flattened <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0477140Z test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0477438Z test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_container <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0477714Z test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_optional <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0478008Z test_jit.py::TestUnion::test_union_redundant_arguments_are_skipped_subtyping <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0478239Z test_jit.py::TestUnion::test_union_return_type <- test/jit/test_union.py PASSED [ 45%] 2023-03-31T05:50:36.0478517Z test_jit.py::TestUnion::test_union_schema_matching_on_internal_type <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0478806Z test_jit.py::TestUnion::test_union_serialization_preserves_type_annotations <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0479060Z test_jit.py::TestUnion::test_union_subclasses_larger_union <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0479313Z test_jit.py::TestUnion::test_union_subtractive_refinement <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0479596Z test_jit.py::TestUnion::test_union_subtractive_refinement_with_container <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0479836Z test_jit.py::TestUnion::test_union_type_refinement <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0480108Z test_jit.py::TestUnion::test_union_type_refinement_internal_declaration <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0480418Z test_jit.py::TestUnion::test_union_type_refinement_partial_static_refinement_tuple_rhs <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0480723Z test_jit.py::TestUnion::test_union_type_refinement_partial_static_refinement_union_rhs <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0480993Z test_jit.py::TestUnion::test_union_type_refinement_statically_false <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0481261Z test_jit.py::TestUnion::test_union_type_refinement_statically_true <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0481521Z test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0481813Z test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs_noncontained_type <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0482086Z test_jit.py::TestUnion::test_union_type_refinement_tuple_rhs_union <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0482346Z test_jit.py::TestUnion::test_union_type_refinement_union_rhs <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0482595Z test_jit.py::TestUnion::test_union_variable_can_be_reassigned <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0482834Z test_jit.py::TestUnion::test_union_with_collections <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0483231Z test_jit.py::TestUnion::test_union_with_dict_assignment <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0483460Z test_jit.py::TestUnion::test_union_with_enum <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0484151Z test_jit.py::TestUnion::test_union_with_list_assignment <- test/jit/test_union.py [W ir_emitter.cpp:4434] Warning: List consists of heterogeneous types, which means that it has been typed as containing Union[Tensor, int]. To use any of the values in this List, it will be necessary to add an `assert isinstance` statement before first use to trigger type refinement. 2023-03-31T05:50:36.0484327Z File "", line 3 2023-03-31T05:50:36.0484337Z 2023-03-31T05:50:36.0484404Z def fn(): 2023-03-31T05:50:36.0484642Z x: Union[List[str], List[torch.Tensor]] = [torch.add(1, x) for x in [torch.arange(5), 1]] 2023-03-31T05:50:36.0484843Z ~~~~~~~~~~~~~~~~~~~ <--- HERE 2023-03-31T05:50:36.0484958Z if torch.jit.isinstance(x, List[torch.Tensor]): 2023-03-31T05:50:36.0485052Z x.append(torch.tensor(3)) 2023-03-31T05:50:36.0485144Z (function emitListLiteral) 2023-03-31T05:50:36.0485510Z [W ir_emitter.cpp:4434] Warning: List consists of heterogeneous types, which means that it has been typed as containing Union[Tensor, int]. To use any of the values in this List, it will be necessary to add an `assert isinstance` statement before first use to trigger type refinement. 2023-03-31T05:50:36.0485596Z File "", line 3 2023-03-31T05:50:36.0485602Z 2023-03-31T05:50:36.0485669Z def fn(): 2023-03-31T05:50:36.0485824Z x: Union[List[torch.Tensor], int] = [torch.add(1, x) for x in [torch.arange(5), 1]] 2023-03-31T05:50:36.0486019Z ~~~~~~~~~~~~~~~~~~~ <--- HERE 2023-03-31T05:50:36.0486144Z if torch.jit.isinstance(x, List[torch.Tensor]): 2023-03-31T05:50:36.0486222Z x.append(torch.tensor(3)) 2023-03-31T05:50:36.0486313Z (function emitListLiteral) 2023-03-31T05:50:36.0486387Z PASSED [ 46%] 2023-03-31T05:50:36.0486636Z test_jit.py::TestUnion::test_union_with_scalar_values <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0486901Z test_jit.py::TestUnion::test_unions_of_a_single_argument_vanish <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0487157Z test_jit.py::TestUnion::test_unions_of_unions_are_flattened <- test/jit/test_union.py PASSED [ 46%] 2023-03-31T05:50:36.0487409Z test_jit.py::TestBatchMM::test_batch_mm_no_mutation <- test/jit/test_batch_mm.py PASSED [ 46%] 2023-03-31T05:50:36.0487661Z test_jit.py::TestBatchMM::test_batch_mm_permitted_mutation <- test/jit/test_batch_mm.py PASSED [ 46%] 2023-03-31T05:50:36.0487929Z test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation <- test/jit/test_batch_mm.py PASSED [ 46%] 2023-03-31T05:50:36.0488208Z test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation_if_node <- test/jit/test_batch_mm.py PASSED [ 47%] 2023-03-31T05:50:36.0488500Z test_jit.py::TestBatchMM::test_batch_mm_prohibited_mutation_multiple_adds <- test/jit/test_batch_mm.py PASSED [ 47%] 2023-03-31T05:50:36.0488772Z test_jit.py::TestBatchMM::test_batch_mm_side_permitted_mutation <- test/jit/test_batch_mm.py PASSED [ 47%] 2023-03-31T05:50:36.0489067Z test_jit.py::TestBatchMM::test_batch_mm_side_prohibited_mutation_common_side <- test/jit/test_batch_mm.py PASSED [ 47%] 2023-03-31T05:50:36.0489364Z test_jit.py::TestBatchMM::test_batch_mm_side_prohibited_mutation_uncommon_side <- test/jit/test_batch_mm.py PASSED [ 47%] 2023-03-31T05:50:36.0489633Z test_jit.py::TestDtypeAnalysis::test_binary_scalar <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0489906Z test_jit.py::TestDtypeAnalysis::test_binary_tensors <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0490161Z test_jit.py::TestDtypeAnalysis::test_combined <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0490418Z test_jit.py::TestDtypeAnalysis::test_conv_no_mixed_args <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0490678Z test_jit.py::TestDtypeAnalysis::test_custom_rules <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0490925Z test_jit.py::TestDtypeAnalysis::test_unary <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0491317Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_expected_failure_nn_functional_adaptive_max_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py XFAIL [ 47%] 2023-03-31T05:50:36.0491764Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_expected_failure_nn_functional_adaptive_max_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py XFAIL [ 47%] 2023-03-31T05:50:36.0492180Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0492549Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0492913Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0493282Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0493648Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0494007Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0494368Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float16 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0494729Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0495089Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_avg_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0495457Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0495821Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0496186Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 47%] 2023-03-31T05:50:36.0496549Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0496906Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0497266Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_adaptive_max_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0497614Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0497966Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0498303Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0498646Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool1d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0498996Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0499345Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0499767Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0500271Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0500618Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0500956Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0501298Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_avg_pool3d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0501650Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0501997Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0502333Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_batch_norm_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0502676Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0503109Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_complex128 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0503460Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_complex64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0503801Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0504144Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0504488Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv1d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0504833Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0505178Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_complex128 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0505528Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_complex64 <- test/jit/test_dtype_analysis.py PASSED [ 48%] 2023-03-31T05:50:36.0505869Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0506202Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0506541Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0506901Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0507268Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_complex128 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0507631Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_complex64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0507989Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0508406Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0508824Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_conv_transpose2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0509174Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0509519Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0509869Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_hardswish_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0510202Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0510554Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0510904Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0511250Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0511593Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0511935Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0512274Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0512619Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_max_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0512960Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0513301Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0513638Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_ints_nn_functional_prelu_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0513984Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0514346Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 49%] 2023-03-31T05:50:36.0514706Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0515064Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0515417Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0515774Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0516126Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0516544Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0516951Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_avg_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0517307Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0517659Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0518001Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0518356Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0518710Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0519066Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_adaptive_max_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0519405Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0519754Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0520090Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0520425Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool1d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0520771Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0521114Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0521447Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0521769Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0522102Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0522438Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0522784Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_avg_pool3d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0523258Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0523594Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 50%] 2023-03-31T05:50:36.0523930Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_batch_norm_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0524264Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0524602Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_complex128 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0525064Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_complex64 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0525460Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0525778Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0526114Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv1d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0526442Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0526780Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_complex128 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0527121Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_complex64 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0527460Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0527790Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0528118Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0528469Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0528819Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_complex128 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0529172Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_complex64 <- test/jit/test_dtype_analysis.py XFAIL [ 51%] 2023-03-31T05:50:36.0529515Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0529864Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0530209Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_conv_transpose2d_cpu_int64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0530550Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0530888Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0531237Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_hardswish_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0531577Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0531910Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0532246Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool1d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0532589Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 51%] 2023-03-31T05:50:36.0532920Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0533341Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool2d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0533674Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool3d_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0534009Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_max_pool3d_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0534346Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_bfloat16 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0534679Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_float32 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0535010Z test_jit.py::TestDtypeCustomRulesCPU::test_custom_rules_nn_functional_prelu_cpu_float64 <- test/jit/test_dtype_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0535292Z test_jit.py::TestDeviceAnalysis::test_custom_device_op <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0535568Z test_jit.py::TestDeviceAnalysis::test_device_apply <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0535831Z test_jit.py::TestDeviceAnalysis::test_device_arg <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0536116Z test_jit.py::TestDeviceAnalysis::test_device_if_propagation <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0536364Z test_jit.py::TestDeviceAnalysis::test_if_loop_mix <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0536642Z test_jit.py::TestDeviceAnalysis::test_loop_device_change <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0536905Z test_jit.py::TestDeviceAnalysis::test_loop_simple <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0537169Z test_jit.py::TestDeviceAnalysis::test_mobilenet <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0537437Z test_jit.py::TestDeviceAnalysis::test_nested_loops <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0537699Z test_jit.py::TestDeviceAnalysis::test_set_dtype <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0537954Z test_jit.py::TestDeviceAnalysis::test_simple <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0538220Z test_jit.py::TestDeviceAnalysis::test_tensor_as_fns <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0538474Z test_jit.py::TestDeviceAnalysis::test_while_change <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0538736Z test_jit.py::TestDeviceAnalysis::test_zerodim_cpu <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0539017Z test_jit.py::TestDeviceAnalysis::test_zerodim_gpu <- test/jit/test_device_analysis.py SKIPPED (No CUDA) [ 52%] 2023-03-31T05:50:36.0539299Z test_jit.py::TestDeviceAnalysis::test_zerodim_no_device <- test/jit/test_device_analysis.py PASSED [ 52%] 2023-03-31T05:50:36.0539527Z test_jit.py::TestDCE::test_setattr_no_aliasdb <- test/jit/test_dce.py PASSED [ 52%] 2023-03-31T05:50:36.0539753Z test_jit.py::TestDCE::test_setattr_removed <- test/jit/test_dce.py PASSED [ 52%] 2023-03-31T05:50:36.0539991Z test_jit.py::TestSparse::test_freeze_sparse_coo <- test/jit/test_sparse.py PASSED [ 53%] 2023-03-31T05:50:36.0540229Z test_jit.py::TestSparse::test_freeze_sparse_csr <- test/jit/test_sparse.py PASSED [ 53%] 2023-03-31T05:50:36.0540470Z test_jit.py::TestSparse::test_serialize_sparse_coo <- test/jit/test_sparse.py PASSED [ 53%] 2023-03-31T05:50:36.0540700Z test_jit.py::TestSparse::test_serialize_sparse_csr <- test/jit/test_sparse.py PASSED [ 53%] 2023-03-31T05:50:36.0540951Z test_jit.py::TestTensorMethods::test_getitem <- test/jit/test_tensor_methods.py PASSED [ 53%] 2023-03-31T05:50:36.0541218Z test_jit.py::TestTensorMethods::test_getitem_invalid <- test/jit/test_tensor_methods.py PASSED [ 53%] 2023-03-31T05:50:36.0541525Z test_jit.py::TestDataclasses::test__post_init__ <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0541834Z test_jit.py::TestDataclasses::test_comparators <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0542084Z test_jit.py::TestDataclasses::test_custom__eq__ <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0542348Z test_jit.py::TestDataclasses::test_default_factories <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0542591Z test_jit.py::TestDataclasses::test_init_vars <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0542835Z test_jit.py::TestDataclasses::test_no_source <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0543193Z test_jit.py::TestDataclasses::test_use_unregistered_dataclass_raises <- test/jit/test_dataclasses.py PASSED [ 53%] 2023-03-31T05:50:36.0543917Z test_jit.py::TestJitProfiler::test_profiler SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/65521 for allplatform(s) . If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 53%] 2023-03-31T05:50:36.0544063Z test_jit.py::TestJit::test_ModuleList PASSED [ 53%] 2023-03-31T05:50:36.0544200Z test_jit.py::TestJit::test_Sequential PASSED [ 53%] 2023-03-31T05:50:36.0544334Z test_jit.py::TestJit::test_T_mT_H_mH PASSED [ 53%] 2023-03-31T05:50:36.0544472Z test_jit.py::TestJit::test_add_relu_fusion PASSED [ 53%] 2023-03-31T05:50:36.0544659Z test_jit.py::TestJit::test_arg_configurations SKIPPED (Need to be adjusted to Graph Executor) [ 53%] 2023-03-31T05:50:36.0544789Z test_jit.py::TestJit::test_attrs PASSED [ 53%] 2023-03-31T05:50:36.0544920Z test_jit.py::TestJit::test_batchnorm PASSED [ 53%] 2023-03-31T05:50:36.0545058Z test_jit.py::TestJit::test_big SKIPPED (Requires a lot of RAM) [ 53%] 2023-03-31T05:50:36.0545197Z test_jit.py::TestJit::test_conj_transpose PASSED [ 53%] 2023-03-31T05:50:36.0545335Z test_jit.py::TestJit::test_constant_insertion PASSED [ 53%] 2023-03-31T05:50:36.0545479Z test_jit.py::TestJit::test_constant_prop_aliasing_type PASSED [ 53%] 2023-03-31T05:50:36.0545622Z test_jit.py::TestJit::test_constant_prop_exception PASSED [ 53%] 2023-03-31T05:50:36.0545763Z test_jit.py::TestJit::test_constant_prop_if_constant PASSED [ 54%] 2023-03-31T05:50:36.0545899Z test_jit.py::TestJit::test_constant_prop_if_inline PASSED [ 54%] 2023-03-31T05:50:36.0557567Z test_jit.py::TestJit::test_constant_prop_loop_constant PASSED [ 54%] 2023-03-31T05:50:36.0557800Z test_jit.py::TestJit::test_constant_prop_nested PASSED [ 54%] 2023-03-31T05:50:36.0557959Z test_jit.py::TestJit::test_constant_prop_none PASSED [ 54%] 2023-03-31T05:50:36.0558092Z test_jit.py::TestJit::test_constant_prop_print PASSED [ 54%] 2023-03-31T05:50:36.0558237Z test_jit.py::TestJit::test_constant_prop_rand PASSED [ 54%] 2023-03-31T05:50:36.0558382Z test_jit.py::TestJit::test_constant_prop_remove_output PASSED [ 54%] 2023-03-31T05:50:36.0558520Z test_jit.py::TestJit::test_constant_prop_simple PASSED [ 54%] 2023-03-31T05:50:36.0558656Z test_jit.py::TestJit::test_constants_pkl PASSED [ 54%] 2023-03-31T05:50:36.0558781Z test_jit.py::TestJit::test_cpp PASSED [ 54%] 2023-03-31T05:50:36.0558905Z test_jit.py::TestJit::test_cse PASSED [ 54%] 2023-03-31T05:50:36.0559047Z test_jit.py::TestJit::test_cse_not_introduce_aliasing PASSED [ 54%] 2023-03-31T05:50:36.0559318Z test_jit.py::TestJit::test_cu_escaped_number PASSED [ 54%] 2023-03-31T05:50:36.0559469Z test_jit.py::TestJit::test_cuda_export_restore SKIPPED (requires CUDA) [ 54%] 2023-03-31T05:50:36.0559690Z test_jit.py::TestJit::test_debug_flush_compilation_cache PASSED [ 54%] 2023-03-31T05:50:36.0559829Z test_jit.py::TestJit::test_decompose_addmm PASSED [ 54%] 2023-03-31T05:50:36.0559975Z test_jit.py::TestJit::test_device_not_equal SKIPPED (requires CUDA) [ 54%] 2023-03-31T05:50:36.0560120Z test_jit.py::TestJit::test_diff_subgraph_clones_constants PASSED [ 54%] 2023-03-31T05:50:36.0560252Z test_jit.py::TestJit::test_disabled PASSED [ 54%] 2023-03-31T05:50:36.0560377Z test_jit.py::TestJit::test_dropout PASSED [ 54%] 2023-03-31T05:50:36.0560532Z test_jit.py::TestJit::test_dropout_cuda SKIPPED (test_dropout_cuda require CUDA) [ 54%] 2023-03-31T05:50:36.0560948Z test_jit.py::TestJit::test_dropout_func_requires_grad STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0561224Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0561506Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0561766Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0562025Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0562293Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0562545Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0562800Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0563304Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0563564Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:50:36.0563821Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:50:36.0564090Z STAGE:2023-03-31 05:49:42 2274:2274 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:50:36.0564170Z PASSED [ 54%] 2023-03-31T05:50:36.0564363Z test_jit.py::TestJit::test_dropout_module_requires_grad SKIPPED (Testing differentiable graph) [ 54%] 2023-03-31T05:50:36.0564494Z test_jit.py::TestJit::test_einsum PASSED [ 54%] 2023-03-31T05:50:36.0564628Z test_jit.py::TestJit::test_element_size PASSED [ 55%] 2023-03-31T05:50:36.0564759Z test_jit.py::TestJit::test_expand_fold_quant_inputs PASSED [ 55%] 2023-03-31T05:50:36.0564900Z test_jit.py::TestJit::test_expand_quantlint PASSED [ 55%] 2023-03-31T05:50:36.0565108Z test_jit.py::TestJit::test_export_batchnorm SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 55%] 2023-03-31T05:50:36.0565245Z test_jit.py::TestJit::test_export_dropout PASSED [ 55%] 2023-03-31T05:50:36.0565377Z test_jit.py::TestJit::test_export_lstm PASSED [ 55%] 2023-03-31T05:50:36.0565512Z test_jit.py::TestJit::test_export_opnames PASSED [ 55%] 2023-03-31T05:50:36.0565638Z test_jit.py::TestJit::test_export_rnn PASSED [ 55%] 2023-03-31T05:50:36.0565815Z test_jit.py::TestJit::test_flags SKIPPED (Need to instrument GraphExecutors a bit more) [ 55%] 2023-03-31T05:50:36.0565944Z test_jit.py::TestJit::test_function_default_values PASSED [ 55%] 2023-03-31T05:50:36.0566094Z test_jit.py::TestJit::test_hide_source_ranges_context_manager PASSED [ 55%] 2023-03-31T05:50:36.0566328Z test_jit.py::TestJit::test_import_method PASSED [ 55%] 2023-03-31T05:50:36.0566468Z test_jit.py::TestJit::test_inferred_as_tensor PASSED [ 55%] 2023-03-31T05:50:36.0566668Z test_jit.py::TestJit::test_layout PASSED [ 55%] 2023-03-31T05:50:36.0566806Z test_jit.py::TestJit::test_matrix_conj_transpose PASSED [ 55%] 2023-03-31T05:50:36.0566942Z test_jit.py::TestJit::test_matrix_transpose PASSED [ 55%] 2023-03-31T05:50:36.0567078Z test_jit.py::TestJit::test_module_default_values PASSED [ 55%] 2023-03-31T05:50:36.0567205Z test_jit.py::TestJit::test_mutable_default_values PASSED [ 55%] 2023-03-31T05:50:36.0567370Z test_jit.py::TestJit::test_native_dropout_corner_case SKIPPED (test requires CUDA) [ 55%] 2023-03-31T05:50:36.0567496Z test_jit.py::TestJit::test_nn_conv PASSED [ 55%] 2023-03-31T05:50:36.0567631Z test_jit.py::TestJit::test_nn_lp_pool1d PASSED [ 55%] 2023-03-31T05:50:36.0567762Z test_jit.py::TestJit::test_nn_lp_pool2d PASSED [ 55%] 2023-03-31T05:50:36.0567898Z test_jit.py::TestJit::test_nn_padding PASSED [ 55%] 2023-03-31T05:50:36.0568036Z test_jit.py::TestJit::test_nn_padding_functional PASSED [ 55%] 2023-03-31T05:50:36.0568172Z test_jit.py::TestJit::test_no_erroneous_warnings PASSED [ 55%] 2023-03-31T05:50:36.0568370Z test_jit.py::TestJit::test_non_ascii_string SKIPPED (temporarily disable the test for fwd compatibility) [ 55%] 2023-03-31T05:50:36.0568483Z test_jit.py::TestJit::test_numel PASSED [ 56%] 2023-03-31T05:50:36.0568624Z test_jit.py::TestJit::test_pattern_based_module_rewrite PASSED [ 56%] 2023-03-31T05:50:36.0568759Z test_jit.py::TestJit::test_pattern_based_rewrite PASSED [ 56%] 2023-03-31T05:50:36.0568933Z test_jit.py::TestJit::test_pattern_based_rewrite_with_source_range_preserved PASSED [ 56%] 2023-03-31T05:50:36.0569238Z test_jit.py::TestJit::test_peephole_optimize_shape_ops SKIPPED (Simple executor doesn't have shape information) [ 56%] 2023-03-31T05:50:36.0569385Z test_jit.py::TestJit::test_permute_inputs_binding PASSED [ 56%] 2023-03-31T05:50:36.0569520Z test_jit.py::TestJit::test_pretty_printer PASSED [ 56%] 2023-03-31T05:50:36.0569660Z test_jit.py::TestJit::test_print_classes_module PASSED [ 56%] 2023-03-31T05:50:36.0569780Z test_jit.py::TestJit::test_print_op_module PASSED [ 56%] 2023-03-31T05:50:36.0569917Z test_jit.py::TestJit::test_print_torch_ops_modules PASSED [ 56%] 2023-03-31T05:50:36.0570051Z test_jit.py::TestJit::test_python_bindings PASSED [ 56%] 2023-03-31T05:50:36.0570179Z test_jit.py::TestJit::test_python_ir PASSED [ 56%] 2023-03-31T05:50:36.0570315Z test_jit.py::TestJit::test_python_ir_utils PASSED [ 56%] 2023-03-31T05:50:36.0570450Z test_jit.py::TestJit::test_python_ir_utils_graph PASSED [ 56%] 2023-03-31T05:50:36.0570586Z test_jit.py::TestJit::test_python_ivalue PASSED [ 56%] 2023-03-31T05:50:36.0570720Z test_jit.py::TestJit::test_pytorch_jit_env_off PASSED [ 56%] 2023-03-31T05:50:36.0570839Z test_jit.py::TestJit::test_recursive_cse PASSED [ 56%] 2023-03-31T05:50:36.0570980Z test_jit.py::TestJit::test_repeat_interleave_script PASSED [ 56%] 2023-03-31T05:50:36.0571110Z test_jit.py::TestJit::test_restore_device PASSED [ 56%] 2023-03-31T05:50:36.0571280Z test_jit.py::TestJit::test_restore_device_cuda SKIPPED (restore device requires CUDA) [ 56%] 2023-03-31T05:50:36.0571466Z test_jit.py::TestJit::test_restore_shared_storage_on_cuda SKIPPED (restore device requires CUDA) [ 56%] 2023-03-31T05:50:36.0571661Z test_jit.py::TestJit::test_script_autograd_grad PASSED [ 56%] 2023-03-31T05:50:36.0571793Z test_jit.py::TestJit::test_script_backward PASSED [ 56%] 2023-03-31T05:50:36.0571992Z test_jit.py::TestJit::test_script_backward_twice PASSED [ 56%] 2023-03-31T05:50:36.0572112Z test_jit.py::TestJit::test_script_fn_pkl PASSED [ 56%] 2023-03-31T05:50:36.0572246Z test_jit.py::TestJit::test_script_tensor_type PASSED [ 56%] 2023-03-31T05:50:36.0572385Z test_jit.py::TestJit::test_shape_analysis_broadcast PASSED [ 57%] 2023-03-31T05:50:36.0572524Z test_jit.py::TestJit::test_shape_analysis_masked_select PASSED [ 57%] 2023-03-31T05:50:36.0572668Z test_jit.py::TestJit::test_shape_analysis_unsqueeze_in_loop PASSED [ 57%] 2023-03-31T05:50:36.0572802Z test_jit.py::TestJit::test_sparse_csr_tensors PASSED [ 57%] 2023-03-31T05:50:36.0572935Z test_jit.py::TestJit::test_sparse_tensors PASSED [ 57%] 2023-03-31T05:50:36.0573071Z test_jit.py::TestJit::test_torch_complex PASSED [ 57%] 2023-03-31T05:50:36.0573410Z test_jit.py::TestJit::test_torch_load_error SKIPPED (TODO: re-enable with https://github.com/pytorch/pytorch/pull/29339) [ 57%] 2023-03-31T05:50:36.0573541Z test_jit.py::TestJit::test_torch_load_zipfile_check PASSED [ 57%] 2023-03-31T05:50:36.0573673Z test_jit.py::TestJit::test_torch_ops_kwonly PASSED [ 57%] 2023-03-31T05:50:36.0573813Z test_jit.py::TestJit::test_torch_ops_overloaded PASSED [ 57%] 2023-03-31T05:50:36.0573940Z test_jit.py::TestJit::test_torch_sum PASSED [ 57%] 2023-03-31T05:50:36.0574074Z test_jit.py::TestJit::test_trace_retains_train PASSED [ 57%] 2023-03-31T05:50:36.0574201Z test_jit.py::TestJit::test_train_eval PASSED [ 57%] 2023-03-31T05:50:36.0574329Z test_jit.py::TestJit::test_transpose PASSED [ 57%] 2023-03-31T05:50:36.0574466Z test_jit.py::TestJit::test_unchecked_cast PASSED [ 57%] 2023-03-31T05:50:36.0574591Z test_jit.py::TestJit::test_unique_state_dict PASSED [ 57%] 2023-03-31T05:50:36.0574783Z test_jit.py::TestJit::test_verify SKIPPED (verify needs to be updated to work with GraphExecutors) [ 57%] 2023-03-31T05:50:36.0574910Z test_jit.py::TestJit::test_warnings PASSED [ 57%] 2023-03-31T05:50:36.0575081Z test_jit.py::TestFrontend::test_dictionary_as_example_inputs_for_jit_trace PASSED [ 57%] 2023-03-31T05:50:36.0575223Z test_jit.py::TestFrontend::test_instancing_error PASSED [ 57%] 2023-03-31T05:50:36.0575353Z test_jit.py::TestScript::test_add_out PASSED [ 57%] 2023-03-31T05:50:36.0575497Z test_jit.py::TestScript::test_add_tuple_different_types PASSED [ 57%] 2023-03-31T05:50:36.0575637Z test_jit.py::TestScript::test_add_tuple_non_optional PASSED [ 57%] 2023-03-31T05:50:36.0575771Z test_jit.py::TestScript::test_add_tuple_optional PASSED [ 57%] 2023-03-31T05:50:36.0575914Z test_jit.py::TestScript::test_add_tuple_same_types PASSED [ 57%] 2023-03-31T05:50:36.0576045Z test_jit.py::TestScript::test_addmm_grad PASSED [ 58%] 2023-03-31T05:50:36.0576198Z test_jit.py::TestScript::test_alias_covariant_type_containers PASSED [ 58%] 2023-03-31T05:50:36.0576323Z test_jit.py::TestScript::test_all PASSED [ 58%] 2023-03-31T05:50:36.0576459Z test_jit.py::TestScript::test_annot_ast_mypy_fn PASSED [ 58%] 2023-03-31T05:50:36.0576596Z test_jit.py::TestScript::test_annot_ast_mypy_method PASSED [ 58%] 2023-03-31T05:50:36.0576731Z test_jit.py::TestScript::test_annot_ast_py3_fn PASSED [ 58%] 2023-03-31T05:50:36.0576856Z test_jit.py::TestScript::test_annot_ast_py3_method PASSED [ 58%] 2023-03-31T05:50:36.0577064Z test_jit.py::TestScript::test_annot_string_mypy_fn PASSED [ 58%] 2023-03-31T05:50:36.0577204Z test_jit.py::TestScript::test_annot_string_mypy_method PASSED [ 58%] 2023-03-31T05:50:36.0577402Z test_jit.py::TestScript::test_annot_string_py3_fn PASSED [ 58%] 2023-03-31T05:50:36.0577544Z test_jit.py::TestScript::test_annot_string_py3_method PASSED [ 58%] 2023-03-31T05:50:36.0577682Z test_jit.py::TestScript::test_annotated_script_fn PASSED [ 58%] 2023-03-31T05:50:36.0577827Z test_jit.py::TestScript::test_annotated_script_fn_arg_mismatch PASSED [ 58%] 2023-03-31T05:50:36.0577977Z test_jit.py::TestScript::test_annotated_script_fn_return_mismatch PASSED [ 58%] 2023-03-31T05:50:36.0578106Z test_jit.py::TestScript::test_annotated_script_method PASSED [ 58%] 2023-03-31T05:50:36.0578244Z test_jit.py::TestScript::test_annoying_doubles PASSED [ 58%] 2023-03-31T05:50:36.0578374Z test_jit.py::TestScript::test_any PASSED [ 58%] 2023-03-31T05:50:36.0578522Z test_jit.py::TestScript::test_assert_is_scripting_metacompile PASSED [ 58%] 2023-03-31T05:50:36.0578674Z test_jit.py::TestScript::test_assertion_optional_refinement PASSED [ 58%] 2023-03-31T05:50:36.0578811Z test_jit.py::TestScript::test_attr_module_constants PASSED [ 58%] 2023-03-31T05:50:36.0578948Z test_jit.py::TestScript::test_attr_qscheme_script PASSED [ 58%] 2023-03-31T05:50:36.0579083Z test_jit.py::TestScript::test_attribute_in_init PASSED [ 58%] 2023-03-31T05:50:36.0579216Z test_jit.py::TestScript::test_attribute_serialization PASSED [ 58%] 2023-03-31T05:50:36.0579357Z test_jit.py::TestScript::test_attribute_unpickling PASSED [ 58%] 2023-03-31T05:50:36.0579491Z test_jit.py::TestScript::test_augmented_assign PASSED [ 58%] 2023-03-31T05:50:36.0579635Z test_jit.py::TestScript::test_autodiff_complex SKIPPED (no CUDA) [ 58%] 2023-03-31T05:50:36.0579779Z test_jit.py::TestScript::test_backend_cudnn_enabled PASSED [ 59%] 2023-03-31T05:50:36.0579923Z test_jit.py::TestScript::test_bad_multiline_annotations PASSED [ 59%] 2023-03-31T05:50:36.0580086Z test_jit.py::TestScript::test_bailout_loop_carried_deps_name_clash PASSED [ 59%] 2023-03-31T05:50:36.0580237Z test_jit.py::TestScript::test_bailout_loop_counter_transition PASSED [ 59%] 2023-03-31T05:50:36.0580443Z test_jit.py::TestScript::test_batch_norm_inference_backward_cuda SKIPPED (running tests on cuda to verify cudnn fix) [ 59%] 2023-03-31T05:50:36.0580806Z test_jit.py::TestScript::test_batchnorm_fuser_cpu <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/jit_utils.py PASSED [ 59%] 2023-03-31T05:50:36.0580947Z test_jit.py::TestScript::test_big_float_literals PASSED [ 59%] 2023-03-31T05:50:36.0581083Z test_jit.py::TestScript::test_big_int_literals PASSED [ 59%] 2023-03-31T05:50:36.0581224Z test_jit.py::TestScript::test_binary_op_shape PASSED [ 59%] 2023-03-31T05:50:36.0581356Z test_jit.py::TestScript::test_bitwise_ops PASSED [ 59%] 2023-03-31T05:50:36.0581504Z test_jit.py::TestScript::test_block_input_grad_in_loop PASSED [ 59%] 2023-03-31T05:50:36.0581649Z test_jit.py::TestScript::test_bool_augassign_bitwise_and PASSED [ 59%] 2023-03-31T05:50:36.0581791Z test_jit.py::TestScript::test_bool_augassign_bitwise_or PASSED [ 59%] 2023-03-31T05:50:36.0581923Z test_jit.py::TestScript::test_bool_augassign_bitwise_xor PASSED [ 59%] 2023-03-31T05:50:36.0582060Z test_jit.py::TestScript::test_bool_dispatch PASSED [ 59%] 2023-03-31T05:50:36.0582218Z test_jit.py::TestScript::test_boolean_literal_constant_metacompile PASSED [ 59%] 2023-03-31T05:50:36.0582358Z test_jit.py::TestScript::test_break_continue_error PASSED [ 59%] 2023-03-31T05:50:36.0582553Z test_jit.py::TestScript::test_breaks_continues PASSED [ 59%] 2023-03-31T05:50:36.0582687Z test_jit.py::TestScript::test_builtin_args PASSED [ 59%] 2023-03-31T05:50:36.0583291Z test_jit.py::TestScript::test_builtin_args_fails [W ir_emitter.cpp:4434] Warning: List consists of heterogeneous types, which means that it has been typed as containing Union[List[int], int]. To use any of the values in this List, it will be necessary to add an `assert isinstance` statement before first use to trigger type refinement. 2023-03-31T05:50:36.0583389Z File "test_jit.py", line 10743 2023-03-31T05:50:36.0583476Z @torch.jit.script 2023-03-31T05:50:36.0583537Z def f6(a): 2023-03-31T05:50:36.0583627Z a.expand(size=[3, [4]]) 2023-03-31T05:50:36.0583769Z ~~~~~~ <--- HERE 2023-03-31T05:50:36.0583861Z (function emitListLiteral) 2023-03-31T05:50:36.0583941Z PASSED [ 59%] 2023-03-31T05:50:36.0584098Z test_jit.py::TestScript::test_builtin_function_attributes PASSED [ 59%] 2023-03-31T05:50:36.0584239Z test_jit.py::TestScript::test_builtin_use_as_value PASSED [ 59%] 2023-03-31T05:50:36.0584363Z test_jit.py::TestScript::test_call_ge PASSED [ 59%] 2023-03-31T05:50:36.0584509Z test_jit.py::TestScript::test_call_python_fn_from_script_fn PASSED [ 59%] 2023-03-31T05:50:36.0584656Z test_jit.py::TestScript::test_call_python_fn_from_script_module PASSED [ 59%] 2023-03-31T05:50:36.0584804Z test_jit.py::TestScript::test_call_python_fn_from_traced_module PASSED [ 60%] 2023-03-31T05:50:36.0584948Z test_jit.py::TestScript::test_call_python_fn_from_tracing_fn PASSED [ 60%] 2023-03-31T05:50:36.0585093Z test_jit.py::TestScript::test_call_python_mod_from_script_fn PASSED [ 60%] 2023-03-31T05:50:36.0585242Z test_jit.py::TestScript::test_call_python_mod_from_script_module PASSED [ 60%] 2023-03-31T05:50:36.0585388Z test_jit.py::TestScript::test_call_python_mod_from_traced_module PASSED [ 60%] 2023-03-31T05:50:36.0585526Z test_jit.py::TestScript::test_call_python_mod_from_tracing_fn PASSED [ 60%] 2023-03-31T05:50:36.0585671Z test_jit.py::TestScript::test_call_script_fn_from_script_fn PASSED [ 60%] 2023-03-31T05:50:36.0585824Z test_jit.py::TestScript::test_call_script_fn_from_script_module PASSED [ 60%] 2023-03-31T05:50:36.0585969Z test_jit.py::TestScript::test_call_script_fn_from_tracing_fn PASSED [ 60%] 2023-03-31T05:50:36.0586111Z test_jit.py::TestScript::test_call_script_mod_from_script_fn PASSED [ 60%] 2023-03-31T05:50:36.0586259Z test_jit.py::TestScript::test_call_script_mod_from_script_module PASSED [ 60%] 2023-03-31T05:50:36.0586445Z test_jit.py::TestScript::test_call_script_mod_from_tracing_fn SKIPPED (error in first class mode) [ 60%] 2023-03-31T05:50:36.0586590Z test_jit.py::TestScript::test_call_traced_fn_from_tracing_fn PASSED [ 60%] 2023-03-31T05:50:36.0586758Z test_jit.py::TestScript::test_call_traced_mod_from_tracing_fn SKIPPED (error in first class mode) [ 60%] 2023-03-31T05:50:36.0586907Z test_jit.py::TestScript::test_calls_in_type_annotations PASSED [ 60%] 2023-03-31T05:50:36.0587055Z test_jit.py::TestScript::test_canonicalize_control_outputs PASSED [ 60%] 2023-03-31T05:50:36.0587248Z test_jit.py::TestScript::test_cast SKIPPED (RuntimeError: VariableType::ID() not implemented) [ 60%] 2023-03-31T05:50:36.0587376Z test_jit.py::TestScript::test_cat PASSED [ 60%] 2023-03-31T05:50:36.0587507Z test_jit.py::TestScript::test_cat_lifts PASSED [ 60%] 2023-03-31T05:50:36.0587635Z test_jit.py::TestScript::test_chr PASSED [ 60%] 2023-03-31T05:50:36.0587778Z test_jit.py::TestScript::test_circular_dependency PASSED [ 60%] 2023-03-31T05:50:36.0587915Z test_jit.py::TestScript::test_class_as_attribute PASSED [ 60%] 2023-03-31T05:50:36.0588038Z test_jit.py::TestScript::test_class_attribute PASSED [ 60%] 2023-03-31T05:50:36.0588278Z test_jit.py::TestScript::test_class_attribute_in_script PASSED [ 60%] 2023-03-31T05:50:36.0588440Z test_jit.py::TestScript::test_class_with_comment_at_lower_indentation PASSED [ 60%] 2023-03-31T05:50:36.0588632Z test_jit.py::TestScript::test_code_with_constants PASSED [ 60%] 2023-03-31T05:50:36.0588778Z test_jit.py::TestScript::test_code_with_constants_restore PASSED [ 61%] 2023-03-31T05:50:36.0588920Z test_jit.py::TestScript::test_comment_ignore_indent PASSED [ 61%] 2023-03-31T05:50:36.0589062Z test_jit.py::TestScript::test_compare_two_bool_inputs PASSED [ 61%] 2023-03-31T05:50:36.0589289Z test_jit.py::TestScript::test_compile_module_with_constant SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 61%] 2023-03-31T05:50:36.0589417Z test_jit.py::TestScript::test_conditional_casting PASSED [ 61%] 2023-03-31T05:50:36.0589553Z test_jit.py::TestScript::test_constant_as_attr PASSED [ 61%] 2023-03-31T05:50:36.0589711Z test_jit.py::TestScript::test_constant_pooling_introduce_aliasing PASSED [ 61%] 2023-03-31T05:50:36.0589852Z test_jit.py::TestScript::test_constant_pooling_none PASSED [ 61%] 2023-03-31T05:50:36.0590006Z test_jit.py::TestScript::test_constant_pooling_same_identity PASSED [ 61%] 2023-03-31T05:50:36.0590141Z test_jit.py::TestScript::test_context_manager PASSED [ 61%] 2023-03-31T05:50:36.0590274Z test_jit.py::TestScript::test_conv_error PASSED [ 61%] 2023-03-31T05:50:36.0590408Z test_jit.py::TestScript::test_convert_base PASSED [ 61%] 2023-03-31T05:50:36.0590537Z test_jit.py::TestScript::test_cpp_function_tensor_str PASSED [ 61%] 2023-03-31T05:50:36.0590675Z test_jit.py::TestScript::test_cpp_module_iterator PASSED [ 61%] 2023-03-31T05:50:36.0590809Z test_jit.py::TestScript::test_desugar_module PASSED [ 61%] 2023-03-31T05:50:36.0590951Z test_jit.py::TestScript::test_device_kwarg PASSED [ 61%] 2023-03-31T05:50:36.0591082Z test_jit.py::TestScript::test_device_type PASSED [ 61%] 2023-03-31T05:50:36.0591232Z test_jit.py::TestScript::test_device_type_cuda SKIPPED (Requires CUDA) [ 61%] 2023-03-31T05:50:36.0591362Z test_jit.py::TestScript::test_dict_str PASSED [ 61%] 2023-03-31T05:50:36.0591488Z test_jit.py::TestScript::test_dir PASSED [ 61%] 2023-03-31T05:50:36.0591605Z test_jit.py::TestScript::test_divmod PASSED [ 61%] 2023-03-31T05:50:36.0591782Z test_jit.py::TestScript::test_dominated_bailout SKIPPED (bailouts are being deprecated) [ 61%] 2023-03-31T05:50:36.0591915Z test_jit.py::TestScript::test_dropout_eval PASSED [ 61%] 2023-03-31T05:50:36.0592051Z test_jit.py::TestScript::test_dtype_attr PASSED [ 61%] 2023-03-31T05:50:36.0592190Z test_jit.py::TestScript::test_dtype_op_shape PASSED [ 61%] 2023-03-31T05:50:36.0592327Z test_jit.py::TestScript::test_dtype_op_shape2 PASSED [ 61%] 2023-03-31T05:50:36.0592469Z test_jit.py::TestScript::test_early_return_closure PASSED [ 62%] 2023-03-31T05:50:36.0592611Z test_jit.py::TestScript::test_early_return_fork_join PASSED [ 62%] 2023-03-31T05:50:36.0592737Z test_jit.py::TestScript::test_early_return_rewrite PASSED [ 62%] 2023-03-31T05:50:36.0592882Z test_jit.py::TestScript::test_early_return_type_refinement PASSED [ 62%] 2023-03-31T05:50:36.0593019Z test_jit.py::TestScript::test_early_returns_loops PASSED [ 62%] 2023-03-31T05:50:36.0593158Z test_jit.py::TestScript::test_ellipsis_const_end PASSED [ 62%] 2023-03-31T05:50:36.0593297Z test_jit.py::TestScript::test_ellipsis_const_mid PASSED [ 62%] 2023-03-31T05:50:36.0593438Z test_jit.py::TestScript::test_ellipsis_const_mid_select PASSED [ 62%] 2023-03-31T05:50:36.0593626Z test_jit.py::TestScript::test_ellipsis_const_start PASSED [ 62%] 2023-03-31T05:50:36.0593760Z test_jit.py::TestScript::test_ellipsis_end PASSED [ 62%] 2023-03-31T05:50:36.0593944Z test_jit.py::TestScript::test_ellipsis_mid PASSED [ 62%] 2023-03-31T05:50:36.0594084Z test_jit.py::TestScript::test_ellipsis_mid_select PASSED [ 62%] 2023-03-31T05:50:36.0594218Z test_jit.py::TestScript::test_ellipsis_start PASSED [ 62%] 2023-03-31T05:50:36.0594363Z test_jit.py::TestScript::test_embedding_renorm_grad_error PASSED [ 62%] 2023-03-31T05:50:36.0594504Z test_jit.py::TestScript::test_empty_like_memory_format_bc PASSED [ 62%] 2023-03-31T05:50:36.0594639Z test_jit.py::TestScript::test_empty_tuple_str PASSED [ 62%] 2023-03-31T05:50:36.0594781Z test_jit.py::TestScript::test_enumerate_modlist_range PASSED [ 62%] 2023-03-31T05:50:36.0594923Z test_jit.py::TestScript::test_erase_number_types PASSED [ 62%] 2023-03-31T05:50:36.0595050Z test_jit.py::TestScript::test_error PASSED [ 62%] 2023-03-31T05:50:36.0595179Z test_jit.py::TestScript::test_error_stacktrace PASSED [ 62%] 2023-03-31T05:50:36.0595325Z test_jit.py::TestScript::test_error_stacktrace_interface PASSED [ 62%] 2023-03-31T05:50:36.0595456Z test_jit.py::TestScript::test_eval_python PASSED [ 62%] 2023-03-31T05:50:36.0595596Z test_jit.py::TestScript::test_exception_exits_closure PASSED [ 62%] 2023-03-31T05:50:36.0595741Z test_jit.py::TestScript::test_exceptions_with_control_flow PASSED [ 62%] 2023-03-31T05:50:36.0595870Z test_jit.py::TestScript::test_expand PASSED [ 62%] 2023-03-31T05:50:36.0595997Z test_jit.py::TestScript::test_fibb PASSED [ 63%] 2023-03-31T05:50:36.0596134Z test_jit.py::TestScript::test_fibb_totally_better PASSED [ 63%] 2023-03-31T05:50:36.0596271Z test_jit.py::TestScript::test_file_format_serialization PASSED [ 63%] 2023-03-31T05:50:36.0596409Z test_jit.py::TestScript::test_file_line_error PASSED [ 63%] 2023-03-31T05:50:36.0596549Z test_jit.py::TestScript::test_file_line_error_class_defn PASSED [ 63%] 2023-03-31T05:50:36.0596684Z test_jit.py::TestScript::test_file_line_graph PASSED [ 63%] 2023-03-31T05:50:36.0596820Z test_jit.py::TestScript::test_file_line_save_load PASSED [ 63%] 2023-03-31T05:50:36.0596955Z test_jit.py::TestScript::test_file_line_string PASSED [ 63%] 2023-03-31T05:50:36.0597087Z test_jit.py::TestScript::test_file_line_trace PASSED [ 63%] 2023-03-31T05:50:36.0597217Z test_jit.py::TestScript::test_filecheck PASSED [ 63%] 2023-03-31T05:50:36.0597342Z test_jit.py::TestScript::test_filecheck_parse PASSED [ 63%] 2023-03-31T05:50:36.0597481Z test_jit.py::TestScript::test_first_class_calls PASSED [ 63%] 2023-03-31T05:50:36.0597621Z test_jit.py::TestScript::test_first_class_module PASSED [ 63%] 2023-03-31T05:50:36.0597755Z test_jit.py::TestScript::test_floor_div PASSED [ 63%] 2023-03-31T05:50:36.0597885Z test_jit.py::TestScript::test_floordiv PASSED [ 63%] 2023-03-31T05:50:36.0598014Z test_jit.py::TestScript::test_for_else PASSED [ 63%] 2023-03-31T05:50:36.0598145Z test_jit.py::TestScript::test_for_in_dict PASSED [ 63%] 2023-03-31T05:50:36.0598279Z test_jit.py::TestScript::test_for_in_enumerate PASSED [ 63%] 2023-03-31T05:50:36.0598397Z test_jit.py::TestScript::test_for_in_range PASSED [ 63%] 2023-03-31T05:50:36.0598530Z test_jit.py::TestScript::test_for_in_range_ast PASSED [ 63%] 2023-03-31T05:50:36.0598718Z test_jit.py::TestScript::test_for_in_range_dynamic PASSED [ 63%] 2023-03-31T05:50:36.0598852Z test_jit.py::TestScript::test_for_in_range_if_ast PASSED [ 63%] 2023-03-31T05:50:36.0599044Z test_jit.py::TestScript::test_for_in_range_start_end PASSED [ 63%] 2023-03-31T05:50:36.0599186Z test_jit.py::TestScript::test_for_in_range_start_end_step PASSED [ 63%] 2023-03-31T05:50:36.0599325Z test_jit.py::TestScript::test_for_in_range_zero_step PASSED [ 63%] 2023-03-31T05:50:36.0599456Z test_jit.py::TestScript::test_for_in_string PASSED [ 63%] 2023-03-31T05:50:36.0599576Z test_jit.py::TestScript::test_for_in_tensors PASSED [ 64%] 2023-03-31T05:50:36.0599717Z test_jit.py::TestScript::test_for_in_tensors_fail_scalar PASSED [ 64%] 2023-03-31T05:50:36.0599853Z test_jit.py::TestScript::test_for_in_tensors_nested PASSED [ 64%] 2023-03-31T05:50:36.0599989Z test_jit.py::TestScript::test_for_in_tensors_rank0 PASSED [ 64%] 2023-03-31T05:50:36.0600122Z test_jit.py::TestScript::test_for_in_zip PASSED [ 64%] 2023-03-31T05:50:36.0600261Z test_jit.py::TestScript::test_for_in_zip_enumerate PASSED [ 64%] 2023-03-31T05:50:36.0600399Z test_jit.py::TestScript::test_for_tuple_assign PASSED [ 64%] 2023-03-31T05:50:36.0600533Z test_jit.py::TestScript::test_for_tuple_unpack PASSED [ 64%] 2023-03-31T05:50:36.0600650Z test_jit.py::TestScript::test_format PASSED [ 64%] 2023-03-31T05:50:36.0600779Z test_jit.py::TestScript::test_func_call PASSED [ 64%] 2023-03-31T05:50:36.0600926Z test_jit.py::TestScript::test_function_compilation_caching PASSED [ 64%] 2023-03-31T05:50:36.0601067Z test_jit.py::TestScript::test_function_overload_misuse PASSED [ 64%] 2023-03-31T05:50:36.0601215Z test_jit.py::TestScript::test_function_overloading_isinstance PASSED [ 64%] 2023-03-31T05:50:36.0601358Z test_jit.py::TestScript::test_function_overloads PASSED [ 64%] 2023-03-31T05:50:36.0601865Z test_jit.py::TestScript::test_fuser_double_float_codegen <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/jit_utils.py SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 64%] 2023-03-31T05:50:36.0602259Z test_jit.py::TestScript::test_fuser_double_literal_precision <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/jit_utils.py PASSED [ 64%] 2023-03-31T05:50:36.0602403Z test_jit.py::TestScript::test_fuser_multiple_blocks PASSED [ 64%] 2023-03-31T05:50:36.0602534Z test_jit.py::TestScript::test_gather_dynamic_index PASSED [ 64%] 2023-03-31T05:50:36.0602673Z test_jit.py::TestScript::test_generic_list_errors PASSED [ 64%] 2023-03-31T05:50:36.0602807Z test_jit.py::TestScript::test_get_set_state PASSED [ 64%] 2023-03-31T05:50:36.0602954Z test_jit.py::TestScript::test_get_set_state_with_tensors PASSED [ 64%] 2023-03-31T05:50:36.0603264Z test_jit.py::TestScript::test_grad_from_script PASSED [ 64%] 2023-03-31T05:50:36.0603401Z test_jit.py::TestScript::test_hash PASSED [ 64%] 2023-03-31T05:50:36.0603536Z test_jit.py::TestScript::test_hex_literals PASSED [ 64%] 2023-03-31T05:50:36.0603664Z test_jit.py::TestScript::test_id PASSED [ 64%] 2023-03-31T05:50:36.0603778Z test_jit.py::TestScript::test_if PASSED [ 65%] 2023-03-31T05:50:36.0603911Z test_jit.py::TestScript::test_if_define PASSED [ 65%] 2023-03-31T05:50:36.0604049Z test_jit.py::TestScript::test_if_different_type PASSED [ 65%] 2023-03-31T05:50:36.0604185Z test_jit.py::TestScript::test_if_for_in_range PASSED [ 65%] 2023-03-31T05:50:36.0604323Z test_jit.py::TestScript::test_if_is_none_dispatch PASSED [ 65%] 2023-03-31T05:50:36.0604559Z test_jit.py::TestScript::test_if_list_cat PASSED [ 65%] 2023-03-31T05:50:36.0604767Z test_jit.py::TestScript::test_if_nest_while PASSED [ 65%] 2023-03-31T05:50:36.0604899Z test_jit.py::TestScript::test_if_noelse PASSED [ 65%] 2023-03-31T05:50:36.0605025Z test_jit.py::TestScript::test_if_not_defined_error PASSED [ 65%] 2023-03-31T05:50:36.0605160Z test_jit.py::TestScript::test_if_supertype PASSED [ 65%] 2023-03-31T05:50:36.0605297Z test_jit.py::TestScript::test_ignore_decorator PASSED [ 65%] 2023-03-31T05:50:36.0605433Z test_jit.py::TestScript::test_ignored_as_value PASSED [ 65%] 2023-03-31T05:50:36.0605574Z test_jit.py::TestScript::test_ignored_method_binding PASSED [ 65%] 2023-03-31T05:50:36.0605711Z test_jit.py::TestScript::test_ignored_props PASSED [ 65%] 2023-03-31T05:50:36.0605866Z test_jit.py::TestScript::test_import_constants_not_specialized PASSED [ 65%] 2023-03-31T05:50:36.0606003Z test_jit.py::TestScript::test_in_for_and_comp_expr PASSED [ 65%] 2023-03-31T05:50:36.0606140Z test_jit.py::TestScript::test_in_operator_with_two_strings PASSED [ 65%] 2023-03-31T05:50:36.0606269Z test_jit.py::TestScript::test_index PASSED [ 65%] 2023-03-31T05:50:36.0606413Z test_jit.py::TestScript::test_index_select_shape_prop PASSED [ 65%] 2023-03-31T05:50:36.0606549Z test_jit.py::TestScript::test_index_with_tuple PASSED [ 65%] 2023-03-31T05:50:36.0606684Z test_jit.py::TestScript::test_indexing_error PASSED [ 65%] 2023-03-31T05:50:36.0606816Z test_jit.py::TestScript::test_infer_size PASSED [ 65%] 2023-03-31T05:50:36.0606955Z test_jit.py::TestScript::test_inferred_error_msg PASSED [ 65%] 2023-03-31T05:50:36.0607096Z test_jit.py::TestScript::test_inherit_method PASSED [ 65%] 2023-03-31T05:50:36.0607329Z test_jit.py::TestScript::test_inline_and_run_annotated_script_fn SKIPPED (https://github.com/pytorch/pytorch/issues/9595) [ 65%] 2023-03-31T05:50:36.0607457Z test_jit.py::TestScript::test_inlined_graph PASSED [ 65%] 2023-03-31T05:50:36.0607595Z test_jit.py::TestScript::test_inlining_cleanup PASSED [ 66%] 2023-03-31T05:50:36.0607728Z test_jit.py::TestScript::test_inplace_add PASSED [ 66%] 2023-03-31T05:50:36.0607867Z test_jit.py::TestScript::test_inplace_copy_script PASSED [ 66%] 2023-03-31T05:50:36.0608008Z test_jit.py::TestScript::test_input_keyword_in_schema PASSED [ 66%] 2023-03-31T05:50:36.0608136Z test_jit.py::TestScript::test_int_cast PASSED [ 66%] 2023-03-31T05:50:36.0608278Z test_jit.py::TestScript::test_integral_shape_inference PASSED [ 66%] 2023-03-31T05:50:36.0608418Z test_jit.py::TestScript::test_interpret_graph PASSED [ 66%] 2023-03-31T05:50:36.0608544Z test_jit.py::TestScript::test_interpreter_fuzz PASSED [ 66%] 2023-03-31T05:50:36.0608682Z test_jit.py::TestScript::test_intlist_args PASSED [ 66%] 2023-03-31T05:50:36.0608825Z test_jit.py::TestScript::test_invalid_call_arguments PASSED [ 66%] 2023-03-31T05:50:36.0608966Z test_jit.py::TestScript::test_invalid_lhs_assignment PASSED [ 66%] 2023-03-31T05:50:36.0609109Z test_jit.py::TestScript::test_invalid_prefix_annotation PASSED [ 66%] 2023-03-31T05:50:36.0609241Z test_jit.py::TestScript::test_irparser PASSED [ 66%] 2023-03-31T05:50:36.0609371Z test_jit.py::TestScript::test_is_after_use PASSED [ 66%] 2023-03-31T05:50:36.0609500Z test_jit.py::TestScript::test_is_isnot PASSED [ 66%] 2023-03-31T05:50:36.0609620Z test_jit.py::TestScript::test_is_optional PASSED [ 66%] 2023-03-31T05:50:36.0609815Z test_jit.py::TestScript::test_is_scripting PASSED [ 66%] 2023-03-31T05:50:36.0609958Z test_jit.py::TestScript::test_is_scripting_metacompile PASSED [ 66%] 2023-03-31T05:50:36.0610142Z test_jit.py::TestScript::test_isinstance PASSED [ 66%] 2023-03-31T05:50:36.0610282Z test_jit.py::TestScript::test_isinstance_dynamic PASSED [ 66%] 2023-03-31T05:50:36.0610423Z test_jit.py::TestScript::test_isinstance_metacompile PASSED [ 66%] 2023-03-31T05:50:36.0610564Z test_jit.py::TestScript::test_isinstance_refinement PASSED [ 66%] 2023-03-31T05:50:36.0610695Z test_jit.py::TestScript::test_jitter_bug PASSED [ 66%] 2023-03-31T05:50:36.0610813Z test_jit.py::TestScript::test_keyword PASSED [ 66%] 2023-03-31T05:50:36.0610954Z test_jit.py::TestScript::test_kwarg_expansion_error PASSED [ 66%] 2023-03-31T05:50:36.0611095Z test_jit.py::TestScript::test_kwargs_error_msg PASSED [ 66%] 2023-03-31T05:50:36.0611228Z test_jit.py::TestScript::test_lazy_script PASSED [ 67%] 2023-03-31T05:50:36.0611382Z test_jit.py::TestScript::test_lhs_advanced_indexing_assignment PASSED [ 67%] 2023-03-31T05:50:36.0611550Z test_jit.py::TestScript::test_lhs_advanced_indexing_augmented_assignment PASSED [ 67%] 2023-03-31T05:50:36.0611682Z test_jit.py::TestScript::test_lhs_indexing PASSED [ 67%] 2023-03-31T05:50:36.0611823Z test_jit.py::TestScript::test_lhs_indexing_increment PASSED [ 67%] 2023-03-31T05:50:36.0611955Z test_jit.py::TestScript::test_lhs_indexing_increment_list PASSED [ 67%] 2023-03-31T05:50:36.0612102Z test_jit.py::TestScript::test_lhs_indexing_increment_list_prim PASSED [ 67%] 2023-03-31T05:50:36.0612237Z test_jit.py::TestScript::test_lhs_indexing_list PASSED [ 67%] 2023-03-31T05:50:36.0612373Z test_jit.py::TestScript::test_lhs_indexing_multi PASSED [ 67%] 2023-03-31T05:50:36.0612510Z test_jit.py::TestScript::test_linear_grad PASSED [ 67%] 2023-03-31T05:50:36.0612662Z test_jit.py::TestScript::test_list_comprehension_modulelist PASSED [ 67%] 2023-03-31T05:50:36.0612812Z test_jit.py::TestScript::test_list_comprehension_variable_write PASSED [ 67%] 2023-03-31T05:50:36.0612948Z test_jit.py::TestScript::test_list_iterables PASSED [ 67%] 2023-03-31T05:50:36.0613070Z test_jit.py::TestScript::test_list_python_op PASSED [ 67%] 2023-03-31T05:50:36.0613200Z test_jit.py::TestScript::test_list_unify PASSED [ 67%] 2023-03-31T05:50:36.0613328Z test_jit.py::TestScript::test_literal PASSED [ 67%] 2023-03-31T05:50:36.0613458Z test_jit.py::TestScript::test_literals PASSED [ 67%] 2023-03-31T05:50:36.0613596Z test_jit.py::TestScript::test_logical_short_circuit PASSED [ 67%] 2023-03-31T05:50:36.0613734Z test_jit.py::TestScript::test_loop_liveness PASSED [ 67%] 2023-03-31T05:50:36.0613873Z test_jit.py::TestScript::test_loop_unroll_negative PASSED [ 67%] 2023-03-31T05:50:36.0614019Z test_jit.py::TestScript::test_loop_unroll_unused_counter PASSED [ 67%] 2023-03-31T05:50:36.0614142Z test_jit.py::TestScript::test_loop_unrolling PASSED [ 67%] 2023-03-31T05:50:36.0614282Z test_jit.py::TestScript::test_loop_unrolling_const PASSED [ 67%] 2023-03-31T05:50:36.0614421Z test_jit.py::TestScript::test_loop_unrolling_nested PASSED [ 67%] 2023-03-31T05:50:36.0614561Z test_jit.py::TestScript::test_lower_nested_tuples PASSED [ 67%] 2023-03-31T05:50:36.0614689Z test_jit.py::TestScript::test_math_ops PASSED [ 68%] 2023-03-31T05:50:36.0614877Z test_jit.py::TestScript::test_maxpool_guard_elimination SKIPPED (bailouts are being deprecated) [ 68%] 2023-03-31T05:50:36.0615171Z test_jit.py::TestScript::test_meshgrid SKIPPED (Profiling executor fails to recognize that tensors in a list require gradients) [ 68%] 2023-03-31T05:50:36.0615363Z test_jit.py::TestScript::test_method_casts_script PASSED [ 68%] 2023-03-31T05:50:36.0615501Z test_jit.py::TestScript::test_method_no_self PASSED [ 68%] 2023-03-31T05:50:36.0615629Z test_jit.py::TestScript::test_method_overloading PASSED [ 68%] 2023-03-31T05:50:36.0615767Z test_jit.py::TestScript::test_missing_getstate PASSED [ 68%] 2023-03-31T05:50:36.0615898Z test_jit.py::TestScript::test_mm_batching PASSED [ 68%] 2023-03-31T05:50:36.0616032Z test_jit.py::TestScript::test_module_apis PASSED [ 68%] 2023-03-31T05:50:36.0616164Z test_jit.py::TestScript::test_module_attrs PASSED [ 68%] 2023-03-31T05:50:36.0616306Z test_jit.py::TestScript::test_module_copy_with_attributes PASSED [ 68%] 2023-03-31T05:50:36.0616447Z test_jit.py::TestScript::test_module_copying PASSED [ 68%] 2023-03-31T05:50:36.0616580Z test_jit.py::TestScript::test_module_error PASSED [ 68%] 2023-03-31T05:50:36.0616716Z test_jit.py::TestScript::test_module_method_reassignment PASSED [ 68%] 2023-03-31T05:50:36.0616850Z test_jit.py::TestScript::test_module_none_attrs PASSED [ 68%] 2023-03-31T05:50:36.0616995Z test_jit.py::TestScript::test_module_parameters_and_buffers PASSED [ 68%] 2023-03-31T05:50:36.0617125Z test_jit.py::TestScript::test_module_str PASSED [ 68%] 2023-03-31T05:50:36.0617269Z test_jit.py::TestScript::test_module_with_params_called_fails PASSED [ 68%] 2023-03-31T05:50:36.0617405Z test_jit.py::TestScript::test_multi_reduction PASSED [ 68%] 2023-03-31T05:50:36.0617544Z test_jit.py::TestScript::test_multi_starred_expr_lhs PASSED [ 68%] 2023-03-31T05:50:36.0617690Z test_jit.py::TestScript::test_multiline_annot_ast_py3_fn PASSED [ 68%] 2023-03-31T05:50:36.0617837Z test_jit.py::TestScript::test_multiline_optional_future_refinement PASSED [ 68%] 2023-03-31T05:50:36.0617985Z test_jit.py::TestScript::test_multiline_string_dedents PASSED [ 68%] 2023-03-31T05:50:36.0618121Z test_jit.py::TestScript::test_multiple_assign PASSED [ 68%] 2023-03-31T05:50:36.0618261Z test_jit.py::TestScript::test_multiple_assignment PASSED [ 68%] 2023-03-31T05:50:36.0618394Z test_jit.py::TestScript::test_mutable_dce PASSED [ 68%] 2023-03-31T05:50:36.0618534Z test_jit.py::TestScript::test_mutable_dce_block PASSED [ 69%] 2023-03-31T05:50:36.0618674Z test_jit.py::TestScript::test_mutable_dce_graph_input PASSED [ 69%] 2023-03-31T05:50:36.0618822Z test_jit.py::TestScript::test_mutable_dce_indirect_wildcard_write PASSED [ 69%] 2023-03-31T05:50:36.0618956Z test_jit.py::TestScript::test_mutable_dce_indirect_wildcards PASSED [ 69%] 2023-03-31T05:50:36.0619097Z test_jit.py::TestScript::test_mutable_dce_list PASSED [ 69%] 2023-03-31T05:50:36.0619229Z test_jit.py::TestScript::test_mutable_dce_loop PASSED [ 69%] 2023-03-31T05:50:36.0619374Z test_jit.py::TestScript::test_mutable_dce_wildcards PASSED [ 69%] 2023-03-31T05:50:36.0619509Z test_jit.py::TestScript::test_mutate_constant PASSED [ 69%] 2023-03-31T05:50:36.0619643Z test_jit.py::TestScript::test_mypy_type_ignore PASSED [ 69%] 2023-03-31T05:50:36.0619787Z test_jit.py::TestScript::test_named_buffers_are_iterable PASSED [ 69%] 2023-03-31T05:50:36.0619924Z test_jit.py::TestScript::test_namedtuple_attr PASSED [ 69%] 2023-03-31T05:50:36.0620072Z test_jit.py::TestScript::test_namedtuple_default_values_Tensor_type PASSED [ 69%] 2023-03-31T05:50:36.0620238Z test_jit.py::TestScript::test_namedtuple_default_values_container_type PASSED [ 69%] 2023-03-31T05:50:36.0620445Z test_jit.py::TestScript::test_namedtuple_default_values_missing PASSED [ 69%] 2023-03-31T05:50:36.0620604Z test_jit.py::TestScript::test_namedtuple_default_values_simple_type PASSED [ 69%] 2023-03-31T05:50:36.0620839Z test_jit.py::TestScript::test_namedtuple_default_values_using_factory_constructor PASSED [ 69%] 2023-03-31T05:50:36.0620984Z test_jit.py::TestScript::test_namedtuple_python PASSED [ 69%] 2023-03-31T05:50:36.0621129Z test_jit.py::TestScript::test_namedtuple_type_inference PASSED [ 69%] 2023-03-31T05:50:36.0621262Z test_jit.py::TestScript::test_narrow_copy PASSED [ 69%] 2023-03-31T05:50:36.0621387Z test_jit.py::TestScript::test_nested_aug_assign PASSED [ 69%] 2023-03-31T05:50:36.0621523Z test_jit.py::TestScript::test_nested_bailouts PASSED [ 69%] 2023-03-31T05:50:36.0621656Z test_jit.py::TestScript::test_nested_breaks PASSED [ 69%] 2023-03-31T05:50:36.0621802Z test_jit.py::TestScript::test_nested_list_construct PASSED [ 69%] 2023-03-31T05:50:36.0621939Z test_jit.py::TestScript::test_nested_select_assign PASSED [ 69%] 2023-03-31T05:50:36.0622072Z test_jit.py::TestScript::test_nn_GRU PASSED [ 69%] 2023-03-31T05:50:36.0622200Z test_jit.py::TestScript::test_nn_LSTM PASSED [ 70%] 2023-03-31T05:50:36.0622335Z test_jit.py::TestScript::test_nn_LSTM_with_layers PASSED [ 70%] 2023-03-31T05:50:36.0622463Z test_jit.py::TestScript::test_nn_init PASSED [ 70%] 2023-03-31T05:50:36.0622586Z test_jit.py::TestScript::test_no_dtype_shape PASSED [ 70%] 2023-03-31T05:50:36.0622727Z test_jit.py::TestScript::test_no_self_arg_ignore_function PASSED [ 70%] 2023-03-31T05:50:36.0622860Z test_jit.py::TestScript::test_non_final_return PASSED [ 70%] 2023-03-31T05:50:36.0623108Z test_jit.py::TestScript::test_none_type_str PASSED [ 70%] 2023-03-31T05:50:36.0623248Z test_jit.py::TestScript::test_not PASSED [ 70%] 2023-03-31T05:50:36.0623388Z test_jit.py::TestScript::test_not_initialized_err PASSED [ 70%] 2023-03-31T05:50:36.0623529Z test_jit.py::TestScript::test_ntuple_builtins PASSED [ 70%] 2023-03-31T05:50:36.0623660Z test_jit.py::TestScript::test_number_abs PASSED [ 70%] 2023-03-31T05:50:36.0623786Z test_jit.py::TestScript::test_number_augassign PASSED [ 70%] 2023-03-31T05:50:36.0623935Z test_jit.py::TestScript::test_number_augassign_bitwise_lshift PASSED [ 70%] 2023-03-31T05:50:36.0624080Z test_jit.py::TestScript::test_number_augassign_bitwise_pow PASSED [ 70%] 2023-03-31T05:50:36.0624229Z test_jit.py::TestScript::test_number_augassign_bitwise_rshift PASSED [ 70%] 2023-03-31T05:50:36.0624362Z test_jit.py::TestScript::test_number_div PASSED [ 70%] 2023-03-31T05:50:36.0624501Z test_jit.py::TestScript::test_number_math PASSED [ 70%] 2023-03-31T05:50:36.0624631Z test_jit.py::TestScript::test_number_neg PASSED [ 70%] 2023-03-31T05:50:36.0624819Z test_jit.py::TestScript::test_old_models_bc SKIPPED (PyTorch is build without Caffe2 support) [ 70%] 2023-03-31T05:50:36.0624939Z test_jit.py::TestScript::test_oneline_func PASSED [ 70%] 2023-03-31T05:50:36.0625071Z test_jit.py::TestScript::test_op_dtype PASSED [ 70%] 2023-03-31T05:50:36.0625214Z test_jit.py::TestScript::test_operator_precedence PASSED [ 70%] 2023-03-31T05:50:36.0625560Z test_jit.py::TestScript::test_optional_list SKIPPED (the current version of Profiler doesn't profile/specialize Optionals) [ 70%] 2023-03-31T05:50:36.0625885Z test_jit.py::TestScript::test_optional_tensor SKIPPED (the current version of Profiler doesn't profile/specialize Optionals) [ 70%] 2023-03-31T05:50:36.0626084Z test_jit.py::TestScript::test_ord PASSED [ 70%] 2023-03-31T05:50:36.0626220Z test_jit.py::TestScript::test_override_magic PASSED [ 70%] 2023-03-31T05:50:36.0626455Z test_jit.py::TestScript::test_pack_tuple_into_non_var PASSED [ 71%] 2023-03-31T05:50:36.0626585Z test_jit.py::TestScript::test_pack_unpack_nested PASSED [ 71%] 2023-03-31T05:50:36.0626720Z test_jit.py::TestScript::test_pack_unpack_state PASSED [ 71%] 2023-03-31T05:50:36.0626858Z test_jit.py::TestScript::test_parameter_order PASSED [ 71%] 2023-03-31T05:50:36.0627004Z test_jit.py::TestScript::test_parse_empty_tuple_annotation PASSED [ 71%] 2023-03-31T05:50:36.0627171Z test_jit.py::TestScript::test_parse_empty_tuple_annotation_element_error PASSED [ 71%] 2023-03-31T05:50:36.0627311Z test_jit.py::TestScript::test_parse_nested_names PASSED [ 71%] 2023-03-31T05:50:36.0627455Z test_jit.py::TestScript::test_parse_none_type_annotation PASSED [ 71%] 2023-03-31T05:50:36.0627603Z test_jit.py::TestScript::test_parse_tensor_constants PASSED [ 71%] 2023-03-31T05:50:36.0627728Z test_jit.py::TestScript::test_parser_kwargonly PASSED [ 71%] 2023-03-31T05:50:36.0627876Z test_jit.py::TestScript::test_parser_type_annotations PASSED [ 71%] 2023-03-31T05:50:36.0628025Z test_jit.py::TestScript::test_parser_type_annotations_comment PASSED [ 71%] 2023-03-31T05:50:36.0628200Z test_jit.py::TestScript::test_parser_type_annotations_incompatible_expression PASSED [ 71%] 2023-03-31T05:50:36.0628369Z test_jit.py::TestScript::test_parser_type_annotations_subscript_non_ident PASSED [ 71%] 2023-03-31T05:50:36.0628537Z test_jit.py::TestScript::test_parser_type_annotations_subscript_tensor PASSED [ 71%] 2023-03-31T05:50:36.0628695Z test_jit.py::TestScript::test_parser_type_annotations_unknown_type PASSED [ 71%] 2023-03-31T05:50:36.0628833Z test_jit.py::TestScript::test_partial_returns PASSED [ 71%] 2023-03-31T05:50:36.0628967Z test_jit.py::TestScript::test_pass PASSED [ 71%] 2023-03-31T05:50:36.0629095Z test_jit.py::TestScript::test_pickle_checkpoint PASSED [ 71%] 2023-03-31T05:50:36.0629249Z test_jit.py::TestScript::test_pickle_checkpoint_cuda SKIPPED (no CUDA) [ 71%] 2023-03-31T05:50:36.0629390Z test_jit.py::TestScript::test_pickle_checkpoint_tup PASSED [ 71%] 2023-03-31T05:50:36.0629568Z test_jit.py::TestScript::test_pow_scalar_backward_cuda SKIPPED (device tests require CUDA) [ 71%] 2023-03-31T05:50:36.0629714Z test_jit.py::TestScript::test_pretty_print_function PASSED [ 71%] 2023-03-31T05:50:36.0629905Z test_jit.py::TestScript::test_prim_grad_undefined SKIPPED (shape analysis is only enabled in Legacy) [ 71%] 2023-03-31T05:50:36.0630037Z test_jit.py::TestScript::test_print PASSED [ 71%] 2023-03-31T05:50:36.0630171Z test_jit.py::TestScript::test_print_kwargs PASSED [ 71%] 2023-03-31T05:50:36.0630349Z test_jit.py::TestScript::test_profiling_graph_executor SKIPPED (bailouts are being deprecated) [ 72%] 2023-03-31T05:50:36.0630492Z test_jit.py::TestScript::test_profiling_merge PASSED [ 72%] 2023-03-31T05:50:36.0630638Z test_jit.py::TestScript::test_pybind_type_comparisons PASSED [ 72%] 2023-03-31T05:50:36.0630772Z test_jit.py::TestScript::test_python_call PASSED [ 72%] 2023-03-31T05:50:36.0630917Z test_jit.py::TestScript::test_python_call_annotation PASSED [ 72%] 2023-03-31T05:50:36.0631066Z test_jit.py::TestScript::test_python_call_annoytation_failure PASSED [ 72%] 2023-03-31T05:50:36.0631204Z test_jit.py::TestScript::test_python_call_failure PASSED [ 72%] 2023-03-31T05:50:36.0631343Z test_jit.py::TestScript::test_python_call_non_tensor PASSED [ 72%] 2023-03-31T05:50:36.0631473Z test_jit.py::TestScript::test_python_call_non_tensor_wrong PASSED [ 72%] 2023-03-31T05:50:36.0631659Z test_jit.py::TestScript::test_python_frontend PASSED [ 72%] 2023-03-31T05:50:36.0631798Z test_jit.py::TestScript::test_python_frontend_py3 PASSED [ 72%] 2023-03-31T05:50:36.0632000Z test_jit.py::TestScript::test_python_frontend_source_range PASSED [ 72%] 2023-03-31T05:50:36.0632139Z test_jit.py::TestScript::test_python_op_builtins PASSED [ 72%] 2023-03-31T05:50:36.0632276Z test_jit.py::TestScript::test_python_op_name PASSED [ 72%] 2023-03-31T05:50:36.0632418Z test_jit.py::TestScript::test_python_val_doesnt_have_attr PASSED [ 72%] 2023-03-31T05:50:36.0632581Z test_jit.py::TestScript::test_rand SKIPPED (the original version of test_rand) [ 72%] 2023-03-31T05:50:36.0632706Z test_jit.py::TestScript::test_rand_profiling PASSED [ 72%] 2023-03-31T05:50:36.0632838Z test_jit.py::TestScript::test_range_args PASSED [ 72%] 2023-03-31T05:50:36.0632982Z test_jit.py::TestScript::test_reassign_module_lhs PASSED [ 72%] 2023-03-31T05:50:36.0633120Z test_jit.py::TestScript::test_reassign_module_rhs PASSED [ 72%] 2023-03-31T05:50:36.0633262Z test_jit.py::TestScript::test_refine_tuple_types PASSED [ 72%] 2023-03-31T05:50:36.0633398Z test_jit.py::TestScript::test_remove_dropout PASSED [ 72%] 2023-03-31T05:50:36.0633542Z test_jit.py::TestScript::test_repeated_script_on_function PASSED [ 72%] 2023-03-31T05:50:36.0633677Z test_jit.py::TestScript::test_request_bailout PASSED [ 72%] 2023-03-31T05:50:36.0633835Z test_jit.py::TestScript::test_requires_grad_loop SKIPPED (Peeling is now disabled) [ 72%] 2023-03-31T05:50:36.0633980Z test_jit.py::TestScript::test_rescripting_loaded_modules PASSED [ 73%] 2023-03-31T05:50:36.0634116Z test_jit.py::TestScript::test_resize_input_ops PASSED [ 73%] 2023-03-31T05:50:36.0634244Z test_jit.py::TestScript::test_return PASSED [ 73%] 2023-03-31T05:50:36.0634387Z test_jit.py::TestScript::test_return_stmt_not_at_end PASSED [ 73%] 2023-03-31T05:50:36.0634523Z test_jit.py::TestScript::test_return_tuple PASSED [ 73%] 2023-03-31T05:50:36.0634662Z test_jit.py::TestScript::test_robust_op_resolution PASSED [ 73%] 2023-03-31T05:50:36.0634790Z test_jit.py::TestScript::test_round PASSED [ 73%] 2023-03-31T05:50:36.0634926Z test_jit.py::TestScript::test_save_load_attr_error PASSED [ 73%] 2023-03-31T05:50:36.0635053Z test_jit.py::TestScript::test_script_annotation PASSED [ 73%] 2023-03-31T05:50:36.0635191Z test_jit.py::TestScript::test_script_bool_constant PASSED [ 73%] 2023-03-31T05:50:36.0635322Z test_jit.py::TestScript::test_script_chunk PASSED [ 73%] 2023-03-31T05:50:36.0635460Z test_jit.py::TestScript::test_script_clamp_none PASSED [ 73%] 2023-03-31T05:50:36.0635597Z test_jit.py::TestScript::test_script_copy PASSED [ 73%] 2023-03-31T05:50:36.0635728Z test_jit.py::TestScript::test_script_cu PASSED [ 73%] 2023-03-31T05:50:36.0635870Z test_jit.py::TestScript::test_script_define_order PASSED [ 73%] 2023-03-31T05:50:36.0636018Z test_jit.py::TestScript::test_script_define_order_recursive_fail PASSED [ 73%] 2023-03-31T05:50:36.0636143Z test_jit.py::TestScript::test_script_docstring PASSED [ 73%] 2023-03-31T05:50:36.0636291Z test_jit.py::TestScript::test_script_forward_method_replacement PASSED [ 73%] 2023-03-31T05:50:36.0636451Z test_jit.py::TestScript::test_script_get_device_cuda SKIPPED (requires CUDA) [ 73%] 2023-03-31T05:50:36.0636592Z test_jit.py::TestScript::test_script_get_tracing_state PASSED [ 73%] 2023-03-31T05:50:36.0636728Z test_jit.py::TestScript::test_script_is_tracing PASSED [ 73%] 2023-03-31T05:50:36.0636917Z test_jit.py::TestScript::test_script_kwargs_fn_call PASSED [ 73%] 2023-03-31T05:50:36.0637059Z test_jit.py::TestScript::test_script_method_docstring PASSED [ 73%] 2023-03-31T05:50:36.0637268Z test_jit.py::TestScript::test_script_method_torch_function_overload PASSED [ 73%] 2023-03-31T05:50:36.0637392Z test_jit.py::TestScript::test_script_module PASSED [ 73%] 2023-03-31T05:50:36.0637536Z test_jit.py::TestScript::test_script_module_call_noscript PASSED [ 73%] 2023-03-31T05:50:36.0637676Z test_jit.py::TestScript::test_script_module_const PASSED [ 74%] 2023-03-31T05:50:36.0637824Z test_jit.py::TestScript::test_script_module_const_submodule_fail PASSED [ 74%] 2023-03-31T05:50:36.0637966Z test_jit.py::TestScript::test_script_module_export_blocks PASSED [ 74%] 2023-03-31T05:50:36.0638115Z test_jit.py::TestScript::test_script_module_export_shared_storage PASSED [ 74%] 2023-03-31T05:50:36.0638262Z test_jit.py::TestScript::test_script_module_export_submodule PASSED [ 74%] 2023-03-31T05:50:36.0638465Z test_jit.py::TestScript::test_script_module_export_tensor_cuda SKIPPED (testing cuda tensors require CUDA) [ 74%] 2023-03-31T05:50:36.0638605Z test_jit.py::TestScript::test_script_module_export_tensor_type PASSED [ 74%] 2023-03-31T05:50:36.0638743Z test_jit.py::TestScript::test_script_module_fail_exist PASSED [ 74%] 2023-03-31T05:50:36.0638879Z test_jit.py::TestScript::test_script_module_for PASSED [ 74%] 2023-03-31T05:50:36.0639019Z test_jit.py::TestScript::test_script_module_for2 PASSED [ 74%] 2023-03-31T05:50:36.0639162Z test_jit.py::TestScript::test_script_module_invalid_consts PASSED [ 74%] 2023-03-31T05:50:36.0639308Z test_jit.py::TestScript::test_script_module_nochange_submodule PASSED [ 74%] 2023-03-31T05:50:36.0639687Z test_jit.py::TestScript::test_script_module_none_exist_fail SKIPPED ([module dedupe] currently NoneType refinement on optional attributes doesn't work.) [ 74%] 2023-03-31T05:50:36.0639833Z test_jit.py::TestScript::test_script_module_not_tuple PASSED [ 74%] 2023-03-31T05:50:36.0639970Z test_jit.py::TestScript::test_script_module_param_buffer_mutation PASSED [ 74%] 2023-03-31T05:50:36.0640119Z test_jit.py::TestScript::test_script_module_star_assign2 PASSED [ 74%] 2023-03-31T05:50:36.0640268Z test_jit.py::TestScript::test_script_module_star_assign2_inplace PASSED [ 74%] 2023-03-31T05:50:36.0640429Z test_jit.py::TestScript::test_script_module_star_assign_fail_builtin PASSED [ 74%] 2023-03-31T05:50:36.0640590Z test_jit.py::TestScript::test_script_module_star_assign_fail_pythonop PASSED [ 74%] 2023-03-31T05:50:36.0640750Z test_jit.py::TestScript::test_script_module_tensor_subclass_argument PASSED [ 74%] 2023-03-31T05:50:36.0640890Z test_jit.py::TestScript::test_script_nested_mod_list PASSED [ 74%] 2023-03-31T05:50:36.0641039Z test_jit.py::TestScript::test_script_non_tensor_args_outputs PASSED [ 74%] 2023-03-31T05:50:36.0641182Z test_jit.py::TestScript::test_script_optional_none PASSED [ 74%] 2023-03-31T05:50:36.0641308Z test_jit.py::TestScript::test_script_outputs PASSED [ 74%] 2023-03-31T05:50:36.0641456Z test_jit.py::TestScript::test_script_pack_padded_sequence PASSED [ 74%] 2023-03-31T05:50:36.0641604Z test_jit.py::TestScript::test_script_pad_sequence_pack_sequence PASSED [ 75%] 2023-03-31T05:50:36.0641738Z test_jit.py::TestScript::test_script_scope PASSED [ 75%] 2023-03-31T05:50:36.0641877Z test_jit.py::TestScript::test_script_sequential_for PASSED [ 75%] 2023-03-31T05:50:36.0642023Z test_jit.py::TestScript::test_script_sequential_in_mod_list PASSED [ 75%] 2023-03-31T05:50:36.0642175Z test_jit.py::TestScript::test_script_sequential_multi_output_fail PASSED [ 75%] 2023-03-31T05:50:36.0642320Z test_jit.py::TestScript::test_script_sequential_orderdict PASSED [ 75%] 2023-03-31T05:50:36.0642458Z test_jit.py::TestScript::test_script_sequential_sliced_iteration PASSED [ 75%] 2023-03-31T05:50:36.0642659Z test_jit.py::TestScript::test_script_star_assign PASSED [ 75%] 2023-03-31T05:50:36.0642795Z test_jit.py::TestScript::test_script_star_expr PASSED [ 75%] 2023-03-31T05:50:36.0642989Z test_jit.py::TestScript::test_script_star_expr_string PASSED [ 75%] 2023-03-31T05:50:36.0643293Z test_jit.py::TestScript::test_scriptable_fn_as_attr PASSED [ 75%] 2023-03-31T05:50:36.0643462Z test_jit.py::TestScript::test_scriptmodule_multi_head_attn_cuda SKIPPED (no CUDA) [ 75%] 2023-03-31T05:50:36.0643631Z test_jit.py::TestScript::test_scriptmodule_releases_tensors_cuda SKIPPED (no CUDA) [ 75%] 2023-03-31T05:50:36.0643796Z test_jit.py::TestScript::test_scriptmodule_transformer_cuda SKIPPED (no CUDA) [ 75%] 2023-03-31T05:50:36.0643923Z test_jit.py::TestScript::test_select_after_chunk PASSED [ 75%] 2023-03-31T05:50:36.0644062Z test_jit.py::TestScript::test_sequence_parsing PASSED [ 75%] 2023-03-31T05:50:36.0644219Z test_jit.py::TestScript::test_sequential_intermediary_types PASSED [ 75%] 2023-03-31T05:50:36.0644362Z test_jit.py::TestScript::test_serialization_big_ints PASSED [ 75%] 2023-03-31T05:50:36.0644508Z test_jit.py::TestScript::test_serialization_sharing PASSED [ 75%] 2023-03-31T05:50:36.0644648Z test_jit.py::TestScript::test_serialize_long_lines PASSED [ 75%] 2023-03-31T05:50:36.0644792Z test_jit.py::TestScript::test_serialized_source_ranges PASSED [ 75%] 2023-03-31T05:50:36.0644936Z test_jit.py::TestScript::test_serialized_source_ranges2 PASSED [ 75%] 2023-03-31T05:50:36.0645082Z test_jit.py::TestScript::test_serialized_source_ranges_dont_jitter PASSED [ 75%] 2023-03-31T05:50:36.0645231Z test_jit.py::TestScript::test_serialized_source_ranges_graph PASSED [ 75%] 2023-03-31T05:50:36.0645377Z test_jit.py::TestScript::test_serialized_source_ranges_no_dups PASSED [ 75%] 2023-03-31T05:50:36.0645531Z test_jit.py::TestScript::test_set_attribute_through_optional PASSED [ 75%] 2023-03-31T05:50:36.0645677Z test_jit.py::TestScript::test_shape_analysis_grad_property PASSED [ 76%] 2023-03-31T05:50:36.0645820Z test_jit.py::TestScript::test_shape_analysis_loop PASSED [ 76%] 2023-03-31T05:50:36.0645964Z test_jit.py::TestScript::test_shape_prop_promote_scalar_arg PASSED [ 76%] 2023-03-31T05:50:36.0646104Z test_jit.py::TestScript::test_shape_prop_promotion PASSED [ 76%] 2023-03-31T05:50:36.0646227Z test_jit.py::TestScript::test_signed_float_zero PASSED [ 76%] 2023-03-31T05:50:36.0646370Z test_jit.py::TestScript::test_single_starred_expr_for_loop PASSED [ 76%] 2023-03-31T05:50:36.0646506Z test_jit.py::TestScript::test_single_starred_lhs PASSED [ 76%] 2023-03-31T05:50:36.0646651Z test_jit.py::TestScript::test_singleton_tuple_unpack PASSED [ 76%] 2023-03-31T05:50:36.0646835Z test_jit.py::TestScript::test_slice_guard_elimination SKIPPED (bailouts are being deprecated) [ 76%] 2023-03-31T05:50:36.0646968Z test_jit.py::TestScript::test_split PASSED [ 76%] 2023-03-31T05:50:36.0647099Z test_jit.py::TestScript::test_stack PASSED [ 76%] 2023-03-31T05:50:36.0647236Z test_jit.py::TestScript::test_static_if_prop PASSED [ 76%] 2023-03-31T05:50:36.0647363Z test_jit.py::TestScript::test_static_method_on_module PASSED [ 76%] 2023-03-31T05:50:36.0647498Z test_jit.py::TestScript::test_static_methods PASSED [ 76%] 2023-03-31T05:50:36.0647626Z test_jit.py::TestScript::test_str_cast PASSED [ 76%] 2023-03-31T05:50:36.0647757Z test_jit.py::TestScript::test_string_cu PASSED [ 76%] 2023-03-31T05:50:36.0647906Z test_jit.py::TestScript::test_string_device_implicit_conversion PASSED [ 76%] 2023-03-31T05:50:36.0648043Z test_jit.py::TestScript::test_string_frontend_elif PASSED [ 76%] 2023-03-31T05:50:36.0648266Z test_jit.py::TestScript::test_string_index PASSED [ 76%] 2023-03-31T05:50:36.0648396Z test_jit.py::TestScript::test_string_len PASSED [ 76%] 2023-03-31T05:50:36.0648579Z test_jit.py::TestScript::test_string_list PASSED [ 76%] 2023-03-31T05:50:36.0648719Z test_jit.py::TestScript::test_string_new_line PASSED [ 76%] 2023-03-31T05:50:36.0648849Z test_jit.py::TestScript::test_string_ops PASSED [ 76%] 2023-03-31T05:50:36.0648982Z test_jit.py::TestScript::test_string_print PASSED [ 76%] 2023-03-31T05:50:36.0649118Z test_jit.py::TestScript::test_string_single_escape PASSED [ 76%] 2023-03-31T05:50:36.0649253Z test_jit.py::TestScript::test_string_slicing PASSED [ 76%] 2023-03-31T05:50:36.0649385Z test_jit.py::TestScript::test_string_sort PASSED [ 77%] 2023-03-31T05:50:36.0649523Z test_jit.py::TestScript::test_string_sorted PASSED [ 77%] 2023-03-31T05:50:36.0649675Z test_jit.py::TestScript::test_submodule_attribute_serialization PASSED [ 77%] 2023-03-31T05:50:36.0649804Z test_jit.py::TestScript::test_submodule_twice PASSED [ 77%] 2023-03-31T05:50:36.0649930Z test_jit.py::TestScript::test_sum PASSED [ 77%] 2023-03-31T05:50:36.0650065Z test_jit.py::TestScript::test_sum_list_diff_elms PASSED [ 77%] 2023-03-31T05:50:36.0650197Z test_jit.py::TestScript::test_sum_list_empty PASSED [ 77%] 2023-03-31T05:50:36.0650331Z test_jit.py::TestScript::test_sum_list_literal PASSED [ 77%] 2023-03-31T05:50:36.0650461Z test_jit.py::TestScript::test_sum_list_one PASSED [ 77%] 2023-03-31T05:50:36.0650598Z test_jit.py::TestScript::test_sum_list_wrong_type PASSED [ 77%] 2023-03-31T05:50:36.0650734Z test_jit.py::TestScript::test_sys_stdout_override PASSED [ 77%] 2023-03-31T05:50:36.0651050Z test_jit.py::TestScript::test_tensor_as_tensor_shape_prop SKIPPED (Simple Executor doesn't have any shapes to propagate) [ 77%] 2023-03-31T05:50:36.0651186Z test_jit.py::TestScript::test_tensor_data PASSED [ 77%] 2023-03-31T05:50:36.0651348Z test_jit.py::TestScript::test_tensor_device SKIPPED (device tests require CUDA) [ 77%] 2023-03-31T05:50:36.0651481Z test_jit.py::TestScript::test_tensor_dtype PASSED [ 77%] 2023-03-31T05:50:36.0651613Z test_jit.py::TestScript::test_tensor_grad PASSED [ 77%] 2023-03-31T05:50:36.0651752Z test_jit.py::TestScript::test_tensor_import_export PASSED [ 77%] 2023-03-31T05:50:36.0651882Z test_jit.py::TestScript::test_tensor_len PASSED [ 77%] 2023-03-31T05:50:36.0652019Z test_jit.py::TestScript::test_tensor_number_math PASSED [ 77%] 2023-03-31T05:50:36.0652155Z test_jit.py::TestScript::test_tensor_number_math_cuda SKIPPED (No CUDA) [ 77%] 2023-03-31T05:50:36.0652329Z test_jit.py::TestScript::test_tensor_requires_grad SKIPPED (testing legacy behavior) [ 77%] 2023-03-31T05:50:36.0652466Z test_jit.py::TestScript::test_tensor_shape PASSED [ 77%] 2023-03-31T05:50:36.0652603Z test_jit.py::TestScript::test_tensor_subclasses PASSED [ 77%] 2023-03-31T05:50:36.0652737Z test_jit.py::TestScript::test_tensor_to PASSED [ 77%] 2023-03-31T05:50:36.0652870Z test_jit.py::TestScript::test_tensor_to_cpu PASSED [ 77%] 2023-03-31T05:50:36.0653034Z test_jit.py::TestScript::test_tensor_to_cuda SKIPPED (device tests require CUDA) [ 78%] 2023-03-31T05:50:36.0653197Z test_jit.py::TestScript::test_tensor_to_device SKIPPED (device tests require CUDA) [ 78%] 2023-03-31T05:50:36.0653318Z test_jit.py::TestScript::test_ternary PASSED [ 78%] 2023-03-31T05:50:36.0653460Z test_jit.py::TestScript::test_ternary_module_type_hint PASSED [ 78%] 2023-03-31T05:50:36.0653672Z test_jit.py::TestScript::test_ternary_right_associative PASSED [ 78%] 2023-03-31T05:50:36.0653810Z test_jit.py::TestScript::test_ternary_static_if PASSED [ 78%] 2023-03-31T05:50:36.0653991Z test_jit.py::TestScript::test_torch_any PASSED [ 78%] 2023-03-31T05:50:36.0654169Z test_jit.py::TestScript::test_torch_functional SKIPPED (Skipping while landing PR stack) [ 78%] 2023-03-31T05:50:36.0654314Z test_jit.py::TestScript::test_torch_functional_tensordot_int PASSED [ 78%] 2023-03-31T05:50:36.0654463Z test_jit.py::TestScript::test_torch_functional_tensordot_list PASSED [ 78%] 2023-03-31T05:50:36.0654599Z test_jit.py::TestScript::test_torch_functional_tensordot_tensor PASSED [ 78%] 2023-03-31T05:50:36.0654748Z test_jit.py::TestScript::test_torch_functional_tensordot_tuple PASSED [ 78%] 2023-03-31T05:50:36.0654894Z test_jit.py::TestScript::test_torch_ignore_conversion_to_none PASSED [ 78%] 2023-03-31T05:50:36.0655033Z test_jit.py::TestScript::test_torch_manual_seed PASSED [ 78%] 2023-03-31T05:50:36.0655164Z test_jit.py::TestScript::test_torch_pow PASSED [ 78%] 2023-03-31T05:50:36.0655305Z test_jit.py::TestScript::test_torch_tensor_as_tensor PASSED [ 78%] 2023-03-31T05:50:36.0655450Z test_jit.py::TestScript::test_torch_tensor_as_tensor_empty_list PASSED [ 78%] 2023-03-31T05:50:36.0655587Z test_jit.py::TestScript::test_torch_tensor_bad_input PASSED [ 78%] 2023-03-31T05:50:36.0655710Z test_jit.py::TestScript::test_torch_tensor_dtype PASSED [ 78%] 2023-03-31T05:50:36.0655856Z test_jit.py::TestScript::test_torchscript_memoryformat PASSED [ 78%] 2023-03-31T05:50:36.0655997Z test_jit.py::TestScript::test_torchscript_multi_head_attn PASSED [ 78%] 2023-03-31T05:50:36.0656157Z test_jit.py::TestScript::test_torchscript_multi_head_attn_fast_path PASSED [ 78%] 2023-03-31T05:50:36.0656295Z test_jit.py::TestScript::test_training_param PASSED [ 78%] 2023-03-31T05:50:36.0656433Z test_jit.py::TestScript::test_tuple_assignments PASSED [ 78%] 2023-03-31T05:50:36.0656571Z test_jit.py::TestScript::test_tuple_error_msg PASSED [ 78%] 2023-03-31T05:50:36.0656707Z test_jit.py::TestScript::test_tuple_index_to_list PASSED [ 78%] 2023-03-31T05:50:36.0656841Z test_jit.py::TestScript::test_tuple_indexing PASSED [ 79%] 2023-03-31T05:50:36.0656958Z test_jit.py::TestScript::test_tuple_len PASSED [ 79%] 2023-03-31T05:50:36.0657094Z test_jit.py::TestScript::test_tuple_nested_sort PASSED [ 79%] 2023-03-31T05:50:36.0657223Z test_jit.py::TestScript::test_tuple_sort PASSED [ 79%] 2023-03-31T05:50:36.0657358Z test_jit.py::TestScript::test_tuple_sort_reverse PASSED [ 79%] 2023-03-31T05:50:36.0657491Z test_jit.py::TestScript::test_tuple_sorted PASSED [ 79%] 2023-03-31T05:50:36.0657623Z test_jit.py::TestScript::test_tuple_str PASSED [ 79%] 2023-03-31T05:50:36.0657759Z test_jit.py::TestScript::test_tuple_to_opt_list PASSED [ 79%] 2023-03-31T05:50:36.0657906Z test_jit.py::TestScript::test_tuple_unsortable_diff_type PASSED [ 79%] 2023-03-31T05:50:36.0658042Z test_jit.py::TestScript::test_tuple_unsortable_element_type PASSED [ 79%] 2023-03-31T05:50:36.0658189Z test_jit.py::TestScript::test_tuple_unsortable_nested_diff_type PASSED [ 79%] 2023-03-31T05:50:36.0658321Z test_jit.py::TestScript::test_type_annotate PASSED [ 79%] 2023-03-31T05:50:36.0658462Z test_jit.py::TestScript::test_type_annotation_module PASSED [ 79%] 2023-03-31T05:50:36.0658599Z test_jit.py::TestScript::test_type_annotation_py3 PASSED [ 79%] 2023-03-31T05:50:36.0658736Z test_jit.py::TestScript::test_type_annotations PASSED [ 79%] 2023-03-31T05:50:36.0658943Z test_jit.py::TestScript::test_type_annotations_repeated_list PASSED [ 79%] 2023-03-31T05:50:36.0659086Z test_jit.py::TestScript::test_type_annotations_varargs PASSED [ 79%] 2023-03-31T05:50:36.0659271Z test_jit.py::TestScript::test_type_call_in_script PASSED [ 79%] 2023-03-31T05:50:36.0659401Z test_jit.py::TestScript::test_type_cast PASSED [ 79%] 2023-03-31T05:50:36.0659541Z test_jit.py::TestScript::test_type_comments_in_body PASSED [ 79%] 2023-03-31T05:50:36.0659690Z test_jit.py::TestScript::test_type_inferred_from_empty_annotation PASSED [ 79%] 2023-03-31T05:50:36.0659936Z test_jit.py::TestScript::test_unbind SKIPPED (Profiling executor will be using different heuristics for constructing differentiable graphs) [ 79%] 2023-03-31T05:50:36.0660072Z test_jit.py::TestScript::test_unfold_zero_dim PASSED [ 79%] 2023-03-31T05:50:36.0660209Z test_jit.py::TestScript::test_unicode_comments PASSED [ 79%] 2023-03-31T05:50:36.0660353Z test_jit.py::TestScript::test_uninitialized PASSED [ 79%] 2023-03-31T05:50:36.0660477Z test_jit.py::TestScript::test_union_to_number PASSED [ 80%] 2023-03-31T05:50:36.0660615Z test_jit.py::TestScript::test_unknown_builtin PASSED [ 80%] 2023-03-31T05:50:36.0660758Z test_jit.py::TestScript::test_unmatched_type_annotation PASSED [ 80%] 2023-03-31T05:50:36.0660903Z test_jit.py::TestScript::test_unspecialized_any_binding PASSED [ 80%] 2023-03-31T05:50:36.0661093Z test_jit.py::TestScript::test_unsqueeze_guard_elimination SKIPPED (bailouts are being deprecated) [ 80%] 2023-03-31T05:50:36.0661238Z test_jit.py::TestScript::test_unsupported_builtin_error PASSED [ 80%] 2023-03-31T05:50:36.0661374Z test_jit.py::TestScript::test_unused_decorator PASSED [ 80%] 2023-03-31T05:50:36.0661519Z test_jit.py::TestScript::test_unwrap_optional_builtin PASSED [ 80%] 2023-03-31T05:50:36.0661644Z test_jit.py::TestScript::test_var_aug_assign PASSED [ 80%] 2023-03-31T05:50:36.0661775Z test_jit.py::TestScript::test_vararg_zeros PASSED [ 80%] 2023-03-31T05:50:36.0661925Z test_jit.py::TestScript::test_view_listconstruct_shape_prop PASSED [ 80%] 2023-03-31T05:50:36.0662061Z test_jit.py::TestScript::test_view_shape_prop PASSED [ 80%] 2023-03-31T05:50:36.0662192Z test_jit.py::TestScript::test_view_write PASSED [ 80%] 2023-03-31T05:50:36.0662328Z test_jit.py::TestScript::test_weak_cuda SKIPPED (no CUDA) [ 80%] 2023-03-31T05:50:36.0662459Z test_jit.py::TestScript::test_where PASSED [ 80%] 2023-03-31T05:50:36.0662592Z test_jit.py::TestScript::test_where_method PASSED [ 80%] 2023-03-31T05:50:36.0662706Z test_jit.py::TestScript::test_while PASSED [ 80%] 2023-03-31T05:50:36.0662839Z test_jit.py::TestScript::test_while_nest_if PASSED [ 80%] 2023-03-31T05:50:36.0663102Z test_jit.py::TestScript::test_while_nonexistent_cond_value PASSED [ 80%] 2023-03-31T05:50:36.0663254Z test_jit.py::TestScript::test_while_nonexistent_value PASSED [ 80%] 2023-03-31T05:50:36.0663397Z test_jit.py::TestScript::test_while_write_outer_then_read PASSED [ 80%] 2023-03-31T05:50:36.0663533Z test_jit.py::TestScript::test_wrong_attr_lookup PASSED [ 80%] 2023-03-31T05:50:36.0663674Z test_jit.py::TestScript::test_wrong_implicit_expand PASSED [ 80%] 2023-03-31T05:50:36.0663816Z test_jit.py::TestScript::test_wrong_method_call_inputs PASSED [ 80%] 2023-03-31T05:50:36.0663956Z test_jit.py::TestScript::test_wrong_module_attr_lookup PASSED [ 80%] 2023-03-31T05:50:36.0664080Z test_jit.py::TestScript::test_wrong_return_type PASSED [ 80%] 2023-03-31T05:50:36.0664220Z test_jit.py::TestScript::test_wrong_use_as_callable PASSED [ 81%] 2023-03-31T05:50:36.0664414Z test_jit.py::TestScript::test_wrong_use_as_tuple PASSED [ 81%] 2023-03-31T05:50:36.0664540Z test_jit.py::TestScript::test_zeros PASSED [ 81%] 2023-03-31T05:50:36.0664743Z test_jit.py::TestScript::test_zip_enumerate_modulelist PASSED [ 81%] 2023-03-31T05:50:36.0664899Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d PASSED [ 81%] 2023-03-31T05:50:36.0665067Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_no_batch_dim PASSED [ 81%] 2023-03-31T05:50:36.0665223Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_stride PASSED [ 81%] 2023-03-31T05:50:36.0665368Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool1d_stride_pad PASSED [ 81%] 2023-03-31T05:50:36.0665519Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d PASSED [ 81%] 2023-03-31T05:50:36.0665675Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor PASSED [ 81%] 2023-03-31T05:50:36.0665849Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor_stride PASSED [ 81%] 2023-03-31T05:50:36.0666022Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_divisor_stride_pad PASSED [ 81%] 2023-03-31T05:50:36.0666190Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_no_batch_dim PASSED [ 81%] 2023-03-31T05:50:36.0666345Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_stride PASSED [ 81%] 2023-03-31T05:50:36.0666502Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool2d_stride_pad PASSED [ 81%] 2023-03-31T05:50:36.0666640Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d PASSED [ 81%] 2023-03-31T05:50:36.0666797Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor PASSED [ 81%] 2023-03-31T05:50:36.0666966Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride PASSED [ 81%] 2023-03-31T05:50:36.0667153Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride1_pad0_gpu_input PASSED [ 81%] 2023-03-31T05:50:36.0667331Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad PASSED [ 81%] 2023-03-31T05:50:36.0667531Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output PASSED [ 81%] 2023-03-31T05:50:36.0667735Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_general_output PASSED [ 81%] 2023-03-31T05:50:36.0667935Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap PASSED [ 81%] 2023-03-31T05:50:36.0668087Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_no_batch_dim PASSED [ 81%] 2023-03-31T05:50:36.0668244Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride PASSED [ 81%] 2023-03-31T05:50:36.0668421Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride1_pad0_gpu_input PASSED [ 81%] 2023-03-31T05:50:36.0668583Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad PASSED [ 82%] 2023-03-31T05:50:36.0668769Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_fixedkw_output PASSED [ 82%] 2023-03-31T05:50:36.0668957Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_general_output PASSED [ 82%] 2023-03-31T05:50:36.0669153Z test_jit.py::TestJitGeneratedModule::test_nn_AvgPool3d_stride_pad_gpu_input_nooverlap PASSED [ 82%] 2023-03-31T05:50:36.0669306Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss PASSED [ 82%] 2023-03-31T05:50:36.0669476Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_mean PASSED [ 82%] 2023-03-31T05:50:36.0669627Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_none PASSED [ 82%] 2023-03-31T05:50:36.0669792Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_batch_dim_sum PASSED [ 82%] 2023-03-31T05:50:36.0669951Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_reduce PASSED [ 82%] 2023-03-31T05:50:36.0670116Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_no_reduce_scalar PASSED [ 82%] 2023-03-31T05:50:36.0670280Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_scalar_weights PASSED [ 82%] 2023-03-31T05:50:36.0670493Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights PASSED [ 82%] 2023-03-31T05:50:36.0670758Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights_no_reduce PASSED [ 82%] 2023-03-31T05:50:36.0670935Z test_jit.py::TestJitGeneratedModule::test_nn_BCELoss_weights_no_reduce_scalar PASSED [ 82%] 2023-03-31T05:50:36.0671087Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss PASSED [ 82%] 2023-03-31T05:50:36.0671265Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_legacy_enum PASSED [ 82%] 2023-03-31T05:50:36.0671452Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_mean PASSED [ 82%] 2023-03-31T05:50:36.0671639Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_none PASSED [ 82%] 2023-03-31T05:50:36.0671824Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_batch_dim_sum PASSED [ 82%] 2023-03-31T05:50:36.0672005Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_reduce PASSED [ 82%] 2023-03-31T05:50:36.0672189Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_no_reduce_scalar PASSED [ 82%] 2023-03-31T05:50:36.0672376Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_scalar_weights PASSED [ 82%] 2023-03-31T05:50:36.0672553Z test_jit.py::TestJitGeneratedModule::test_nn_BCEWithLogitsLoss_weights PASSED [ 82%] 2023-03-31T05:50:36.0672703Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_3d_input PASSED [ 82%] 2023-03-31T05:50:36.0672880Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_3d_input_not_affine PASSED [ 82%] 2023-03-31T05:50:36.0673039Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_affine PASSED [ 83%] 2023-03-31T05:50:36.0673221Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_affine_simple_average PASSED [ 83%] 2023-03-31T05:50:36.0673384Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_not_affine PASSED [ 83%] 2023-03-31T05:50:36.0673563Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_not_tracking_stats PASSED [ 83%] 2023-03-31T05:50:36.0673728Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm1d_zero_batch PASSED [ 83%] 2023-03-31T05:50:36.0673889Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d PASSED [ 83%] 2023-03-31T05:50:36.0674052Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_2d_simple_average PASSED [ 83%] 2023-03-31T05:50:36.0674213Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_momentum PASSED [ 83%] 2023-03-31T05:50:36.0674377Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_not_affine PASSED [ 83%] 2023-03-31T05:50:36.0674549Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_not_tracking_stats PASSED [ 83%] 2023-03-31T05:50:36.0674713Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm2d_zero_batch PASSED [ 83%] 2023-03-31T05:50:36.0674867Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d PASSED [ 83%] 2023-03-31T05:50:36.0675044Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_3d_simple_average PASSED [ 83%] 2023-03-31T05:50:36.0675204Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_momentum PASSED [ 83%] 2023-03-31T05:50:36.0675361Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_not_affine PASSED [ 83%] 2023-03-31T05:50:36.0675536Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_not_tracking_stats PASSED [ 83%] 2023-03-31T05:50:36.0675699Z test_jit.py::TestJitGeneratedModule::test_nn_BatchNorm3d_zero_batch PASSED [ 83%] 2023-03-31T05:50:36.0675852Z test_jit.py::TestJitGeneratedModule::test_nn_Bilinear PASSED [ 83%] 2023-03-31T05:50:36.0676009Z test_jit.py::TestJitGeneratedModule::test_nn_CELU_no_batch_dim PASSED [ 83%] 2023-03-31T05:50:36.0676228Z test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_int_target_lengths_intlists SKIPPED (module test skipped on JIT) [ 83%] 2023-03-31T05:50:36.0676410Z test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_int_target_lengths_tensors PASSED [ 83%] 2023-03-31T05:50:36.0676631Z test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_2d_lengths_tensors PASSED [ 83%] 2023-03-31T05:50:36.0676886Z test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_lengths_intlists SKIPPED (module test skipped on JIT) [ 83%] 2023-03-31T05:50:36.0677043Z test_jit.py::TestJitGeneratedModule::test_nn_CTCLoss_lengths_tensors PASSED [ 83%] 2023-03-31T05:50:36.0677194Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d PASSED [ 83%] 2023-03-31T05:50:36.0677365Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_circular_stride2_pad2 PASSED [ 83%] 2023-03-31T05:50:36.0677524Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_dilated PASSED [ 84%] 2023-03-31T05:50:36.0677679Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_groups PASSED [ 84%] 2023-03-31T05:50:36.0677832Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad1 PASSED [ 84%] 2023-03-31T05:50:36.0677988Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad1size1 PASSED [ 84%] 2023-03-31T05:50:36.0678146Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad2 PASSED [ 84%] 2023-03-31T05:50:36.0678293Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad2size1 PASSED [ 84%] 2023-03-31T05:50:36.0678445Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same PASSED [ 84%] 2023-03-31T05:50:36.0678599Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same2 PASSED [ 84%] 2023-03-31T05:50:36.0678765Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_same_dilated PASSED [ 84%] 2023-03-31T05:50:36.0678918Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_pad_valid PASSED [ 84%] 2023-03-31T05:50:36.0679090Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_reflect_stride2_pad2 PASSED [ 84%] 2023-03-31T05:50:36.0679264Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_replicate_stride2_pad2 PASSED [ 84%] 2023-03-31T05:50:36.0679415Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_stride PASSED [ 84%] 2023-03-31T05:50:36.0679564Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_zero_batch PASSED [ 84%] 2023-03-31T05:50:36.0679731Z test_jit.py::TestJitGeneratedModule::test_nn_Conv1d_zeros_stride2_pad2 PASSED [ 84%] 2023-03-31T05:50:36.0679882Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d PASSED [ 84%] 2023-03-31T05:50:36.0680055Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_circular_stride2_pad2 PASSED [ 84%] 2023-03-31T05:50:36.0680210Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise PASSED [ 84%] 2023-03-31T05:50:36.0680377Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_dilated PASSED [ 84%] 2023-03-31T05:50:36.0680543Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_padded PASSED [ 84%] 2023-03-31T05:50:36.0680710Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_strided PASSED [ 84%] 2023-03-31T05:50:36.0680877Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_depthwise_with_multiplier PASSED [ 84%] 2023-03-31T05:50:36.0681035Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_dilated PASSED [ 84%] 2023-03-31T05:50:36.0681189Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_groups PASSED [ 84%] 2023-03-31T05:50:36.0681350Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_groups_thnn PASSED [ 84%] 2023-03-31T05:50:36.0681500Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_no_bias PASSED [ 85%] 2023-03-31T05:50:36.0681653Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_same PASSED [ 85%] 2023-03-31T05:50:36.0681817Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_same_dilated PASSED [ 85%] 2023-03-31T05:50:36.0681972Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_pad_valid PASSED [ 85%] 2023-03-31T05:50:36.0682114Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_padding PASSED [ 85%] 2023-03-31T05:50:36.0682284Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_reflect_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0682508Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_replicate_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0682662Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_strided PASSED [ 85%] 2023-03-31T05:50:36.0682866Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_zero_batch PASSED [ 85%] 2023-03-31T05:50:36.0683195Z test_jit.py::TestJitGeneratedModule::test_nn_Conv2d_zeros_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0683348Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d PASSED [ 85%] 2023-03-31T05:50:36.0683505Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_1x1x1_no_bias PASSED [ 85%] 2023-03-31T05:50:36.0683677Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_circular_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0683817Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_dilated PASSED [ 85%] 2023-03-31T05:50:36.0683983Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_dilated_strided PASSED [ 85%] 2023-03-31T05:50:36.0684144Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_groups PASSED [ 85%] 2023-03-31T05:50:36.0684297Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_no_bias PASSED [ 85%] 2023-03-31T05:50:36.0684455Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_same PASSED [ 85%] 2023-03-31T05:50:36.0684622Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_same_dilated PASSED [ 85%] 2023-03-31T05:50:36.0684778Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_pad_valid PASSED [ 85%] 2023-03-31T05:50:36.0684952Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_replicate_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0685092Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_stride PASSED [ 85%] 2023-03-31T05:50:36.0685256Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_stride_padding PASSED [ 85%] 2023-03-31T05:50:36.0685412Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_zero_batch PASSED [ 85%] 2023-03-31T05:50:36.0685582Z test_jit.py::TestJitGeneratedModule::test_nn_Conv3d_zeros_stride2_pad2 PASSED [ 85%] 2023-03-31T05:50:36.0685745Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d PASSED [ 85%] 2023-03-31T05:50:36.0685917Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_dilated PASSED [ 86%] 2023-03-31T05:50:36.0686085Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_groups PASSED [ 86%] 2023-03-31T05:50:36.0686254Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose1d_no_bias PASSED [ 86%] 2023-03-31T05:50:36.0686400Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d PASSED [ 86%] 2023-03-31T05:50:36.0686568Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_dilated PASSED [ 86%] 2023-03-31T05:50:36.0686735Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_groups PASSED [ 86%] 2023-03-31T05:50:36.0686902Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose2d_no_bias PASSED [ 86%] 2023-03-31T05:50:36.0687063Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose3d PASSED [ 86%] 2023-03-31T05:50:36.0687235Z test_jit.py::TestJitGeneratedModule::test_nn_ConvTranspose3d_dilated PASSED [ 86%] 2023-03-31T05:50:36.0687399Z test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss PASSED [ 86%] 2023-03-31T05:50:36.0687580Z test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_margin PASSED [ 86%] 2023-03-31T05:50:36.0687757Z test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_mean PASSED [ 86%] 2023-03-31T05:50:36.0687944Z test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_none PASSED [ 86%] 2023-03-31T05:50:36.0688130Z test_jit.py::TestJitGeneratedModule::test_nn_CosineEmbeddingLoss_no_batch_dim_sum PASSED [ 86%] 2023-03-31T05:50:36.0688293Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss PASSED [ 86%] 2023-03-31T05:50:36.0688461Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d PASSED [ 86%] 2023-03-31T05:50:36.0688639Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_ignore_index PASSED [ 86%] 2023-03-31T05:50:36.0688953Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing PASSED [ 86%] 2023-03-31T05:50:36.0689236Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index PASSED [ 86%] 2023-03-31T05:50:36.0689452Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction PASSED [ 86%] 2023-03-31T05:50:36.0689647Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight PASSED [ 86%] 2023-03-31T05:50:36.0689826Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target PASSED [ 86%] 2023-03-31T05:50:36.0690019Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing PASSED [ 86%] 2023-03-31T05:50:36.0690227Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction PASSED [ 86%] 2023-03-31T05:50:36.0690434Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight PASSED [ 86%] 2023-03-31T05:50:36.0690622Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_prob_target_weights PASSED [ 86%] 2023-03-31T05:50:36.0690800Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_2d_weights PASSED [ 87%] 2023-03-31T05:50:36.0691000Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing PASSED [ 87%] 2023-03-31T05:50:36.0691210Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index PASSED [ 87%] 2023-03-31T05:50:36.0691409Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction PASSED [ 87%] 2023-03-31T05:50:36.0691635Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index PASSED [ 87%] 2023-03-31T05:50:36.0691812Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target PASSED [ 87%] 2023-03-31T05:50:36.0692007Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_smoothing PASSED [ 87%] 2023-03-31T05:50:36.0692220Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction PASSED [ 87%] 2023-03-31T05:50:36.0692409Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_3d_prob_target_weights PASSED [ 87%] 2023-03-31T05:50:36.0692587Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_4d_prob_target PASSED [ 87%] 2023-03-31T05:50:36.0692775Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_4d_prob_target_weights PASSED [ 87%] 2023-03-31T05:50:36.0692950Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_dim_is_3 PASSED [ 87%] 2023-03-31T05:50:36.0693112Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_higher_dim PASSED [ 87%] 2023-03-31T05:50:36.0693285Z test_jit.py::TestJitGeneratedModule::test_nn_CrossEntropyLoss_weights PASSED [ 87%] 2023-03-31T05:50:36.0693449Z test_jit.py::TestJitGeneratedModule::test_nn_CrossMapLRN2d PASSED [ 87%] 2023-03-31T05:50:36.0693606Z test_jit.py::TestJitGeneratedModule::test_nn_ELU_no_batch_dim PASSED [ 87%] 2023-03-31T05:50:36.0693767Z test_jit.py::TestJitGeneratedModule::test_nn_Embedding PASSED [ 87%] 2023-03-31T05:50:36.0693940Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_discontiguous PASSED [ 87%] 2023-03-31T05:50:36.0694100Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_max PASSED [ 87%] 2023-03-31T05:50:36.0694273Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_max_padding_idx PASSED [ 87%] 2023-03-31T05:50:36.0694419Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_mean PASSED [ 87%] 2023-03-31T05:50:36.0694594Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_mean_padding_idx PASSED [ 87%] 2023-03-31T05:50:36.0694754Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sparse PASSED [ 87%] 2023-03-31T05:50:36.0694977Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sum PASSED [ 87%] 2023-03-31T05:50:36.0695152Z test_jit.py::TestJitGeneratedModule::test_nn_EmbeddingBag_sum_padding_idx PASSED [ 87%] 2023-03-31T05:50:36.0695372Z test_jit.py::TestJitGeneratedModule::test_nn_Embedding_discontiguous PASSED [ 88%] 2023-03-31T05:50:36.0695536Z test_jit.py::TestJitGeneratedModule::test_nn_Embedding_sparse PASSED [ 88%] 2023-03-31T05:50:36.0695688Z test_jit.py::TestJitGeneratedModule::test_nn_Flatten PASSED [ 88%] 2023-03-31T05:50:36.0695834Z test_jit.py::TestJitGeneratedModule::test_nn_Flatten_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0695981Z test_jit.py::TestJitGeneratedModule::test_nn_Fold PASSED [ 88%] 2023-03-31T05:50:36.0696134Z test_jit.py::TestJitGeneratedModule::test_nn_Fold_int_input PASSED [ 88%] 2023-03-31T05:50:36.0696300Z test_jit.py::TestJitGeneratedModule::test_nn_Fold_no_batch_dim_input PASSED [ 88%] 2023-03-31T05:50:36.0696473Z test_jit.py::TestJitGeneratedModule::test_nn_Fold_no_batch_dim_int_input PASSED [ 88%] 2023-03-31T05:50:36.0696649Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio PASSED [ 88%] 2023-03-31T05:50:36.0696842Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0697055Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples PASSED [ 88%] 2023-03-31T05:50:36.0697247Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_ratio_return_indices PASSED [ 88%] 2023-03-31T05:50:36.0697408Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size PASSED [ 88%] 2023-03-31T05:50:36.0697594Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0697804Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples PASSED [ 88%] 2023-03-31T05:50:36.0697988Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_asymsize PASSED [ 88%] 2023-03-31T05:50:36.0698164Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio PASSED [ 88%] 2023-03-31T05:50:36.0698356Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0698566Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples PASSED [ 88%] 2023-03-31T05:50:36.0698756Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_ratio_return_indices PASSED [ 88%] 2023-03-31T05:50:36.0698918Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size PASSED [ 88%] 2023-03-31T05:50:36.0699106Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0699316Z test_jit.py::TestJitGeneratedModule::test_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples PASSED [ 88%] 2023-03-31T05:50:36.0699475Z test_jit.py::TestJitGeneratedModule::test_nn_GELU_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0699638Z test_jit.py::TestJitGeneratedModule::test_nn_GLU_no_batch_dim PASSED [ 88%] 2023-03-31T05:50:36.0699795Z test_jit.py::TestJitGeneratedModule::test_nn_GRUCell PASSED [ 88%] 2023-03-31T05:50:36.0699953Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine PASSED [ 89%] 2023-03-31T05:50:36.0700119Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine_GN PASSED [ 89%] 2023-03-31T05:50:36.0700297Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_affine_large_batch PASSED [ 89%] 2023-03-31T05:50:36.0700452Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_no_affine_IN PASSED [ 89%] 2023-03-31T05:50:36.0700615Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_1d_no_affine_LN PASSED [ 89%] 2023-03-31T05:50:36.0700779Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_affine PASSED [ 89%] 2023-03-31T05:50:36.0700956Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_affine_large_feature PASSED [ 89%] 2023-03-31T05:50:36.0701176Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_IN PASSED [ 89%] 2023-03-31T05:50:36.0701384Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_LN PASSED [ 89%] 2023-03-31T05:50:36.0701566Z test_jit.py::TestJitGeneratedModule::test_nn_GroupNorm_2d_no_affine_large_feature PASSED [ 89%] 2023-03-31T05:50:36.0701732Z test_jit.py::TestJitGeneratedModule::test_nn_Hardshrink_no_batch_dim PASSED [ 89%] 2023-03-31T05:50:36.0701925Z test_jit.py::TestJitGeneratedModule::test_nn_Hardsigmoid_no_batch_dim SKIPPED (module test skipped on JIT) [ 89%] 2023-03-31T05:50:36.0702128Z test_jit.py::TestJitGeneratedModule::test_nn_Hardswish_no_batch_dim SKIPPED (module test skipped on JIT) [ 89%] 2023-03-31T05:50:36.0702290Z test_jit.py::TestJitGeneratedModule::test_nn_Hardtanh_no_batch_dim PASSED [ 89%] 2023-03-31T05:50:36.0702459Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss PASSED [ 89%] 2023-03-31T05:50:36.0702638Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_margin PASSED [ 89%] 2023-03-31T05:50:36.0702823Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_margin_no_reduce PASSED [ 89%] 2023-03-31T05:50:36.0703120Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_mean PASSED [ 89%] 2023-03-31T05:50:36.0703312Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_none PASSED [ 89%] 2023-03-31T05:50:36.0703496Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_batch_dim_sum PASSED [ 89%] 2023-03-31T05:50:36.0703661Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_no_reduce PASSED [ 89%] 2023-03-31T05:50:36.0703844Z test_jit.py::TestJitGeneratedModule::test_nn_HingeEmbeddingLoss_scalar_margin PASSED [ 89%] 2023-03-31T05:50:36.0704000Z test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss PASSED [ 89%] 2023-03-31T05:50:36.0704161Z test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_delta PASSED [ 89%] 2023-03-31T05:50:36.0704337Z test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_mean PASSED [ 89%] 2023-03-31T05:50:36.0704509Z test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_none PASSED [ 90%] 2023-03-31T05:50:36.0704677Z test_jit.py::TestJitGeneratedModule::test_nn_HuberLoss_no_batch_dim_sum PASSED [ 90%] 2023-03-31T05:50:36.0704836Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d PASSED [ 90%] 2023-03-31T05:50:36.0704992Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0705169Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_tracking_stats PASSED [ 90%] 2023-03-31T05:50:36.0705359Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm1d_tracking_stats_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0705519Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d PASSED [ 90%] 2023-03-31T05:50:36.0705691Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0705871Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_tracking_stats PASSED [ 90%] 2023-03-31T05:50:36.0706067Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm2d_tracking_stats_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0706225Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d PASSED [ 90%] 2023-03-31T05:50:36.0706380Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0706552Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_tracking_stats PASSED [ 90%] 2023-03-31T05:50:36.0706740Z test_jit.py::TestJitGeneratedModule::test_nn_InstanceNorm3d_tracking_stats_no_batch_dim PASSED [ 90%] 2023-03-31T05:50:36.0706897Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss PASSED [ 90%] 2023-03-31T05:50:36.0707059Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_log_target PASSED [ 90%] 2023-03-31T05:50:36.0707288Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_mean PASSED [ 90%] 2023-03-31T05:50:36.0707455Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_none PASSED [ 90%] 2023-03-31T05:50:36.0707705Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_batch_dim_sum PASSED [ 90%] 2023-03-31T05:50:36.0707864Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce PASSED [ 90%] 2023-03-31T05:50:36.0708026Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_log_target PASSED [ 90%] 2023-03-31T05:50:36.0708195Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_scalar PASSED [ 90%] 2023-03-31T05:50:36.0708376Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_no_reduce_scalar_log_target PASSED [ 90%] 2023-03-31T05:50:36.0708535Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_scalar PASSED [ 90%] 2023-03-31T05:50:36.0708706Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_scalar_log_target PASSED [ 90%] 2023-03-31T05:50:36.0708889Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_with_log_target_no_reduce PASSED [ 90%] 2023-03-31T05:50:36.0709063Z test_jit.py::TestJitGeneratedModule::test_nn_KLDivLoss_with_target_no_reduce PASSED [ 91%] 2023-03-31T05:50:36.0709218Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss PASSED [ 91%] 2023-03-31T05:50:36.0709373Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_mean PASSED [ 91%] 2023-03-31T05:50:36.0709538Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_none PASSED [ 91%] 2023-03-31T05:50:36.0709702Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_batch_dim_sum PASSED [ 91%] 2023-03-31T05:50:36.0709859Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce PASSED [ 91%] 2023-03-31T05:50:36.0710025Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce_complex PASSED [ 91%] 2023-03-31T05:50:36.0710187Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_no_reduce_scalar PASSED [ 91%] 2023-03-31T05:50:36.0710347Z test_jit.py::TestJitGeneratedModule::test_nn_L1Loss_scalar PASSED [ 91%] 2023-03-31T05:50:36.0710499Z test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d PASSED [ 91%] 2023-03-31T05:50:36.0710656Z test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d_no_batch_dim PASSED [ 91%] 2023-03-31T05:50:36.0710812Z test_jit.py::TestJitGeneratedModule::test_nn_LPPool1d_norm PASSED [ 91%] 2023-03-31T05:50:36.0710961Z test_jit.py::TestJitGeneratedModule::test_nn_LPPool2d PASSED [ 91%] 2023-03-31T05:50:36.0711117Z test_jit.py::TestJitGeneratedModule::test_nn_LPPool2d_norm PASSED [ 91%] 2023-03-31T05:50:36.0711267Z test_jit.py::TestJitGeneratedModule::test_nn_LSTMCell PASSED [ 91%] 2023-03-31T05:50:36.0711445Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_elementwise_affine PASSED [ 91%] 2023-03-31T05:50:36.0711631Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_empty_elementwise_affine PASSED [ 91%] 2023-03-31T05:50:36.0711814Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_1d_no_elementwise_affine PASSED [ 91%] 2023-03-31T05:50:36.0711988Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_elementwise_affine PASSED [ 91%] 2023-03-31T05:50:36.0712161Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_no_affine_large_feature PASSED [ 91%] 2023-03-31T05:50:36.0712342Z test_jit.py::TestJitGeneratedModule::test_nn_LayerNorm_3d_no_elementwise_affine PASSED [ 91%] 2023-03-31T05:50:36.0712506Z test_jit.py::TestJitGeneratedModule::test_nn_LeakyReLU_no_batch_dim PASSED [ 91%] 2023-03-31T05:50:36.0712656Z test_jit.py::TestJitGeneratedModule::test_nn_Linear PASSED [ 91%] 2023-03-31T05:50:36.0712815Z test_jit.py::TestJitGeneratedModule::test_nn_Linear_no_batch_dim PASSED [ 91%] 2023-03-31T05:50:36.0712972Z test_jit.py::TestJitGeneratedModule::test_nn_Linear_no_bias PASSED [ 91%] 2023-03-31T05:50:36.0713136Z test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_1d PASSED [ 91%] 2023-03-31T05:50:36.0713367Z test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_2d_uneven_pad PASSED [ 92%] 2023-03-31T05:50:36.0713538Z test_jit.py::TestJitGeneratedModule::test_nn_LocalResponseNorm_3d_custom_params PASSED [ 92%] 2023-03-31T05:50:36.0713774Z test_jit.py::TestJitGeneratedModule::test_nn_LogSigmoid_no_batch_dim PASSED [ 92%] 2023-03-31T05:50:36.0713927Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss PASSED [ 92%] 2023-03-31T05:50:36.0714094Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_mean PASSED [ 92%] 2023-03-31T05:50:36.0714257Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_none PASSED [ 92%] 2023-03-31T05:50:36.0714421Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_batch_dim_sum PASSED [ 92%] 2023-03-31T05:50:36.0714579Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_reduce PASSED [ 92%] 2023-03-31T05:50:36.0714745Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_no_reduce_scalar PASSED [ 92%] 2023-03-31T05:50:36.0714893Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_prec PASSED [ 92%] 2023-03-31T05:50:36.0715049Z test_jit.py::TestJitGeneratedModule::test_nn_MSELoss_scalar PASSED [ 92%] 2023-03-31T05:50:36.0715214Z test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss PASSED [ 92%] 2023-03-31T05:50:36.0715387Z test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_margin PASSED [ 92%] 2023-03-31T05:50:36.0715570Z test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_mean PASSED [ 92%] 2023-03-31T05:50:36.0715753Z test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_none PASSED [ 92%] 2023-03-31T05:50:36.0715934Z test_jit.py::TestJitGeneratedModule::test_nn_MarginRankingLoss_no_batch_dim_sum PASSED [ 92%] 2023-03-31T05:50:36.0716088Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d PASSED [ 92%] 2023-03-31T05:50:36.0716244Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d_return_indices PASSED [ 92%] 2023-03-31T05:50:36.0716406Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool1d_stride PASSED [ 92%] 2023-03-31T05:50:36.0716564Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_3d_input PASSED [ 92%] 2023-03-31T05:50:36.0716726Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_4d_input PASSED [ 92%] 2023-03-31T05:50:36.0716892Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool2d_return_indices PASSED [ 92%] 2023-03-31T05:50:36.0717043Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d PASSED [ 92%] 2023-03-31T05:50:36.0717210Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_return_indices PASSED [ 92%] 2023-03-31T05:50:36.0717368Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_stride PASSED [ 92%] 2023-03-31T05:50:36.0717535Z test_jit.py::TestJitGeneratedModule::test_nn_MaxPool3d_stride_padding PASSED [ 93%] 2023-03-31T05:50:36.0717679Z test_jit.py::TestJitGeneratedModule::test_nn_Mish_no_batch_dim PASSED [ 93%] 2023-03-31T05:50:36.0717849Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss PASSED [ 93%] 2023-03-31T05:50:36.0718031Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_0d_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0718208Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_1d PASSED [ 93%] 2023-03-31T05:50:36.0718391Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_1d_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0718571Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_index_neg PASSED [ 93%] 2023-03-31T05:50:36.0718759Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_mean PASSED [ 93%] 2023-03-31T05:50:36.0718949Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_none PASSED [ 93%] 2023-03-31T05:50:36.0719125Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_batch_dim_sum PASSED [ 93%] 2023-03-31T05:50:36.0719306Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelMarginLoss_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0719586Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss PASSED [ 93%] 2023-03-31T05:50:36.0719836Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean PASSED [ 93%] 2023-03-31T05:50:36.0720033Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_none PASSED [ 93%] 2023-03-31T05:50:36.0720232Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum PASSED [ 93%] 2023-03-31T05:50:36.0720422Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0720612Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_weights PASSED [ 93%] 2023-03-31T05:50:36.0720810Z test_jit.py::TestJitGeneratedModule::test_nn_MultiLabelSoftMarginLoss_weights_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0720958Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss PASSED [ 93%] 2023-03-31T05:50:36.0721126Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_1d PASSED [ 93%] 2023-03-31T05:50:36.0721300Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_1d_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0721474Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_margin PASSED [ 93%] 2023-03-31T05:50:36.0721653Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_margin_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0721827Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0721989Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_p PASSED [ 93%] 2023-03-31T05:50:36.0722161Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_p_no_reduce PASSED [ 93%] 2023-03-31T05:50:36.0722319Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_weights PASSED [ 94%] 2023-03-31T05:50:36.0722499Z test_jit.py::TestJitGeneratedModule::test_nn_MultiMarginLoss_weights_no_reduce PASSED [ 94%] 2023-03-31T05:50:36.0722745Z test_jit.py::TestJitGeneratedModule::test_nn_MultiheadAttention SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 94%] 2023-03-31T05:50:36.0722902Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss PASSED [ 94%] 2023-03-31T05:50:36.0723219Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce PASSED [ 94%] 2023-03-31T05:50:36.0723394Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0723565Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss2d_no_reduce_weights PASSED [ 94%] 2023-03-31T05:50:36.0723725Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce PASSED [ 94%] 2023-03-31T05:50:36.0723903Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0724062Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLossNd_no_reduce_weights PASSED [ 94%] 2023-03-31T05:50:36.0724221Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d PASSED [ 94%] 2023-03-31T05:50:36.0724387Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0724547Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_2d_weights PASSED [ 94%] 2023-03-31T05:50:36.0724704Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_dim_is_3 PASSED [ 94%] 2023-03-31T05:50:36.0724864Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_higher_dim PASSED [ 94%] 2023-03-31T05:50:36.0725024Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0725192Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_mean PASSED [ 94%] 2023-03-31T05:50:36.0725348Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_none PASSED [ 94%] 2023-03-31T05:50:36.0725514Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_batch_dim_sum PASSED [ 94%] 2023-03-31T05:50:36.0725675Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce PASSED [ 94%] 2023-03-31T05:50:36.0725939Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0726169Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights PASSED [ 94%] 2023-03-31T05:50:36.0726355Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights_ignore_index PASSED [ 94%] 2023-03-31T05:50:36.0726544Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_no_reduce_weights_ignore_index_neg PASSED [ 94%] 2023-03-31T05:50:36.0726702Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights PASSED [ 94%] 2023-03-31T05:50:36.0726860Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights_ignore_index PASSED [ 95%] 2023-03-31T05:50:36.0727034Z test_jit.py::TestJitGeneratedModule::test_nn_NLLLoss_weights_ignore_index_neg PASSED [ 95%] 2023-03-31T05:50:36.0727195Z test_jit.py::TestJitGeneratedModule::test_nn_PReLU_no_batch_dim PASSED [ 95%] 2023-03-31T05:50:36.0727367Z test_jit.py::TestJitGeneratedModule::test_nn_Padding122112_3dcircular PASSED [ 95%] 2023-03-31T05:50:36.0727533Z test_jit.py::TestJitGeneratedModule::test_nn_Padding1221_2dcircular PASSED [ 95%] 2023-03-31T05:50:36.0727700Z test_jit.py::TestJitGeneratedModule::test_nn_Padding12_1dcircular PASSED [ 95%] 2023-03-31T05:50:36.0727864Z test_jit.py::TestJitGeneratedModule::test_nn_Padding2322_2dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728025Z test_jit.py::TestJitGeneratedModule::test_nn_Padding31_1dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728177Z test_jit.py::TestJitGeneratedModule::test_nn_Padding322112_3dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728342Z test_jit.py::TestJitGeneratedModule::test_nn_Padding332122_3dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728503Z test_jit.py::TestJitGeneratedModule::test_nn_Padding3331_2dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728665Z test_jit.py::TestJitGeneratedModule::test_nn_Padding33_1dcircular PASSED [ 95%] 2023-03-31T05:50:36.0728827Z test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance PASSED [ 95%] 2023-03-31T05:50:36.0729009Z test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_broadcast_lhs PASSED [ 95%] 2023-03-31T05:50:36.0729192Z test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_broadcast_rhs PASSED [ 95%] 2023-03-31T05:50:36.0729368Z test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_no_batch_dim PASSED [ 95%] 2023-03-31T05:50:36.0729555Z test_jit.py::TestJitGeneratedModule::test_nn_PairwiseDistance_with_non_default_args PASSED [ 95%] 2023-03-31T05:50:36.0729702Z test_jit.py::TestJitGeneratedModule::test_nn_PixelShuffle PASSED [ 95%] 2023-03-31T05:50:36.0729863Z test_jit.py::TestJitGeneratedModule::test_nn_PixelUnshuffle PASSED [ 95%] 2023-03-31T05:50:36.0730032Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_full_loss PASSED [ 95%] 2023-03-31T05:50:36.0730218Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_full_loss_no_log_input PASSED [ 95%] 2023-03-31T05:50:36.0730400Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_mean PASSED [ 95%] 2023-03-31T05:50:36.0730576Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_none PASSED [ 95%] 2023-03-31T05:50:36.0730755Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_batch_dim_sum PASSED [ 95%] 2023-03-31T05:50:36.0730929Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_full_loss PASSED [ 95%] 2023-03-31T05:50:36.0731104Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_full_loss_no_log_input PASSED [ 95%] 2023-03-31T05:50:36.0731273Z test_jit.py::TestJitGeneratedModule::test_nn_PoissonNLLLoss_no_reduce PASSED [ 96%] 2023-03-31T05:50:36.0731424Z test_jit.py::TestJitGeneratedModule::test_nn_RNNCell PASSED [ 96%] 2023-03-31T05:50:36.0731575Z test_jit.py::TestJitGeneratedModule::test_nn_RReLU PASSED [ 96%] 2023-03-31T05:50:36.0731732Z test_jit.py::TestJitGeneratedModule::test_nn_RReLU_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0731942Z test_jit.py::TestJitGeneratedModule::test_nn_RReLU_with_up_down PASSED [ 96%] 2023-03-31T05:50:36.0732112Z test_jit.py::TestJitGeneratedModule::test_nn_RReLU_with_up_down_scalar PASSED [ 96%] 2023-03-31T05:50:36.0732326Z test_jit.py::TestJitGeneratedModule::test_nn_ReLU6_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0732470Z test_jit.py::TestJitGeneratedModule::test_nn_ReLU_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0732632Z test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d PASSED [ 96%] 2023-03-31T05:50:36.0732802Z test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d_complex PASSED [ 96%] 2023-03-31T05:50:36.0732978Z test_jit.py::TestJitGeneratedModule::test_nn_ReplicationPad3d_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0733133Z test_jit.py::TestJitGeneratedModule::test_nn_SELU_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0733289Z test_jit.py::TestJitGeneratedModule::test_nn_SiLU_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0733450Z test_jit.py::TestJitGeneratedModule::test_nn_Sigmoid_no_batch_dim PASSED [ 96%] 2023-03-31T05:50:36.0733604Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss PASSED [ 96%] 2023-03-31T05:50:36.0733754Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_beta PASSED [ 96%] 2023-03-31T05:50:36.0733928Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_mean PASSED [ 96%] 2023-03-31T05:50:36.0734097Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_none PASSED [ 96%] 2023-03-31T05:50:36.0734269Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_batch_dim_sum PASSED [ 96%] 2023-03-31T05:50:36.0734433Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_reduce PASSED [ 96%] 2023-03-31T05:50:36.0734603Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_no_reduce_scalar PASSED [ 96%] 2023-03-31T05:50:36.0734766Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_scalar PASSED [ 96%] 2023-03-31T05:50:36.0734934Z test_jit.py::TestJitGeneratedModule::test_nn_SmoothL1Loss_zero_beta PASSED [ 96%] 2023-03-31T05:50:36.0735092Z test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss PASSED [ 96%] 2023-03-31T05:50:36.0735259Z test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_mean PASSED [ 96%] 2023-03-31T05:50:36.0735435Z test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_none PASSED [ 96%] 2023-03-31T05:50:36.0735609Z test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_batch_dim_sum PASSED [ 97%] 2023-03-31T05:50:36.0735778Z test_jit.py::TestJitGeneratedModule::test_nn_SoftMarginLoss_no_reduce PASSED [ 97%] 2023-03-31T05:50:36.0735942Z test_jit.py::TestJitGeneratedModule::test_nn_Softplus_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736107Z test_jit.py::TestJitGeneratedModule::test_nn_Softshrink_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736270Z test_jit.py::TestJitGeneratedModule::test_nn_Softsign_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736429Z test_jit.py::TestJitGeneratedModule::test_nn_Tanh_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736583Z test_jit.py::TestJitGeneratedModule::test_nn_Tanhshrink_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736752Z test_jit.py::TestJitGeneratedModule::test_nn_Threshold_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0736980Z test_jit.py::TestJitGeneratedModule::test_nn_Transformer SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 97%] 2023-03-31T05:50:36.0737179Z test_jit.py::TestJitGeneratedModule::test_nn_TransformerDecoderLayer_gelu_activation PASSED [ 97%] 2023-03-31T05:50:36.0737372Z test_jit.py::TestJitGeneratedModule::test_nn_TransformerDecoderLayer_relu_activation PASSED [ 97%] 2023-03-31T05:50:36.0737562Z test_jit.py::TestJitGeneratedModule::test_nn_TransformerEncoderLayer_gelu_activation PASSED [ 97%] 2023-03-31T05:50:36.0737751Z test_jit.py::TestJitGeneratedModule::test_nn_TransformerEncoderLayer_relu_activation PASSED [ 97%] 2023-03-31T05:50:36.0737924Z test_jit.py::TestJitGeneratedModule::test_nn_Transformer_multilayer_coder PASSED [ 97%] 2023-03-31T05:50:36.0738172Z test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_mean PASSED [ 97%] 2023-03-31T05:50:36.0738393Z test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_none PASSED [ 97%] 2023-03-31T05:50:36.0738575Z test_jit.py::TestJitGeneratedModule::test_nn_TripletMarginLoss_no_batch_dim_sum PASSED [ 97%] 2023-03-31T05:50:36.0738739Z test_jit.py::TestJitGeneratedModule::test_nn_Unflatten_no_batch_dim PASSED [ 97%] 2023-03-31T05:50:36.0738889Z test_jit.py::TestJitGeneratedModule::test_nn_Unfold PASSED [ 97%] 2023-03-31T05:50:36.0739048Z test_jit.py::TestJitGeneratedModule::test_nn_Unfold_int_input PASSED [ 97%] 2023-03-31T05:50:36.0739213Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_2d PASSED [ 97%] 2023-03-31T05:50:36.0739388Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_2d_zero_dim PASSED [ 97%] 2023-03-31T05:50:36.0739567Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_2d PASSED [ 97%] 2023-03-31T05:50:36.0739744Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_shared_2d PASSED [ 97%] 2023-03-31T05:50:36.0739938Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_skewed_2d PASSED [ 97%] 2023-03-31T05:50:36.0740145Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_scale_tuple_skewed_2d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0740316Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_tuple_2d PASSED [ 98%] 2023-03-31T05:50:36.0740509Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bicubic_tuple_2d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0740676Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_2d PASSED [ 98%] 2023-03-31T05:50:36.0740853Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_2d_zero_dim PASSED [ 98%] 2023-03-31T05:50:36.0741026Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_2d PASSED [ 98%] 2023-03-31T05:50:36.0741223Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_shared_2d PASSED [ 98%] 2023-03-31T05:50:36.0741402Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_skewed_2d PASSED [ 98%] 2023-03-31T05:50:36.0741608Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_scale_tuple_skewed_2d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0741780Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_tuple_2d PASSED [ 98%] 2023-03-31T05:50:36.0741972Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_bilinear_tuple_2d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0742137Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d PASSED [ 98%] 2023-03-31T05:50:36.0742316Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0742492Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_1d_zero_dim PASSED [ 98%] 2023-03-31T05:50:36.0742667Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_scale_1d PASSED [ 98%] 2023-03-31T05:50:36.0742846Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_scale_1d_align_corners PASSED [ 98%] 2023-03-31T05:50:36.0743119Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_linear_tuple_1d PASSED [ 98%] 2023-03-31T05:50:36.0743291Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_1d PASSED [ 98%] 2023-03-31T05:50:36.0743467Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_1d_zero_dim PASSED [ 98%] 2023-03-31T05:50:36.0743632Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d PASSED [ 98%] 2023-03-31T05:50:36.0743816Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d_launch_configs PASSED [ 98%] 2023-03-31T05:50:36.0743990Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_2d_zero_dim PASSED [ 98%] 2023-03-31T05:50:36.0744216Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_3d PASSED [ 98%] 2023-03-31T05:50:36.0744388Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_3d_zero_dim PASSED [ 98%] 2023-03-31T05:50:36.0744602Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_1d PASSED [ 98%] 2023-03-31T05:50:36.0744775Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_2d PASSED [ 98%] 2023-03-31T05:50:36.0744950Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_scale_3d PASSED [ 99%] 2023-03-31T05:50:36.0745121Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_1d PASSED [ 99%] 2023-03-31T05:50:36.0745291Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_2d PASSED [ 99%] 2023-03-31T05:50:36.0745463Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_nearest_tuple_3d PASSED [ 99%] 2023-03-31T05:50:36.0745630Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_3d PASSED [ 99%] 2023-03-31T05:50:36.0745812Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_3d_zero_dim PASSED [ 99%] 2023-03-31T05:50:36.0745972Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_scale_3d PASSED [ 99%] 2023-03-31T05:50:36.0746170Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_scale_3d_align_corners PASSED [ 99%] 2023-03-31T05:50:36.0746345Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_tuple_3d PASSED [ 99%] 2023-03-31T05:50:36.0746538Z test_jit.py::TestJitGeneratedModule::test_nn_interpolate_trilinear_tuple_3d_align_corners PASSED [ 99%] 2023-03-31T05:50:36.0746698Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_dim0 PASSED [ 99%] 2023-03-31T05:50:36.0746853Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_dim3 PASSED [ 99%] 2023-03-31T05:50:36.0747017Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_lastdim PASSED [ 99%] 2023-03-31T05:50:36.0747174Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_scalar PASSED [ 99%] 2023-03-31T05:50:36.0747338Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_spatial PASSED [ 99%] 2023-03-31T05:50:36.0747503Z test_jit.py::TestJitGeneratedModule::test_nn_log_softmax_spatial_special PASSED [ 99%] 2023-03-31T05:50:36.0747695Z test_jit.py::TestJitGeneratedModule::test_nn_multimarginloss_1d_input_0d_target_no_reduce PASSED [ 99%] 2023-03-31T05:50:36.0747864Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_dim0 PASSED [ 99%] 2023-03-31T05:50:36.0748031Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_dim3 PASSED [ 99%] 2023-03-31T05:50:36.0748199Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_functional_scalar PASSED [ 99%] 2023-03-31T05:50:36.0748356Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_lastdim PASSED [ 99%] 2023-03-31T05:50:36.0748520Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_lastdim_dtype PASSED [ 99%] 2023-03-31T05:50:36.0748671Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial PASSED [ 99%] 2023-03-31T05:50:36.0748842Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial_dtype PASSED [ 99%] 2023-03-31T05:50:36.0749025Z test_jit.py::TestJitGeneratedModule::test_nn_softmax_spatial_special PASSED [ 99%] 2023-03-31T05:50:36.0749219Z test_jit.py::TestProducerVersion::test_version PASSED [100%]CUDA not available, skipping tests 2023-03-31T05:50:36.0749521Z monkeytype is not installed. Skipping tests for Profile-Directed Typing 2023-03-31T05:50:36.0749533Z 2023-03-31T05:50:36.0749538Z 2023-03-31T05:50:36.0749863Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit/test_jit-47d330549a92143e.xml - 2023-03-31T05:50:36.0749996Z ========== 2430 passed, 119 skipped, 11 xfailed in 116.22s (0:01:56) =========== 2023-03-31T05:50:36.0750179Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:50:36.0750185Z 2023-03-31T05:50:36.0750666Z ##[endgroup] 2023-03-31T05:50:36.0750935Z FINISHED PRINTING LOG FILE of test_jit (/var/lib/jenkins/workspace/test/test-reports/test_jit_144gxr3v.log) 2023-03-31T05:50:36.0750941Z 2023-03-31T05:50:38.0302269Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:50:38.0329519Z Ignoring disabled issues: [] 2023-03-31T05:50:38.0492934Z Running test_foreach ... [2023-03-31 05:50:38.048976] 2023-03-31T05:50:38.0495623Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_foreach.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:50:38.049335] 2023-03-31T05:52:26.3040825Z 2023-03-31T05:52:26.3041354Z Expand the folded group to see the log file of test_foreach 2023-03-31T05:52:26.3042325Z ##[group]PRINTING LOG FILE of test_foreach (/var/lib/jenkins/workspace/test/test-reports/test_foreach_twhugs1k.log) 2023-03-31T05:52:26.3064804Z Test results will be stored in test-reports/python-pytest/test_foreach/test_foreach-b6d6f436a62737dc.xml 2023-03-31T05:52:26.3081061Z ============================= test session starts ============================== 2023-03-31T05:52:26.3081919Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:52:26.3082447Z cachedir: .pytest_cache 2023-03-31T05:52:26.3083363Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:52:26.3084133Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:52:26.3124580Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:52:26.3125020Z collecting ... collected 1055 items 2023-03-31T05:52:26.3283724Z Running 1055 items in this shard: test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_bool, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float16, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float32, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float64, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int16, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int32, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int64, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int8, test/test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_bool, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int16, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int32, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int64, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int8, test/test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_foreach_l2_large_value_input__foreach_norm_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_foreach_l2_large_value_input__foreach_norm_cpu_float16, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float16, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float32, test/test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float64, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_float32, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_float64, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_float32, test/test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cpu_float64, test/test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cpu_float32, test/test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cpu_float64, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_uint8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_complex128, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_complex64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_bfloat16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_float32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_float64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int16, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int32, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int64, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int8, test/test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_uint8 2023-03-31T05:52:26.3402990Z 2023-03-31T05:52:26.3404024Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_bfloat16 PASSED [ 0%] 2023-03-31T05:52:26.3404470Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_bool PASSED [ 0%] 2023-03-31T05:52:26.3404883Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_complex128 PASSED [ 0%] 2023-03-31T05:52:26.3405294Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_complex64 PASSED [ 0%] 2023-03-31T05:52:26.3405707Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float16 PASSED [ 0%] 2023-03-31T05:52:26.3406095Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float32 PASSED [ 0%] 2023-03-31T05:52:26.3406492Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_float64 PASSED [ 0%] 2023-03-31T05:52:26.3406888Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int16 PASSED [ 0%] 2023-03-31T05:52:26.3407279Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int32 PASSED [ 0%] 2023-03-31T05:52:26.3407664Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int64 PASSED [ 0%] 2023-03-31T05:52:26.3408051Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_int8 PASSED [ 1%] 2023-03-31T05:52:26.3408442Z test_foreach.py::TestForeachCPU::test_add_scalar_with_empty_list_and_empty_tensor_cpu_uint8 PASSED [ 1%] 2023-03-31T05:52:26.3408831Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_bfloat16 PASSED [ 1%] 2023-03-31T05:52:26.3409215Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_bool PASSED [ 1%] 2023-03-31T05:52:26.3409613Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_complex128 PASSED [ 1%] 2023-03-31T05:52:26.3410016Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_complex64 PASSED [ 1%] 2023-03-31T05:52:26.3410693Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float16 PASSED [ 1%] 2023-03-31T05:52:26.3411202Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float32 PASSED [ 1%] 2023-03-31T05:52:26.3411596Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_float64 PASSED [ 1%] 2023-03-31T05:52:26.3411994Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int16 PASSED [ 1%] 2023-03-31T05:52:26.3412369Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int32 PASSED [ 1%] 2023-03-31T05:52:26.3412761Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int64 PASSED [ 2%] 2023-03-31T05:52:26.3413146Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_int8 PASSED [ 2%] 2023-03-31T05:52:26.3413530Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_False_cpu_uint8 PASSED [ 2%] 2023-03-31T05:52:26.3413913Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_bfloat16 PASSED [ 2%] 2023-03-31T05:52:26.3414309Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_bool PASSED [ 2%] 2023-03-31T05:52:26.3414709Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_complex128 PASSED [ 2%] 2023-03-31T05:52:26.3415110Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_complex64 PASSED [ 2%] 2023-03-31T05:52:26.3415489Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float16 PASSED [ 2%] 2023-03-31T05:52:26.3415884Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float32 PASSED [ 2%] 2023-03-31T05:52:26.3416268Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_float64 PASSED [ 2%] 2023-03-31T05:52:26.3416639Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int16 PASSED [ 3%] 2023-03-31T05:52:26.3417027Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int32 PASSED [ 3%] 2023-03-31T05:52:26.3417415Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int64 PASSED [ 3%] 2023-03-31T05:52:26.3417793Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_int8 PASSED [ 3%] 2023-03-31T05:52:26.3418163Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_add_is_fastpath_True_cpu_uint8 PASSED [ 3%] 2023-03-31T05:52:26.3418559Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_bfloat16 PASSED [ 3%] 2023-03-31T05:52:26.3418966Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_float32 PASSED [ 3%] 2023-03-31T05:52:26.3419365Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_float64 PASSED [ 3%] 2023-03-31T05:52:26.3419759Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int16 PASSED [ 3%] 2023-03-31T05:52:26.3420161Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int32 PASSED [ 3%] 2023-03-31T05:52:26.3420557Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int64 PASSED [ 3%] 2023-03-31T05:52:26.3420950Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_int8 PASSED [ 4%] 2023-03-31T05:52:26.3421335Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_False_cpu_uint8 PASSED [ 4%] 2023-03-31T05:52:26.3421738Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_bfloat16 PASSED [ 4%] 2023-03-31T05:52:26.3422142Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_float32 PASSED [ 4%] 2023-03-31T05:52:26.3422611Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_float64 PASSED [ 4%] 2023-03-31T05:52:26.3435919Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int16 PASSED [ 4%] 2023-03-31T05:52:26.3436653Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int32 PASSED [ 4%] 2023-03-31T05:52:26.3437055Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int64 PASSED [ 4%] 2023-03-31T05:52:26.3437458Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_int8 PASSED [ 4%] 2023-03-31T05:52:26.3437856Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_max_is_fastpath_True_cpu_uint8 PASSED [ 4%] 2023-03-31T05:52:26.3438249Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_bfloat16 PASSED [ 5%] 2023-03-31T05:52:26.3438666Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_float32 PASSED [ 5%] 2023-03-31T05:52:26.3439083Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_float64 PASSED [ 5%] 2023-03-31T05:52:26.3439497Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int16 PASSED [ 5%] 2023-03-31T05:52:26.3439883Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int32 PASSED [ 5%] 2023-03-31T05:52:26.3440288Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int64 PASSED [ 5%] 2023-03-31T05:52:26.3440685Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_int8 PASSED [ 5%] 2023-03-31T05:52:26.3441081Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_False_cpu_uint8 PASSED [ 5%] 2023-03-31T05:52:26.3441468Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_bfloat16 PASSED [ 5%] 2023-03-31T05:52:26.3441872Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_float32 PASSED [ 5%] 2023-03-31T05:52:26.3442272Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_float64 PASSED [ 5%] 2023-03-31T05:52:26.3442680Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int16 PASSED [ 6%] 2023-03-31T05:52:26.3443246Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int32 PASSED [ 6%] 2023-03-31T05:52:26.3443662Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int64 PASSED [ 6%] 2023-03-31T05:52:26.3444052Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_int8 PASSED [ 6%] 2023-03-31T05:52:26.3444430Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_clamp_min_is_fastpath_True_cpu_uint8 PASSED [ 6%] 2023-03-31T05:52:26.3444822Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_bfloat16 PASSED [ 6%] 2023-03-31T05:52:26.3445223Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_bool PASSED [ 6%] 2023-03-31T05:52:26.3445624Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_complex128 PASSED [ 6%] 2023-03-31T05:52:26.3446009Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_complex64 PASSED [ 6%] 2023-03-31T05:52:26.3446404Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float16 PASSED [ 6%] 2023-03-31T05:52:26.3446800Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float32 PASSED [ 7%] 2023-03-31T05:52:26.3447190Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_float64 PASSED [ 7%] 2023-03-31T05:52:26.3447565Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int16 PASSED [ 7%] 2023-03-31T05:52:26.3448098Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int32 PASSED [ 7%] 2023-03-31T05:52:26.3448483Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int64 PASSED [ 7%] 2023-03-31T05:52:26.3448969Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_int8 PASSED [ 7%] 2023-03-31T05:52:26.3449342Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_False_cpu_uint8 PASSED [ 7%] 2023-03-31T05:52:26.3449730Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_bfloat16 PASSED [ 7%] 2023-03-31T05:52:26.3450117Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_bool PASSED [ 7%] 2023-03-31T05:52:26.3450495Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_complex128 PASSED [ 7%] 2023-03-31T05:52:26.3450891Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_complex64 PASSED [ 7%] 2023-03-31T05:52:26.3451287Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float16 PASSED [ 8%] 2023-03-31T05:52:26.3451684Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float32 PASSED [ 8%] 2023-03-31T05:52:26.3452058Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_float64 PASSED [ 8%] 2023-03-31T05:52:26.3452440Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int16 PASSED [ 8%] 2023-03-31T05:52:26.3452824Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int32 PASSED [ 8%] 2023-03-31T05:52:26.3453199Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int64 PASSED [ 8%] 2023-03-31T05:52:26.3453562Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_int8 PASSED [ 8%] 2023-03-31T05:52:26.3453941Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_div_is_fastpath_True_cpu_uint8 PASSED [ 8%] 2023-03-31T05:52:26.3454338Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_bfloat16 PASSED [ 8%] 2023-03-31T05:52:26.3454730Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_float32 PASSED [ 8%] 2023-03-31T05:52:26.3455127Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_float64 PASSED [ 9%] 2023-03-31T05:52:26.3455530Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int16 PASSED [ 9%] 2023-03-31T05:52:26.3455925Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int32 PASSED [ 9%] 2023-03-31T05:52:26.3456302Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int64 PASSED [ 9%] 2023-03-31T05:52:26.3456698Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_int8 PASSED [ 9%] 2023-03-31T05:52:26.3457094Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_False_cpu_uint8 PASSED [ 9%] 2023-03-31T05:52:26.3457496Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_bfloat16 PASSED [ 9%] 2023-03-31T05:52:26.3457892Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_float32 PASSED [ 9%] 2023-03-31T05:52:26.3458293Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_float64 PASSED [ 9%] 2023-03-31T05:52:26.3458682Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int16 PASSED [ 9%] 2023-03-31T05:52:26.3459072Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int32 PASSED [ 9%] 2023-03-31T05:52:26.3459453Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int64 PASSED [ 10%] 2023-03-31T05:52:26.3459839Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_int8 PASSED [ 10%] 2023-03-31T05:52:26.3460396Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_maximum_is_fastpath_True_cpu_uint8 PASSED [ 10%] 2023-03-31T05:52:26.3460910Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_bfloat16 PASSED [ 10%] 2023-03-31T05:52:26.3461312Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_float32 PASSED [ 10%] 2023-03-31T05:52:26.3461718Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_float64 PASSED [ 10%] 2023-03-31T05:52:26.3462110Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int16 PASSED [ 10%] 2023-03-31T05:52:26.3462570Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int32 PASSED [ 10%] 2023-03-31T05:52:26.3462966Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int64 PASSED [ 10%] 2023-03-31T05:52:26.3463355Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_int8 PASSED [ 10%] 2023-03-31T05:52:26.3463753Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_False_cpu_uint8 PASSED [ 10%] 2023-03-31T05:52:26.3464143Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_bfloat16 PASSED [ 11%] 2023-03-31T05:52:26.3464545Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_float32 PASSED [ 11%] 2023-03-31T05:52:26.3464940Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_float64 PASSED [ 11%] 2023-03-31T05:52:26.3465330Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int16 PASSED [ 11%] 2023-03-31T05:52:26.3465703Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int32 PASSED [ 11%] 2023-03-31T05:52:26.3466084Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int64 PASSED [ 11%] 2023-03-31T05:52:26.3466481Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_int8 PASSED [ 11%] 2023-03-31T05:52:26.3466860Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_minimum_is_fastpath_True_cpu_uint8 PASSED [ 11%] 2023-03-31T05:52:26.3467245Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_bfloat16 PASSED [ 11%] 2023-03-31T05:52:26.3467628Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_bool PASSED [ 11%] 2023-03-31T05:52:26.3468020Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_complex128 PASSED [ 12%] 2023-03-31T05:52:26.3468408Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_complex64 PASSED [ 12%] 2023-03-31T05:52:26.3468801Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float16 PASSED [ 12%] 2023-03-31T05:52:26.3469193Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float32 PASSED [ 12%] 2023-03-31T05:52:26.3469589Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_float64 PASSED [ 12%] 2023-03-31T05:52:26.3469969Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int16 PASSED [ 12%] 2023-03-31T05:52:26.3470354Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int32 PASSED [ 12%] 2023-03-31T05:52:26.3470735Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int64 PASSED [ 12%] 2023-03-31T05:52:26.3471119Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_int8 PASSED [ 12%] 2023-03-31T05:52:26.3471486Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_False_cpu_uint8 PASSED [ 12%] 2023-03-31T05:52:26.3471875Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_bfloat16 PASSED [ 12%] 2023-03-31T05:52:26.3472324Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_bool PASSED [ 13%] 2023-03-31T05:52:26.3472702Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_complex128 PASSED [ 13%] 2023-03-31T05:52:26.3473162Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_complex64 PASSED [ 13%] 2023-03-31T05:52:26.3473554Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float16 PASSED [ 13%] 2023-03-31T05:52:26.3473946Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float32 PASSED [ 13%] 2023-03-31T05:52:26.3474320Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_float64 PASSED [ 13%] 2023-03-31T05:52:26.3474705Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int16 PASSED [ 13%] 2023-03-31T05:52:26.3475087Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int32 PASSED [ 13%] 2023-03-31T05:52:26.3475469Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int64 PASSED [ 13%] 2023-03-31T05:52:26.3475841Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_int8 PASSED [ 13%] 2023-03-31T05:52:26.3476222Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_mul_is_fastpath_True_cpu_uint8 PASSED [ 14%] 2023-03-31T05:52:26.3476700Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_bfloat16 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3477238Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_float32 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3477757Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_float64 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3478286Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int16 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3478804Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int32 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3479309Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int64 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3479805Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_int8 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3480326Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_False_cpu_uint8 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3480847Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_bfloat16 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3481378Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_float32 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 14%] 2023-03-31T05:52:26.3481894Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_float64 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3482400Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int16 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3482914Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int32 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3483577Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int64 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3484305Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_int8 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3484810Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_pow_is_fastpath_True_cpu_uint8 SKIPPED (Memory leak https://github.com/pytorch/pytorch/issues/95237) [ 15%] 2023-03-31T05:52:26.3485263Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_bfloat16 PASSED [ 15%] 2023-03-31T05:52:26.3485656Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_bool PASSED [ 15%] 2023-03-31T05:52:26.3486050Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_complex128 PASSED [ 15%] 2023-03-31T05:52:26.3486437Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_complex64 PASSED [ 15%] 2023-03-31T05:52:26.3486837Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float16 PASSED [ 16%] 2023-03-31T05:52:26.3487228Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float32 PASSED [ 16%] 2023-03-31T05:52:26.3487626Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_float64 PASSED [ 16%] 2023-03-31T05:52:26.3488001Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int16 PASSED [ 16%] 2023-03-31T05:52:26.3488386Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int32 PASSED [ 16%] 2023-03-31T05:52:26.3488771Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int64 PASSED [ 16%] 2023-03-31T05:52:26.3489141Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_int8 PASSED [ 16%] 2023-03-31T05:52:26.3489524Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_False_cpu_uint8 PASSED [ 16%] 2023-03-31T05:52:26.3489915Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_bfloat16 PASSED [ 16%] 2023-03-31T05:52:26.3490309Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_bool PASSED [ 16%] 2023-03-31T05:52:26.3490691Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_complex128 PASSED [ 16%] 2023-03-31T05:52:26.3491088Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_complex64 PASSED [ 17%] 2023-03-31T05:52:26.3491479Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float16 PASSED [ 17%] 2023-03-31T05:52:26.3491861Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float32 PASSED [ 17%] 2023-03-31T05:52:26.3492237Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_float64 PASSED [ 17%] 2023-03-31T05:52:26.3492620Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int16 PASSED [ 17%] 2023-03-31T05:52:26.3493005Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int32 PASSED [ 17%] 2023-03-31T05:52:26.3493381Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int64 PASSED [ 17%] 2023-03-31T05:52:26.3493744Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_int8 PASSED [ 17%] 2023-03-31T05:52:26.3494123Z test_foreach.py::TestForeachCPU::test_binary_op__foreach_sub_is_fastpath_True_cpu_uint8 PASSED [ 17%] 2023-03-31T05:52:26.3494496Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_bfloat16 PASSED [ 17%] 2023-03-31T05:52:26.3494861Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float16 PASSED [ 18%] 2023-03-31T05:52:26.3495237Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float32 PASSED [ 18%] 2023-03-31T05:52:26.3495609Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_add_cpu_float64 PASSED [ 18%] 2023-03-31T05:52:26.3496072Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_bfloat16 PASSED [ 18%] 2023-03-31T05:52:26.3496518Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float16 PASSED [ 18%] 2023-03-31T05:52:26.3496912Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float32 PASSED [ 18%] 2023-03-31T05:52:26.3497305Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_max_cpu_float64 PASSED [ 18%] 2023-03-31T05:52:26.3497692Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_bfloat16 PASSED [ 18%] 2023-03-31T05:52:26.3498069Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float16 PASSED [ 18%] 2023-03-31T05:52:26.3498456Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float32 PASSED [ 18%] 2023-03-31T05:52:26.3498847Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_clamp_min_cpu_float64 PASSED [ 18%] 2023-03-31T05:52:26.3499215Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_bfloat16 PASSED [ 19%] 2023-03-31T05:52:26.3499592Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float16 PASSED [ 19%] 2023-03-31T05:52:26.3499969Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float32 PASSED [ 19%] 2023-03-31T05:52:26.3500344Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_div_cpu_float64 PASSED [ 19%] 2023-03-31T05:52:26.3500715Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_bfloat16 PASSED [ 19%] 2023-03-31T05:52:26.3501105Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float16 PASSED [ 19%] 2023-03-31T05:52:26.3501493Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float32 PASSED [ 19%] 2023-03-31T05:52:26.3501887Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_maximum_cpu_float64 PASSED [ 19%] 2023-03-31T05:52:26.3502269Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_bfloat16 PASSED [ 19%] 2023-03-31T05:52:26.3502749Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float16 PASSED [ 19%] 2023-03-31T05:52:26.3503135Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float32 PASSED [ 20%] 2023-03-31T05:52:26.3503519Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_minimum_cpu_float64 PASSED [ 20%] 2023-03-31T05:52:26.3503889Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_bfloat16 PASSED [ 20%] 2023-03-31T05:52:26.3504265Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float16 PASSED [ 20%] 2023-03-31T05:52:26.3504638Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float32 PASSED [ 20%] 2023-03-31T05:52:26.3504999Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_mul_cpu_float64 PASSED [ 20%] 2023-03-31T05:52:26.3505382Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_bfloat16 PASSED [ 20%] 2023-03-31T05:52:26.3505758Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float16 PASSED [ 20%] 2023-03-31T05:52:26.3506137Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float32 PASSED [ 20%] 2023-03-31T05:52:26.3506498Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_pow_cpu_float64 PASSED [ 20%] 2023-03-31T05:52:26.3506876Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_bfloat16 PASSED [ 20%] 2023-03-31T05:52:26.3507249Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float16 PASSED [ 21%] 2023-03-31T05:52:26.3507690Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float32 PASSED [ 21%] 2023-03-31T05:52:26.3508049Z test_foreach.py::TestForeachCPU::test_binary_op_float_inf_nan__foreach_sub_cpu_float64 PASSED [ 21%] 2023-03-31T05:52:26.3508494Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_bfloat16 PASSED [ 21%] 2023-03-31T05:52:26.3508878Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_bool PASSED [ 21%] 2023-03-31T05:52:26.3509254Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_complex128 PASSED [ 21%] 2023-03-31T05:52:26.3509649Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_complex64 PASSED [ 21%] 2023-03-31T05:52:26.3510039Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float16 PASSED [ 21%] 2023-03-31T05:52:26.3510423Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float32 PASSED [ 21%] 2023-03-31T05:52:26.3510797Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_float64 PASSED [ 21%] 2023-03-31T05:52:26.3511187Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int16 PASSED [ 21%] 2023-03-31T05:52:26.3511558Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int32 PASSED [ 22%] 2023-03-31T05:52:26.3511928Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int64 PASSED [ 22%] 2023-03-31T05:52:26.3512289Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_int8 PASSED [ 22%] 2023-03-31T05:52:26.3512667Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_add_cpu_uint8 PASSED [ 22%] 2023-03-31T05:52:26.3513056Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_bfloat16 PASSED [ 22%] 2023-03-31T05:52:26.3513446Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_float32 PASSED [ 22%] 2023-03-31T05:52:26.3513852Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_float64 PASSED [ 22%] 2023-03-31T05:52:26.3514253Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int16 PASSED [ 22%] 2023-03-31T05:52:26.3514645Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int32 PASSED [ 22%] 2023-03-31T05:52:26.3515024Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int64 PASSED [ 22%] 2023-03-31T05:52:26.3515416Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_int8 PASSED [ 23%] 2023-03-31T05:52:26.3515806Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_max_cpu_uint8 PASSED [ 23%] 2023-03-31T05:52:26.3516201Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_bfloat16 PASSED [ 23%] 2023-03-31T05:52:26.3516591Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_float32 PASSED [ 23%] 2023-03-31T05:52:26.3516987Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_float64 PASSED [ 23%] 2023-03-31T05:52:26.3517385Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int16 PASSED [ 23%] 2023-03-31T05:52:26.3517776Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int32 PASSED [ 23%] 2023-03-31T05:52:26.3518147Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int64 PASSED [ 23%] 2023-03-31T05:52:26.3518536Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_int8 PASSED [ 23%] 2023-03-31T05:52:26.3518924Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_clamp_min_cpu_uint8 PASSED [ 23%] 2023-03-31T05:52:26.3519300Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_bfloat16 PASSED [ 23%] 2023-03-31T05:52:26.3519748Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_bool PASSED [ 24%] 2023-03-31T05:52:26.3520197Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_complex128 PASSED [ 24%] 2023-03-31T05:52:26.3520593Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_complex64 PASSED [ 24%] 2023-03-31T05:52:26.3520971Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float16 PASSED [ 24%] 2023-03-31T05:52:26.3521359Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float32 PASSED [ 24%] 2023-03-31T05:52:26.3521745Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_float64 PASSED [ 24%] 2023-03-31T05:52:26.3522126Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int16 PASSED [ 24%] 2023-03-31T05:52:26.3522495Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int32 PASSED [ 24%] 2023-03-31T05:52:26.3522873Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int64 PASSED [ 24%] 2023-03-31T05:52:26.3523388Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_int8 PASSED [ 24%] 2023-03-31T05:52:26.3523765Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_div_cpu_uint8 PASSED [ 25%] 2023-03-31T05:52:26.3524140Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_bfloat16 PASSED [ 25%] 2023-03-31T05:52:26.3524541Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_float32 PASSED [ 25%] 2023-03-31T05:52:26.3524933Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_float64 PASSED [ 25%] 2023-03-31T05:52:26.3525306Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int16 PASSED [ 25%] 2023-03-31T05:52:26.3525696Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int32 PASSED [ 25%] 2023-03-31T05:52:26.3526082Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int64 PASSED [ 25%] 2023-03-31T05:52:26.3526476Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_int8 PASSED [ 25%] 2023-03-31T05:52:26.3526852Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_maximum_cpu_uint8 PASSED [ 25%] 2023-03-31T05:52:26.3527243Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_bfloat16 PASSED [ 25%] 2023-03-31T05:52:26.3527640Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_float32 PASSED [ 25%] 2023-03-31T05:52:26.3528036Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_float64 PASSED [ 26%] 2023-03-31T05:52:26.3528409Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int16 PASSED [ 26%] 2023-03-31T05:52:26.3528796Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int32 PASSED [ 26%] 2023-03-31T05:52:26.3529182Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int64 PASSED [ 26%] 2023-03-31T05:52:26.3529571Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_int8 PASSED [ 26%] 2023-03-31T05:52:26.3529950Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_minimum_cpu_uint8 PASSED [ 26%] 2023-03-31T05:52:26.3530336Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_bfloat16 PASSED [ 26%] 2023-03-31T05:52:26.3530720Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_bool PASSED [ 26%] 2023-03-31T05:52:26.3531100Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_complex128 PASSED [ 26%] 2023-03-31T05:52:26.3531598Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_complex64 PASSED [ 26%] 2023-03-31T05:52:26.3531987Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float16 PASSED [ 27%] 2023-03-31T05:52:26.3532461Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float32 PASSED [ 27%] 2023-03-31T05:52:26.3532830Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_float64 PASSED [ 27%] 2023-03-31T05:52:26.3533212Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int16 PASSED [ 27%] 2023-03-31T05:52:26.3533591Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int32 PASSED [ 27%] 2023-03-31T05:52:26.3533958Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int64 PASSED [ 27%] 2023-03-31T05:52:26.3534322Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_int8 PASSED [ 27%] 2023-03-31T05:52:26.3534707Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_mul_cpu_uint8 PASSED [ 27%] 2023-03-31T05:52:26.3535095Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_bfloat16 PASSED [ 27%] 2023-03-31T05:52:26.3535471Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_float32 PASSED [ 27%] 2023-03-31T05:52:26.3535856Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_float64 PASSED [ 27%] 2023-03-31T05:52:26.3536239Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int16 PASSED [ 28%] 2023-03-31T05:52:26.3536617Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int32 PASSED [ 28%] 2023-03-31T05:52:26.3536975Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int64 PASSED [ 28%] 2023-03-31T05:52:26.3537353Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_int8 PASSED [ 28%] 2023-03-31T05:52:26.3537733Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_pow_cpu_uint8 PASSED [ 28%] 2023-03-31T05:52:26.3538121Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_bfloat16 PASSED [ 28%] 2023-03-31T05:52:26.3538488Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_bool PASSED [ 28%] 2023-03-31T05:52:26.3538872Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_complex128 PASSED [ 28%] 2023-03-31T05:52:26.3539265Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_complex64 PASSED [ 28%] 2023-03-31T05:52:26.3539639Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float16 PASSED [ 28%] 2023-03-31T05:52:26.3540026Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float32 PASSED [ 29%] 2023-03-31T05:52:26.3540405Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_float64 PASSED [ 29%] 2023-03-31T05:52:26.3540787Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int16 PASSED [ 29%] 2023-03-31T05:52:26.3541157Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int32 PASSED [ 29%] 2023-03-31T05:52:26.3541524Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int64 PASSED [ 29%] 2023-03-31T05:52:26.3541904Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_int8 PASSED [ 29%] 2023-03-31T05:52:26.3542281Z test_foreach.py::TestForeachCPU::test_binary_op_list_error_cases__foreach_sub_cpu_uint8 PASSED [ 29%] 2023-03-31T05:52:26.3542756Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_bfloat16 SKIPPED (CUDA not found) [ 29%] 2023-03-31T05:52:26.3543174Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_bool SKIPPED (CUDA not found) [ 29%] 2023-03-31T05:52:26.3543662Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_complex128 SKIPPED (CUDA not found) [ 29%] 2023-03-31T05:52:26.3544151Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_complex64 SKIPPED (CUDA not found) [ 29%] 2023-03-31T05:52:26.3544559Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float16 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3544980Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float32 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3545395Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_float64 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3545808Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int16 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3546201Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int32 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3546606Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int64 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3547016Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_int8 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3547428Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_add_cpu_uint8 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3547836Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_bfloat16 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3548271Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_float32 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3548695Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_float64 SKIPPED (CUDA not found) [ 30%] 2023-03-31T05:52:26.3549104Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int16 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3549528Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int32 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3549949Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int64 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3550363Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_int8 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3550776Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_max_cpu_uint8 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3551196Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_bfloat16 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3551627Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_float32 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3552056Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_float64 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3552472Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int16 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3552885Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int32 SKIPPED (CUDA not found) [ 31%] 2023-03-31T05:52:26.3553302Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int64 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3553715Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_int8 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3554123Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_clamp_min_cpu_uint8 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3554537Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_bfloat16 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3555006Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_bool SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3555483Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_complex128 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3555892Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_complex64 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3556307Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float16 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3556720Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float32 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3557134Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_float64 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3557532Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int16 SKIPPED (CUDA not found) [ 32%] 2023-03-31T05:52:26.3557939Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int32 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3558340Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int64 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3558744Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_int8 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3559141Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_div_cpu_uint8 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3559563Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_bfloat16 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3559986Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_float32 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3560400Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_float64 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3560807Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int16 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3561226Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int32 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3561643Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int64 SKIPPED (CUDA not found) [ 33%] 2023-03-31T05:52:26.3562058Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_int8 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3562459Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_maximum_cpu_uint8 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3562876Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_bfloat16 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3563433Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_float32 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3563857Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_float64 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3564266Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int16 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3564678Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int32 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3565089Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int64 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3565503Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_int8 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3565909Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_minimum_cpu_uint8 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3566414Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_bfloat16 SKIPPED (CUDA not found) [ 34%] 2023-03-31T05:52:26.3566898Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_bool SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3567302Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_complex128 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3567722Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_complex64 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3568137Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float16 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3568550Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float32 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3568951Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_float64 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3569365Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int16 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3569772Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int32 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3570170Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int64 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3570559Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_int8 SKIPPED (CUDA not found) [ 35%] 2023-03-31T05:52:26.3570967Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_mul_cpu_uint8 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3571381Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_bfloat16 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3571798Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_float32 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3572206Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_float64 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3572623Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int16 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3573021Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int32 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3573415Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int64 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3573802Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_int8 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3574210Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_pow_cpu_uint8 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3574626Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_bfloat16 SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3575045Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_bool SKIPPED (CUDA not found) [ 36%] 2023-03-31T05:52:26.3575452Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_complex128 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3575875Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_complex64 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3576292Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float16 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3576704Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float32 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3577103Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_float64 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3577508Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int16 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3578008Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int32 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3578463Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int64 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3578851Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_int8 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3579258Z test_foreach.py::TestForeachCPU::test_binary_op_list_slow_path__foreach_sub_cpu_uint8 SKIPPED (CUDA not found) [ 37%] 2023-03-31T05:52:26.3579679Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_add_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3580102Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_max_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3580546Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_min_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3580986Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_div_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3581426Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_maximum_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3581856Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_minimum_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3582287Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_mul_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3582805Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_pow_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3583232Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_different_tensor_dtypes__foreach_sub_cpu_float32 PASSED [ 38%] 2023-03-31T05:52:26.3583649Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_bfloat16 PASSED [ 38%] 2023-03-31T05:52:26.3584075Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_bool PASSED [ 38%] 2023-03-31T05:52:26.3584505Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_complex128 PASSED [ 39%] 2023-03-31T05:52:26.3584936Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_complex64 PASSED [ 39%] 2023-03-31T05:52:26.3585352Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float16 PASSED [ 39%] 2023-03-31T05:52:26.3585775Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float32 PASSED [ 39%] 2023-03-31T05:52:26.3586198Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_float64 PASSED [ 39%] 2023-03-31T05:52:26.3586624Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int16 PASSED [ 39%] 2023-03-31T05:52:26.3587035Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int32 PASSED [ 39%] 2023-03-31T05:52:26.3587452Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int64 PASSED [ 39%] 2023-03-31T05:52:26.3587873Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_int8 PASSED [ 39%] 2023-03-31T05:52:26.3588288Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cpu_uint8 PASSED [ 39%] 2023-03-31T05:52:26.3588702Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_bfloat16 PASSED [ 40%] 2023-03-31T05:52:26.3589139Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_float32 PASSED [ 40%] 2023-03-31T05:52:26.3589647Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_float64 PASSED [ 40%] 2023-03-31T05:52:26.3590134Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int16 PASSED [ 40%] 2023-03-31T05:52:26.3590549Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int32 PASSED [ 40%] 2023-03-31T05:52:26.3590981Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int64 PASSED [ 40%] 2023-03-31T05:52:26.3591407Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_int8 PASSED [ 40%] 2023-03-31T05:52:26.3591833Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cpu_uint8 PASSED [ 40%] 2023-03-31T05:52:26.3592250Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_bfloat16 PASSED [ 40%] 2023-03-31T05:52:26.3592692Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_float32 PASSED [ 40%] 2023-03-31T05:52:26.3593128Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_float64 PASSED [ 40%] 2023-03-31T05:52:26.3593555Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int16 PASSED [ 41%] 2023-03-31T05:52:26.3593965Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int32 PASSED [ 41%] 2023-03-31T05:52:26.3594387Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int64 PASSED [ 41%] 2023-03-31T05:52:26.3594813Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_int8 PASSED [ 41%] 2023-03-31T05:52:26.3595236Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cpu_uint8 PASSED [ 41%] 2023-03-31T05:52:26.3595655Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_bfloat16 PASSED [ 41%] 2023-03-31T05:52:26.3596085Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_bool PASSED [ 41%] 2023-03-31T05:52:26.3596509Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_complex128 PASSED [ 41%] 2023-03-31T05:52:26.3596939Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_complex64 PASSED [ 41%] 2023-03-31T05:52:26.3597355Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float16 PASSED [ 41%] 2023-03-31T05:52:26.3597779Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float32 PASSED [ 41%] 2023-03-31T05:52:26.3598200Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_float64 PASSED [ 42%] 2023-03-31T05:52:26.3598614Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int16 PASSED [ 42%] 2023-03-31T05:52:26.3599029Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int32 PASSED [ 42%] 2023-03-31T05:52:26.3599445Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int64 PASSED [ 42%] 2023-03-31T05:52:26.3599863Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_int8 PASSED [ 42%] 2023-03-31T05:52:26.3600269Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cpu_uint8 PASSED [ 42%] 2023-03-31T05:52:26.3600691Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_bfloat16 PASSED [ 42%] 2023-03-31T05:52:26.3601126Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_float32 PASSED [ 42%] 2023-03-31T05:52:26.3601615Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_float64 PASSED [ 42%] 2023-03-31T05:52:26.3602080Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int16 PASSED [ 42%] 2023-03-31T05:52:26.3602512Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int32 PASSED [ 43%] 2023-03-31T05:52:26.3602934Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int64 PASSED [ 43%] 2023-03-31T05:52:26.3603496Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_int8 PASSED [ 43%] 2023-03-31T05:52:26.3603913Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cpu_uint8 PASSED [ 43%] 2023-03-31T05:52:26.3604347Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_bfloat16 PASSED [ 43%] 2023-03-31T05:52:26.3604785Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_float32 PASSED [ 43%] 2023-03-31T05:52:26.3605217Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_float64 PASSED [ 43%] 2023-03-31T05:52:26.3605631Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int16 PASSED [ 43%] 2023-03-31T05:52:26.3606055Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int32 PASSED [ 43%] 2023-03-31T05:52:26.3606481Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int64 PASSED [ 43%] 2023-03-31T05:52:26.3606909Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_int8 PASSED [ 43%] 2023-03-31T05:52:26.3607329Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cpu_uint8 PASSED [ 44%] 2023-03-31T05:52:26.3607759Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_bfloat16 PASSED [ 44%] 2023-03-31T05:52:26.3608179Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_bool PASSED [ 44%] 2023-03-31T05:52:26.3608606Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_complex128 PASSED [ 44%] 2023-03-31T05:52:26.3609023Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_complex64 PASSED [ 44%] 2023-03-31T05:52:26.3609447Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float16 PASSED [ 44%] 2023-03-31T05:52:26.3609867Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float32 PASSED [ 44%] 2023-03-31T05:52:26.3610289Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_float64 PASSED [ 44%] 2023-03-31T05:52:26.3610705Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int16 PASSED [ 44%] 2023-03-31T05:52:26.3611121Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int32 PASSED [ 44%] 2023-03-31T05:52:26.3611530Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int64 PASSED [ 45%] 2023-03-31T05:52:26.3611930Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_int8 PASSED [ 45%] 2023-03-31T05:52:26.3612343Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cpu_uint8 PASSED [ 45%] 2023-03-31T05:52:26.3612771Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_bfloat16 PASSED [ 45%] 2023-03-31T05:52:26.3613283Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_float32 PASSED [ 45%] 2023-03-31T05:52:26.3613773Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_float64 PASSED [ 45%] 2023-03-31T05:52:26.3614197Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int16 PASSED [ 45%] 2023-03-31T05:52:26.3614613Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int32 PASSED [ 45%] 2023-03-31T05:52:26.3615024Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int64 PASSED [ 45%] 2023-03-31T05:52:26.3615433Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_int8 PASSED [ 45%] 2023-03-31T05:52:26.3615849Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cpu_uint8 PASSED [ 45%] 2023-03-31T05:52:26.3616273Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_bfloat16 PASSED [ 46%] 2023-03-31T05:52:26.3616694Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_bool PASSED [ 46%] 2023-03-31T05:52:26.3617102Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_complex128 PASSED [ 46%] 2023-03-31T05:52:26.3617530Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_complex64 PASSED [ 46%] 2023-03-31T05:52:26.3617956Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float16 PASSED [ 46%] 2023-03-31T05:52:26.3618377Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float32 PASSED [ 46%] 2023-03-31T05:52:26.3618782Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_float64 PASSED [ 46%] 2023-03-31T05:52:26.3619205Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int16 PASSED [ 46%] 2023-03-31T05:52:26.3619622Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int32 PASSED [ 46%] 2023-03-31T05:52:26.3620030Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int64 PASSED [ 46%] 2023-03-31T05:52:26.3620432Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_int8 PASSED [ 47%] 2023-03-31T05:52:26.3620848Z test_foreach.py::TestForeachCPU::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cpu_uint8 PASSED [ 47%] 2023-03-31T05:52:26.3621286Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3621742Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_bool SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3622187Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_complex128 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3622742Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_complex64 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3623196Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float16 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3623647Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float32 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3624081Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_float64 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3624525Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int16 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3625043Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int32 SKIPPED (Only runs on cuda) [ 47%] 2023-03-31T05:52:26.3625545Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int64 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3625978Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_int8 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3626418Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_add_cpu_uint8 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3626872Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3627337Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_float32 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3627794Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_float64 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3628251Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int16 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3628713Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int32 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3629159Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int64 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3629594Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_int8 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T05:52:26.3630050Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cpu_uint8 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3630516Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3630982Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_float32 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3631435Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_float64 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3631891Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int16 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3632344Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int32 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3632793Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int64 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3633236Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_int8 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3633693Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cpu_uint8 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3634148Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3634593Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_bool SKIPPED (Only runs on cuda) [ 49%] 2023-03-31T05:52:26.3635035Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_complex128 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3635489Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_complex64 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3635994Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float16 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3636499Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float32 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3636946Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_float64 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3637380Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int16 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3637822Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int32 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3638266Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int64 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3638717Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_int8 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3639153Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_div_cpu_uint8 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3639607Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T05:52:26.3640064Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_float32 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3640518Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_float64 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3640954Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int16 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3641411Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int32 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3641860Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int64 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3642304Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_int8 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3642739Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_maximum_cpu_uint8 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3643428Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3643889Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_float32 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3644344Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_float64 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T05:52:26.3644789Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int16 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3645238Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int32 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3645680Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int64 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3646125Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_int8 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3646557Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_minimum_cpu_uint8 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3647000Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3647604Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_bool SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3648052Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_complex128 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3648489Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_complex64 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3648938Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float16 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3649387Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float32 SKIPPED (Only runs on cuda) [ 52%] 2023-03-31T05:52:26.3649832Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_float64 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3650267Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int16 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3650713Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3651154Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int64 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3651594Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_int8 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3652017Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_mul_cpu_uint8 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3652461Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3652914Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3653364Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_float64 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3653795Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int16 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T05:52:26.3654234Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int32 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3654680Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int64 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3655120Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_int8 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3655554Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_pow_cpu_uint8 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3656002Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3656441Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_bool SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3656889Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_complex128 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3657339Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_complex64 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3657773Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float16 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3658287Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float32 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3658788Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_float64 SKIPPED (Only runs on cuda) [ 54%] 2023-03-31T05:52:26.3659234Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int16 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3659660Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int32 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3660096Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int64 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3660535Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_int8 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3660970Z test_foreach.py::TestForeachCPU::test_binary_op_tensors_on_different_devices__foreach_sub_cpu_uint8 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3661397Z test_foreach.py::TestForeachCPU::test_foreach_l2_large_value_input__foreach_norm_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3661836Z test_foreach.py::TestForeachCPU::test_foreach_l2_large_value_input__foreach_norm_cpu_float16 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3662272Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3662790Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_complex128 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3663213Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_complex64 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T05:52:26.3663648Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float16 SKIPPED (Only runs on cuda) [ 56%] 2023-03-31T05:52:26.3664084Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float32 SKIPPED (Only runs on cuda) [ 56%] 2023-03-31T05:52:26.3664517Z test_foreach.py::TestForeachCPU::test_foreach_reduce_large_input__foreach_norm_cpu_float64 SKIPPED (Only runs on cuda) [ 56%] 2023-03-31T05:52:26.3664913Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_complex128 PASSED [ 56%] 2023-03-31T05:52:26.3665310Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_complex64 PASSED [ 56%] 2023-03-31T05:52:26.3665701Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_float32 PASSED [ 56%] 2023-03-31T05:52:26.3666070Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_False__foreach_lerp_cpu_float64 PASSED [ 56%] 2023-03-31T05:52:26.3666461Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_complex128 PASSED [ 56%] 2023-03-31T05:52:26.3666855Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_complex64 PASSED [ 56%] 2023-03-31T05:52:26.3667243Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_float32 PASSED [ 56%] 2023-03-31T05:52:26.3667618Z test_foreach.py::TestForeachCPU::test_lerp_is_fastpath_True__foreach_lerp_cpu_float64 PASSED [ 56%] 2023-03-31T05:52:26.3668022Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_complex128 PASSED [ 57%] 2023-03-31T05:52:26.3668439Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_complex64 PASSED [ 57%] 2023-03-31T05:52:26.3668852Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_float32 PASSED [ 57%] 2023-03-31T05:52:26.3669251Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_float64 PASSED [ 57%] 2023-03-31T05:52:26.3669655Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int16 PASSED [ 57%] 2023-03-31T05:52:26.3670149Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int32 PASSED [ 57%] 2023-03-31T05:52:26.3670610Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int64 PASSED [ 57%] 2023-03-31T05:52:26.3671002Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_int8 PASSED [ 57%] 2023-03-31T05:52:26.3671412Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_False_cpu_uint8 PASSED [ 57%] 2023-03-31T05:52:26.3671822Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_complex128 PASSED [ 57%] 2023-03-31T05:52:26.3672225Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_complex64 PASSED [ 58%] 2023-03-31T05:52:26.3672634Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_float32 PASSED [ 58%] 2023-03-31T05:52:26.3673044Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_float64 PASSED [ 58%] 2023-03-31T05:52:26.3673456Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int16 PASSED [ 58%] 2023-03-31T05:52:26.3673847Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int32 PASSED [ 58%] 2023-03-31T05:52:26.3674245Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int64 PASSED [ 58%] 2023-03-31T05:52:26.3674644Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_int8 PASSED [ 58%] 2023-03-31T05:52:26.3675049Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcdiv_is_fastpath_True_cpu_uint8 PASSED [ 58%] 2023-03-31T05:52:26.3675445Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_complex128 PASSED [ 58%] 2023-03-31T05:52:26.3675868Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_complex64 PASSED [ 58%] 2023-03-31T05:52:26.3676283Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_float32 PASSED [ 58%] 2023-03-31T05:52:26.3676697Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_float64 PASSED [ 59%] 2023-03-31T05:52:26.3677103Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int16 PASSED [ 59%] 2023-03-31T05:52:26.3677494Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int32 PASSED [ 59%] 2023-03-31T05:52:26.3677890Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int64 PASSED [ 59%] 2023-03-31T05:52:26.3678294Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_int8 PASSED [ 59%] 2023-03-31T05:52:26.3678695Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_False_cpu_uint8 PASSED [ 59%] 2023-03-31T05:52:26.3679101Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_complex128 PASSED [ 59%] 2023-03-31T05:52:26.3679520Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_complex64 PASSED [ 59%] 2023-03-31T05:52:26.3679932Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_float32 PASSED [ 59%] 2023-03-31T05:52:26.3680340Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_float64 PASSED [ 59%] 2023-03-31T05:52:26.3680731Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int16 PASSED [ 60%] 2023-03-31T05:52:26.3681135Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int32 PASSED [ 60%] 2023-03-31T05:52:26.3681531Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int64 PASSED [ 60%] 2023-03-31T05:52:26.3681992Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_int8 PASSED [ 60%] 2023-03-31T05:52:26.3682437Z test_foreach.py::TestForeachCPU::test_pointwise_op__foreach_addcmul_is_fastpath_True_cpu_uint8 PASSED [ 60%] 2023-03-31T05:52:26.3682880Z test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cpu_float32 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T05:52:26.3683488Z test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cpu_float64 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T05:52:26.3683955Z test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cpu_float32 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T05:52:26.3684409Z test_foreach.py::TestForeachCPU::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cpu_float64 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T05:52:26.3684834Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_bfloat16 PASSED [ 60%] 2023-03-31T05:52:26.3685243Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_complex128 PASSED [ 60%] 2023-03-31T05:52:26.3685649Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_complex64 PASSED [ 61%] 2023-03-31T05:52:26.3686034Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float16 PASSED [ 61%] 2023-03-31T05:52:26.3686425Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float32 PASSED [ 61%] 2023-03-31T05:52:26.3686815Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_False_cpu_float64 PASSED [ 61%] 2023-03-31T05:52:26.3687194Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_bfloat16 PASSED [ 61%] 2023-03-31T05:52:26.3687591Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_complex128 PASSED [ 61%] 2023-03-31T05:52:26.3687996Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_complex64 PASSED [ 61%] 2023-03-31T05:52:26.3688396Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float16 PASSED [ 61%] 2023-03-31T05:52:26.3688776Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float32 PASSED [ 61%] 2023-03-31T05:52:26.3689166Z test_foreach.py::TestForeachCPU::test_reduce_op__foreach_norm_is_fastpath_True_cpu_float64 PASSED [ 61%] 2023-03-31T05:52:26.3689557Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_bfloat16 PASSED [ 61%] 2023-03-31T05:52:26.3689950Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_complex128 PASSED [ 62%] 2023-03-31T05:52:26.3690338Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_complex64 PASSED [ 62%] 2023-03-31T05:52:26.3690729Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float16 PASSED [ 62%] 2023-03-31T05:52:26.3691122Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float32 PASSED [ 62%] 2023-03-31T05:52:26.3691512Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_float64 PASSED [ 62%] 2023-03-31T05:52:26.3691883Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int16 PASSED [ 62%] 2023-03-31T05:52:26.3692262Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int32 PASSED [ 62%] 2023-03-31T05:52:26.3692632Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int64 PASSED [ 62%] 2023-03-31T05:52:26.3692994Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_int8 PASSED [ 62%] 2023-03-31T05:52:26.3693371Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_False_cpu_uint8 PASSED [ 62%] 2023-03-31T05:52:26.3693760Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_bfloat16 PASSED [ 63%] 2023-03-31T05:52:26.3694244Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_complex128 PASSED [ 63%] 2023-03-31T05:52:26.3694717Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_complex64 PASSED [ 63%] 2023-03-31T05:52:26.3695105Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float16 PASSED [ 63%] 2023-03-31T05:52:26.3695491Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float32 PASSED [ 63%] 2023-03-31T05:52:26.3695872Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_float64 PASSED [ 63%] 2023-03-31T05:52:26.3696241Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int16 PASSED [ 63%] 2023-03-31T05:52:26.3696620Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int32 PASSED [ 63%] 2023-03-31T05:52:26.3697000Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int64 PASSED [ 63%] 2023-03-31T05:52:26.3697364Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_int8 PASSED [ 63%] 2023-03-31T05:52:26.3697748Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_abs_is_fastpath_True_cpu_uint8 PASSED [ 63%] 2023-03-31T05:52:26.3698147Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_complex128 PASSED [ 64%] 2023-03-31T05:52:26.3698548Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_complex64 PASSED [ 64%] 2023-03-31T05:52:26.3698926Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_float32 PASSED [ 64%] 2023-03-31T05:52:26.3699316Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_False_cpu_float64 PASSED [ 64%] 2023-03-31T05:52:26.3699710Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_complex128 PASSED [ 64%] 2023-03-31T05:52:26.3700111Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_complex64 PASSED [ 64%] 2023-03-31T05:52:26.3700491Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_float32 PASSED [ 64%] 2023-03-31T05:52:26.3700877Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_acos_is_fastpath_True_cpu_float64 PASSED [ 64%] 2023-03-31T05:52:26.3701271Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_complex128 PASSED [ 64%] 2023-03-31T05:52:26.3701668Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_complex64 PASSED [ 64%] 2023-03-31T05:52:26.3702048Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_float32 PASSED [ 65%] 2023-03-31T05:52:26.3702516Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_False_cpu_float64 PASSED [ 65%] 2023-03-31T05:52:26.3702918Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_complex128 PASSED [ 65%] 2023-03-31T05:52:26.3703307Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_complex64 PASSED [ 65%] 2023-03-31T05:52:26.3703697Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_float32 PASSED [ 65%] 2023-03-31T05:52:26.3704089Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_asin_is_fastpath_True_cpu_float64 PASSED [ 65%] 2023-03-31T05:52:26.3704481Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_complex128 PASSED [ 65%] 2023-03-31T05:52:26.3704863Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_complex64 PASSED [ 65%] 2023-03-31T05:52:26.3705253Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_float32 PASSED [ 65%] 2023-03-31T05:52:26.3705643Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_False_cpu_float64 PASSED [ 65%] 2023-03-31T05:52:26.3706109Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_complex128 PASSED [ 65%] 2023-03-31T05:52:26.3706493Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_complex64 PASSED [ 66%] 2023-03-31T05:52:26.3707066Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_float32 PASSED [ 66%] 2023-03-31T05:52:26.3707514Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_atan_is_fastpath_True_cpu_float64 PASSED [ 66%] 2023-03-31T05:52:26.3707895Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_bfloat16 PASSED [ 66%] 2023-03-31T05:52:26.3708287Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_float32 PASSED [ 66%] 2023-03-31T05:52:26.3708682Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_float64 PASSED [ 66%] 2023-03-31T05:52:26.3709073Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int16 PASSED [ 66%] 2023-03-31T05:52:26.3709453Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int32 PASSED [ 66%] 2023-03-31T05:52:26.3709840Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int64 PASSED [ 66%] 2023-03-31T05:52:26.3710222Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_int8 PASSED [ 66%] 2023-03-31T05:52:26.3710606Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_False_cpu_uint8 PASSED [ 67%] 2023-03-31T05:52:26.3710981Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_bfloat16 PASSED [ 67%] 2023-03-31T05:52:26.3711371Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_float32 PASSED [ 67%] 2023-03-31T05:52:26.3711753Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_float64 PASSED [ 67%] 2023-03-31T05:52:26.3712140Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int16 PASSED [ 67%] 2023-03-31T05:52:26.3712515Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int32 PASSED [ 67%] 2023-03-31T05:52:26.3712898Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int64 PASSED [ 67%] 2023-03-31T05:52:26.3713277Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_int8 PASSED [ 67%] 2023-03-31T05:52:26.3713647Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_ceil_is_fastpath_True_cpu_uint8 PASSED [ 67%] 2023-03-31T05:52:26.3714046Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_complex128 PASSED [ 67%] 2023-03-31T05:52:26.3714442Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_complex64 PASSED [ 67%] 2023-03-31T05:52:26.3714829Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_float32 PASSED [ 68%] 2023-03-31T05:52:26.3715209Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_False_cpu_float64 PASSED [ 68%] 2023-03-31T05:52:26.3715606Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_complex128 PASSED [ 68%] 2023-03-31T05:52:26.3716005Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_complex64 PASSED [ 68%] 2023-03-31T05:52:26.3716393Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_float32 PASSED [ 68%] 2023-03-31T05:52:26.3716766Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cos_is_fastpath_True_cpu_float64 PASSED [ 68%] 2023-03-31T05:52:26.3717157Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_complex128 PASSED [ 68%] 2023-03-31T05:52:26.3717558Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_complex64 PASSED [ 68%] 2023-03-31T05:52:26.3717950Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_float32 PASSED [ 68%] 2023-03-31T05:52:26.3718396Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_False_cpu_float64 PASSED [ 68%] 2023-03-31T05:52:26.3718844Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_complex128 PASSED [ 69%] 2023-03-31T05:52:26.3719244Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_complex64 PASSED [ 69%] 2023-03-31T05:52:26.3719624Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_float32 PASSED [ 69%] 2023-03-31T05:52:26.3720012Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_cosh_is_fastpath_True_cpu_float64 PASSED [ 69%] 2023-03-31T05:52:26.3720401Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_bfloat16 PASSED [ 69%] 2023-03-31T05:52:26.3720789Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_float32 PASSED [ 69%] 2023-03-31T05:52:26.3721161Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_False_cpu_float64 PASSED [ 69%] 2023-03-31T05:52:26.3721546Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_bfloat16 PASSED [ 69%] 2023-03-31T05:52:26.3721935Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_float32 PASSED [ 69%] 2023-03-31T05:52:26.3722322Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erf_is_fastpath_True_cpu_float64 PASSED [ 69%] 2023-03-31T05:52:26.3722698Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_bfloat16 PASSED [ 69%] 2023-03-31T05:52:26.3723223Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_float32 PASSED [ 70%] 2023-03-31T05:52:26.3723614Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_False_cpu_float64 PASSED [ 70%] 2023-03-31T05:52:26.3723990Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_bfloat16 PASSED [ 70%] 2023-03-31T05:52:26.3724381Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_float32 PASSED [ 70%] 2023-03-31T05:52:26.3724767Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_erfc_is_fastpath_True_cpu_float64 PASSED [ 70%] 2023-03-31T05:52:26.3725164Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_complex128 PASSED [ 70%] 2023-03-31T05:52:26.3725550Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_complex64 PASSED [ 70%] 2023-03-31T05:52:26.3725938Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_float32 PASSED [ 70%] 2023-03-31T05:52:26.3726323Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_False_cpu_float64 PASSED [ 70%] 2023-03-31T05:52:26.3726719Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_complex128 PASSED [ 70%] 2023-03-31T05:52:26.3727099Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_complex64 PASSED [ 70%] 2023-03-31T05:52:26.3727492Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_float32 PASSED [ 71%] 2023-03-31T05:52:26.3727882Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_exp_is_fastpath_True_cpu_float64 PASSED [ 71%] 2023-03-31T05:52:26.3728273Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_bfloat16 PASSED [ 71%] 2023-03-31T05:52:26.3728657Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_complex128 PASSED [ 71%] 2023-03-31T05:52:26.3729063Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_complex64 PASSED [ 71%] 2023-03-31T05:52:26.3729457Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_float32 PASSED [ 71%] 2023-03-31T05:52:26.3729837Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_False_cpu_float64 PASSED [ 71%] 2023-03-31T05:52:26.3730226Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_bfloat16 PASSED [ 71%] 2023-03-31T05:52:26.3730708Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_complex128 PASSED [ 71%] 2023-03-31T05:52:26.3731173Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_complex64 PASSED [ 71%] 2023-03-31T05:52:26.3731555Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_float32 PASSED [ 72%] 2023-03-31T05:52:26.3731945Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_expm1_is_fastpath_True_cpu_float64 PASSED [ 72%] 2023-03-31T05:52:26.3732340Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_bfloat16 PASSED [ 72%] 2023-03-31T05:52:26.3732737Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_float32 PASSED [ 72%] 2023-03-31T05:52:26.3733117Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_float64 PASSED [ 72%] 2023-03-31T05:52:26.3733505Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int16 PASSED [ 72%] 2023-03-31T05:52:26.3733899Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int32 PASSED [ 72%] 2023-03-31T05:52:26.3734273Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int64 PASSED [ 72%] 2023-03-31T05:52:26.3734659Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_int8 PASSED [ 72%] 2023-03-31T05:52:26.3735042Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_False_cpu_uint8 PASSED [ 72%] 2023-03-31T05:52:26.3735425Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_bfloat16 PASSED [ 72%] 2023-03-31T05:52:26.3735803Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_float32 PASSED [ 73%] 2023-03-31T05:52:26.3736188Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_float64 PASSED [ 73%] 2023-03-31T05:52:26.3736578Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int16 PASSED [ 73%] 2023-03-31T05:52:26.3736961Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int32 PASSED [ 73%] 2023-03-31T05:52:26.3737328Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int64 PASSED [ 73%] 2023-03-31T05:52:26.3737713Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_int8 PASSED [ 73%] 2023-03-31T05:52:26.3738098Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_floor_is_fastpath_True_cpu_uint8 PASSED [ 73%] 2023-03-31T05:52:26.3738484Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_bfloat16 PASSED [ 73%] 2023-03-31T05:52:26.3738862Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_float32 PASSED [ 73%] 2023-03-31T05:52:26.3739249Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_False_cpu_float64 PASSED [ 73%] 2023-03-31T05:52:26.3739640Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_bfloat16 PASSED [ 74%] 2023-03-31T05:52:26.3740018Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_float32 PASSED [ 74%] 2023-03-31T05:52:26.3740403Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_frac_is_fastpath_True_cpu_float64 PASSED [ 74%] 2023-03-31T05:52:26.3740798Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_complex128 PASSED [ 74%] 2023-03-31T05:52:26.3741193Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_complex64 PASSED [ 74%] 2023-03-31T05:52:26.3741570Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_float32 PASSED [ 74%] 2023-03-31T05:52:26.3741959Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_False_cpu_float64 PASSED [ 74%] 2023-03-31T05:52:26.3742541Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_complex128 PASSED [ 74%] 2023-03-31T05:52:26.3743021Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_complex64 PASSED [ 74%] 2023-03-31T05:52:26.3743400Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_float32 PASSED [ 74%] 2023-03-31T05:52:26.3743785Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log10_is_fastpath_True_cpu_float64 PASSED [ 74%] 2023-03-31T05:52:26.3744177Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_bfloat16 PASSED [ 75%] 2023-03-31T05:52:26.3744577Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_complex128 PASSED [ 75%] 2023-03-31T05:52:26.3744965Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_complex64 PASSED [ 75%] 2023-03-31T05:52:26.3745365Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_float32 PASSED [ 75%] 2023-03-31T05:52:26.3745758Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_False_cpu_float64 PASSED [ 75%] 2023-03-31T05:52:26.3746139Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_bfloat16 PASSED [ 75%] 2023-03-31T05:52:26.3746533Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_complex128 PASSED [ 75%] 2023-03-31T05:52:26.3746932Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_complex64 PASSED [ 75%] 2023-03-31T05:52:26.3747324Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_float32 PASSED [ 75%] 2023-03-31T05:52:26.3747700Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log1p_is_fastpath_True_cpu_float64 PASSED [ 75%] 2023-03-31T05:52:26.3748093Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_complex128 PASSED [ 76%] 2023-03-31T05:52:26.3748497Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_complex64 PASSED [ 76%] 2023-03-31T05:52:26.3748893Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_float32 PASSED [ 76%] 2023-03-31T05:52:26.3749268Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_False_cpu_float64 PASSED [ 76%] 2023-03-31T05:52:26.3749657Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_complex128 PASSED [ 76%] 2023-03-31T05:52:26.3750050Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_complex64 PASSED [ 76%] 2023-03-31T05:52:26.3750436Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_float32 PASSED [ 76%] 2023-03-31T05:52:26.3750808Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log2_is_fastpath_True_cpu_float64 PASSED [ 76%] 2023-03-31T05:52:26.3751199Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_complex128 PASSED [ 76%] 2023-03-31T05:52:26.3751597Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_complex64 PASSED [ 76%] 2023-03-31T05:52:26.3751981Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_float32 PASSED [ 76%] 2023-03-31T05:52:26.3752367Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_False_cpu_float64 PASSED [ 77%] 2023-03-31T05:52:26.3752759Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_complex128 PASSED [ 77%] 2023-03-31T05:52:26.3753149Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_complex64 PASSED [ 77%] 2023-03-31T05:52:26.3753527Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_float32 PASSED [ 77%] 2023-03-31T05:52:26.3753908Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_log_is_fastpath_True_cpu_float64 PASSED [ 77%] 2023-03-31T05:52:26.3754362Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_complex128 PASSED [ 77%] 2023-03-31T05:52:26.3754759Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_complex64 PASSED [ 77%] 2023-03-31T05:52:26.3755219Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_float32 PASSED [ 77%] 2023-03-31T05:52:26.3755605Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_float64 PASSED [ 77%] 2023-03-31T05:52:26.3755987Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int16 PASSED [ 77%] 2023-03-31T05:52:26.3756354Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int32 PASSED [ 78%] 2023-03-31T05:52:26.3756723Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int64 PASSED [ 78%] 2023-03-31T05:52:26.3757104Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_int8 PASSED [ 78%] 2023-03-31T05:52:26.3757490Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_False_cpu_uint8 PASSED [ 78%] 2023-03-31T05:52:26.3757868Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_complex128 PASSED [ 78%] 2023-03-31T05:52:26.3758262Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_complex64 PASSED [ 78%] 2023-03-31T05:52:26.3758648Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_float32 PASSED [ 78%] 2023-03-31T05:52:26.3759029Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_float64 PASSED [ 78%] 2023-03-31T05:52:26.3759402Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int16 PASSED [ 78%] 2023-03-31T05:52:26.3759783Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int32 PASSED [ 78%] 2023-03-31T05:52:26.3760158Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int64 PASSED [ 78%] 2023-03-31T05:52:26.3760528Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_int8 PASSED [ 79%] 2023-03-31T05:52:26.3760914Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_neg_is_fastpath_True_cpu_uint8 PASSED [ 79%] 2023-03-31T05:52:26.3761313Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_bfloat16 PASSED [ 79%] 2023-03-31T05:52:26.3761719Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_float32 PASSED [ 79%] 2023-03-31T05:52:26.3762109Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_False_cpu_float64 PASSED [ 79%] 2023-03-31T05:52:26.3762513Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_bfloat16 PASSED [ 79%] 2023-03-31T05:52:26.3762913Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_float32 PASSED [ 79%] 2023-03-31T05:52:26.3763473Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_reciprocal_is_fastpath_True_cpu_float64 PASSED [ 79%] 2023-03-31T05:52:26.3763858Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_bfloat16 PASSED [ 79%] 2023-03-31T05:52:26.3764256Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_float32 PASSED [ 79%] 2023-03-31T05:52:26.3764647Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_float64 PASSED [ 80%] 2023-03-31T05:52:26.3765038Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int16 PASSED [ 80%] 2023-03-31T05:52:26.3765413Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int32 PASSED [ 80%] 2023-03-31T05:52:26.3765800Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int64 PASSED [ 80%] 2023-03-31T05:52:26.3766186Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_int8 PASSED [ 80%] 2023-03-31T05:52:26.3766647Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_False_cpu_uint8 PASSED [ 80%] 2023-03-31T05:52:26.3767105Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_bfloat16 PASSED [ 80%] 2023-03-31T05:52:26.3767504Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_float32 PASSED [ 80%] 2023-03-31T05:52:26.3767899Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_float64 PASSED [ 80%] 2023-03-31T05:52:26.3768274Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int16 PASSED [ 80%] 2023-03-31T05:52:26.3768655Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int32 PASSED [ 80%] 2023-03-31T05:52:26.3769037Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int64 PASSED [ 81%] 2023-03-31T05:52:26.3769419Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_int8 PASSED [ 81%] 2023-03-31T05:52:26.3769795Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_round_is_fastpath_True_cpu_uint8 PASSED [ 81%] 2023-03-31T05:52:26.3770194Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_bfloat16 PASSED [ 81%] 2023-03-31T05:52:26.3770594Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_float32 PASSED [ 81%] 2023-03-31T05:52:26.3770994Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_False_cpu_float64 PASSED [ 81%] 2023-03-31T05:52:26.3771379Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_bfloat16 PASSED [ 81%] 2023-03-31T05:52:26.3771776Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_float32 PASSED [ 81%] 2023-03-31T05:52:26.3772172Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sigmoid_is_fastpath_True_cpu_float64 PASSED [ 81%] 2023-03-31T05:52:26.3772558Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_complex128 PASSED [ 81%] 2023-03-31T05:52:26.3772948Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_complex64 PASSED [ 81%] 2023-03-31T05:52:26.3773344Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_float32 PASSED [ 82%] 2023-03-31T05:52:26.3773731Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_False_cpu_float64 PASSED [ 82%] 2023-03-31T05:52:26.3774106Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_complex128 PASSED [ 82%] 2023-03-31T05:52:26.3774496Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_complex64 PASSED [ 82%] 2023-03-31T05:52:26.3774882Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_float32 PASSED [ 82%] 2023-03-31T05:52:26.3775267Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sin_is_fastpath_True_cpu_float64 PASSED [ 82%] 2023-03-31T05:52:26.3775652Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_complex128 PASSED [ 82%] 2023-03-31T05:52:26.3776054Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_complex64 PASSED [ 82%] 2023-03-31T05:52:26.3776449Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_float32 PASSED [ 82%] 2023-03-31T05:52:26.3776838Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_False_cpu_float64 PASSED [ 82%] 2023-03-31T05:52:26.3777216Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_complex128 PASSED [ 83%] 2023-03-31T05:52:26.3777611Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_complex64 PASSED [ 83%] 2023-03-31T05:52:26.3778001Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_float32 PASSED [ 83%] 2023-03-31T05:52:26.3778371Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sinh_is_fastpath_True_cpu_float64 PASSED [ 83%] 2023-03-31T05:52:26.3778818Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_bfloat16 PASSED [ 83%] 2023-03-31T05:52:26.3779285Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_complex128 PASSED [ 83%] 2023-03-31T05:52:26.3779685Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_complex64 PASSED [ 83%] 2023-03-31T05:52:26.3780065Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_float32 PASSED [ 83%] 2023-03-31T05:52:26.3780455Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_False_cpu_float64 PASSED [ 83%] 2023-03-31T05:52:26.3780841Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_bfloat16 PASSED [ 83%] 2023-03-31T05:52:26.3781232Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_complex128 PASSED [ 83%] 2023-03-31T05:52:26.3781615Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_complex64 PASSED [ 84%] 2023-03-31T05:52:26.3782010Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_float32 PASSED [ 84%] 2023-03-31T05:52:26.3782396Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_sqrt_is_fastpath_True_cpu_float64 PASSED [ 84%] 2023-03-31T05:52:26.3782876Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_complex128 PASSED [ 84%] 2023-03-31T05:52:26.3783259Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_complex64 PASSED [ 84%] 2023-03-31T05:52:26.3783647Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_float32 PASSED [ 84%] 2023-03-31T05:52:26.3784032Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_False_cpu_float64 PASSED [ 84%] 2023-03-31T05:52:26.3784408Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_complex128 PASSED [ 84%] 2023-03-31T05:52:26.3784800Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_complex64 PASSED [ 84%] 2023-03-31T05:52:26.3785195Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_float32 PASSED [ 84%] 2023-03-31T05:52:26.3785579Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tan_is_fastpath_True_cpu_float64 PASSED [ 85%] 2023-03-31T05:52:26.3785957Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_complex128 PASSED [ 85%] 2023-03-31T05:52:26.3786354Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_complex64 PASSED [ 85%] 2023-03-31T05:52:26.3786748Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_float32 PASSED [ 85%] 2023-03-31T05:52:26.3787139Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_False_cpu_float64 PASSED [ 85%] 2023-03-31T05:52:26.3787527Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_complex128 PASSED [ 85%] 2023-03-31T05:52:26.3787923Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_complex64 PASSED [ 85%] 2023-03-31T05:52:26.3788315Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_float32 PASSED [ 85%] 2023-03-31T05:52:26.3788684Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_tanh_is_fastpath_True_cpu_float64 PASSED [ 85%] 2023-03-31T05:52:26.3789071Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_bfloat16 PASSED [ 85%] 2023-03-31T05:52:26.3789466Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_float32 PASSED [ 85%] 2023-03-31T05:52:26.3789855Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_float64 PASSED [ 86%] 2023-03-31T05:52:26.3790229Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int16 PASSED [ 86%] 2023-03-31T05:52:26.3790678Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int32 PASSED [ 86%] 2023-03-31T05:52:26.3791120Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int64 PASSED [ 86%] 2023-03-31T05:52:26.3791506Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_int8 PASSED [ 86%] 2023-03-31T05:52:26.3791880Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_False_cpu_uint8 PASSED [ 86%] 2023-03-31T05:52:26.3792262Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_bfloat16 PASSED [ 86%] 2023-03-31T05:52:26.3792657Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_float32 PASSED [ 86%] 2023-03-31T05:52:26.3793048Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_float64 PASSED [ 86%] 2023-03-31T05:52:26.3793419Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int16 PASSED [ 86%] 2023-03-31T05:52:26.3793808Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int32 PASSED [ 87%] 2023-03-31T05:52:26.3794195Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int64 PASSED [ 87%] 2023-03-31T05:52:26.3794564Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_int8 PASSED [ 87%] 2023-03-31T05:52:26.3794946Z test_foreach.py::TestForeachCPU::test_unary_op__foreach_trunc_is_fastpath_True_cpu_uint8 PASSED [ 87%] 2023-03-31T05:52:26.3795371Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3795830Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_complex128 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3796277Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_complex64 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3796735Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float16 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3797189Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float32 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3797637Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_float64 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3798069Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int16 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T05:52:26.3798510Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int32 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3798947Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int64 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3799389Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_int8 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3799824Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_abs_cpu_uint8 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3800274Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_complex128 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3800730Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_complex64 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3801180Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_float32 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3801608Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_acos_cpu_float64 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3802122Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_complex128 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3802638Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_complex64 SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T05:52:26.3803212Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_float32 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3803643Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_asin_cpu_float64 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3804096Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_complex128 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3804548Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_complex64 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3805002Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_float32 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3805450Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_atan_cpu_float64 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3805886Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3806332Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_float32 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3806777Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_float64 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3807226Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3807656Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int32 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T05:52:26.3808114Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int64 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3808556Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_int8 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3808995Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_ceil_cpu_uint8 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3809428Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_complex128 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3809874Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_complex64 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3810319Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_float32 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3810766Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cos_cpu_float64 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3811206Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_complex128 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3811660Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_complex64 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3824867Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_float32 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3825344Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_cosh_cpu_float64 SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T05:52:26.3825809Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3826463Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_float32 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3827001Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erf_cpu_float64 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3827460Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3827916Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_float32 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3828371Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_erfc_cpu_float64 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3828814Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_complex128 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3829279Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_complex64 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3829739Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_float32 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3830186Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_exp_cpu_float64 SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T05:52:26.3830625Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3831084Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_complex128 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3831549Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_complex64 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3832009Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_float32 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3832452Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_expm1_cpu_float64 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3832903Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3833354Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_float32 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3833800Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_float64 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3834249Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int16 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3834681Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int32 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3835132Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int64 SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T05:52:26.3835578Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_int8 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3836023Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_floor_cpu_uint8 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3836460Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3836908Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_float32 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3837356Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_frac_cpu_float64 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3837935Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_complex128 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3838382Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_complex64 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3838834Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_float32 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3839281Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log10_cpu_float64 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3839733Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T05:52:26.3840180Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_complex128 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3840648Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_complex64 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3841100Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_float32 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3841544Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log1p_cpu_float64 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3841987Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_complex128 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3842440Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_complex64 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3842892Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_float32 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3843538Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log2_cpu_float64 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3843979Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_complex128 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3844432Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_complex64 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3844881Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_float32 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T05:52:26.3845328Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_log_cpu_float64 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3845762Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_complex128 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3846220Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_complex64 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3846672Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_float32 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3847116Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_float64 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3847544Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int16 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3847980Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int32 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3848412Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int64 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3848917Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_int8 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3849411Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_neg_cpu_uint8 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T05:52:26.3849869Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3850339Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_float32 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3850569Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cpu_float64 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3850790Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3851017Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_float32 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3851242Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_float64 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3851446Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int16 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3851660Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int32 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3851877Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int64 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3852096Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_int8 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3852321Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_round_cpu_uint8 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:52:26.3852550Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3852775Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_float32 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3852996Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cpu_float64 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3853218Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_complex128 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3853439Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_complex64 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3853661Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_float32 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3853866Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sin_cpu_float64 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3854087Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_complex128 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3854311Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_complex64 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3854528Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_float32 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:52:26.3854745Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sinh_cpu_float64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3854963Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3855320Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_complex128 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3855542Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_complex64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3855760Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_float32 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3855978Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_sqrt_cpu_float64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3856185Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_complex128 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3856405Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_complex64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3856630Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_float32 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3856851Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tan_cpu_float64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3857071Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_complex128 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:52:26.3857293Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_complex64 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3857511Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_float32 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3857731Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_tanh_cpu_float64 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3857955Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3858180Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_float32 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3858395Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_float64 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3858597Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int16 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3858809Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int32 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3859024Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int64 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3859248Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_int8 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T05:52:26.3859468Z test_foreach.py::TestForeachCPU::test_unary_op_tensors_on_different_devices__foreach_trunc_cpu_uint8 SKIPPED (Only runs on cuda) [100%] 2023-03-31T05:52:26.3859478Z 2023-03-31T05:52:26.3859929Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_foreach/test_foreach-b6d6f436a62737dc.xml - 2023-03-31T05:52:26.3860059Z ================= 717 passed, 338 skipped in 105.16s (0:01:45) ================= 2023-03-31T05:52:26.3860243Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:52:26.3860250Z 2023-03-31T05:52:26.3860598Z ##[endgroup] 2023-03-31T05:52:26.3860892Z FINISHED PRINTING LOG FILE of test_foreach (/var/lib/jenkins/workspace/test/test-reports/test_foreach_twhugs1k.log) 2023-03-31T05:52:26.3860951Z 2023-03-31T05:52:28.3709989Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:52:28.3736670Z Ignoring disabled issues: [] 2023-03-31T05:52:28.3898583Z Running dynamo/test_minifier ... [2023-03-31 05:52:28.389473] 2023-03-31T05:52:28.3900301Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_minifier.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:52:28.389778] 2023-03-31T05:53:08.9380381Z 2023-03-31T05:53:08.9380920Z Expand the folded group to see the log file of dynamo/test_minifier 2023-03-31T05:53:08.9381917Z ##[group]PRINTING LOG FILE of dynamo/test_minifier (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_minifier_v84gcr26.log) 2023-03-31T05:53:08.9382995Z Test results will be stored in test-reports/python-pytest/dynamo.test_minifier/dynamo.test_minifier-9a16a7cb30caca6b.xml 2023-03-31T05:53:08.9383511Z ============================= test session starts ============================== 2023-03-31T05:53:08.9384123Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:53:08.9384553Z cachedir: .pytest_cache 2023-03-31T05:53:08.9385230Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:53:08.9385763Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:53:08.9386431Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:53:08.9386881Z collecting ... collected 18 items 2023-03-31T05:53:08.9389764Z Running 18 items in this shard: test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_accuracy_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_accuracy_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_compile_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_compile_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_runtime_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_runtime_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_accuracy_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_accuracy_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_compile_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_compile_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_runtime_backend_passes, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_runtime_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_custom_backend, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_with_modified_config_cpu_accuracy_error, test/dynamo/test_minifier.py::MinifierTests::test_after_dynamo_with_modified_config_cpu_compile_error, test/dynamo/test_minifier.py::MinifierTests::test_cpu_cuda_module_after_dynamo, test/dynamo/test_minifier.py::MinifierTests::test_dynamo_config_serialization, test/dynamo/test_minifier.py::MinifierTests::test_if_graph_minified 2023-03-31T05:53:08.9392520Z 2023-03-31T05:53:08.9392800Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_accuracy_backend_passes PASSED [ 5%] 2023-03-31T05:53:08.9393344Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_accuracy_error PASSED [ 11%] 2023-03-31T05:53:08.9393902Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_compile_backend_passes PASSED [ 16%] 2023-03-31T05:53:08.9394458Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_compile_error PASSED [ 22%] 2023-03-31T05:53:08.9395020Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_runtime_backend_passes PASSED [ 27%] 2023-03-31T05:53:08.9395550Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cpu_runtime_error PASSED [ 33%] 2023-03-31T05:53:08.9396481Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_accuracy_backend_passes SKIPPED (requires cuda) [ 38%] 2023-03-31T05:53:08.9399246Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_accuracy_error SKIPPED (requires cuda) [ 44%] 2023-03-31T05:53:08.9399871Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_compile_backend_passes SKIPPED (requires cuda) [ 50%] 2023-03-31T05:53:08.9400500Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_compile_error SKIPPED (requires cuda) [ 55%] 2023-03-31T05:53:08.9401106Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_runtime_backend_passes SKIPPED (requires cuda) [ 61%] 2023-03-31T05:53:08.9401710Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_cuda_runtime_error SKIPPED (requires cuda) [ 66%] 2023-03-31T05:53:08.9402248Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_custom_backend PASSED [ 72%] 2023-03-31T05:53:08.9402826Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_with_modified_config_cpu_accuracy_error PASSED [ 77%] 2023-03-31T05:53:08.9406329Z dynamo/test_minifier.py::MinifierTests::test_after_dynamo_with_modified_config_cpu_compile_error PASSED [ 83%] 2023-03-31T05:53:08.9408213Z dynamo/test_minifier.py::MinifierTests::test_cpu_cuda_module_after_dynamo SKIPPED (requires cuda) [ 88%] 2023-03-31T05:53:08.9409079Z dynamo/test_minifier.py::MinifierTests::test_dynamo_config_serialization PASSED [ 94%] 2023-03-31T05:53:08.9409746Z dynamo/test_minifier.py::MinifierTests::test_if_graph_minified PASSED [100%] 2023-03-31T05:53:08.9410100Z 2023-03-31T05:53:08.9410887Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_minifier/dynamo.test_minifier-9a16a7cb30caca6b.xml - 2023-03-31T05:53:08.9411622Z ======================== 11 passed, 7 skipped in 38.58s ======================== 2023-03-31T05:53:08.9412204Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:53:08.9412606Z 2023-03-31T05:53:08.9413169Z ##[endgroup] 2023-03-31T05:53:08.9471485Z FINISHED PRINTING LOG FILE of dynamo/test_minifier (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_minifier_v84gcr26.log) 2023-03-31T05:53:08.9471992Z 2023-03-31T05:53:10.9929713Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:53:10.9956950Z Ignoring disabled issues: [] 2023-03-31T05:53:11.0119328Z Running test_tensorboard ... [2023-03-31 05:53:11.011575] 2023-03-31T05:53:11.0121243Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_tensorboard.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:53:11.011887] 2023-03-31T05:53:39.2599156Z 2023-03-31T05:53:39.2601142Z Expand the folded group to see the log file of test_tensorboard 2023-03-31T05:53:39.2602273Z ##[group]PRINTING LOG FILE of test_tensorboard (/var/lib/jenkins/workspace/test/test-reports/test_tensorboard_led2xc_s.log) 2023-03-31T05:53:39.2603596Z Test results will be stored in test-reports/python-pytest/test_tensorboard/test_tensorboard-1df86ba04af3d664.xml 2023-03-31T05:53:39.2603973Z ============================= test session starts ============================== 2023-03-31T05:53:39.2605955Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:53:39.2606376Z cachedir: .pytest_cache 2023-03-31T05:53:39.2644819Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:53:39.2645371Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:53:39.2645837Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:53:39.2646160Z collecting ... collected 53 items 2023-03-31T05:53:39.2651511Z Running 53 items in this shard: test/test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_autograd_np, test/test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_histogram, test/test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_histogram_raw, test/test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_np, test/test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_write, test/test_tensorboard.py::TestTensorBoardUtils::test_convert_to_HWC_dtype_remains_same, test/test_tensorboard.py::TestTensorBoardUtils::test_numpy_vid_uint8, test/test_tensorboard.py::TestTensorBoardUtils::test_prepare_video, test/test_tensorboard.py::TestTensorBoardUtils::test_to_HWC, test/test_tensorboard.py::TestTensorBoardWriter::test_writer, test/test_tensorboard.py::TestTensorBoardSummaryWriter::test_pathlib, test/test_tensorboard.py::TestTensorBoardSummaryWriter::test_summary_writer_close, test/test_tensorboard.py::TestTensorBoardSummaryWriter::test_summary_writer_ctx, test/test_tensorboard.py::TestTensorBoardEmbedding::test_embedding, test/test_tensorboard.py::TestTensorBoardEmbedding::test_embedding_64, test/test_tensorboard.py::TestTensorBoardSummary::test_audio, test/test_tensorboard.py::TestTensorBoardSummary::test_custom_scalars, test/test_tensorboard.py::TestTensorBoardSummary::test_empty_input, test/test_tensorboard.py::TestTensorBoardSummary::test_float32_image, test/test_tensorboard.py::TestTensorBoardSummary::test_histogram_auto, test/test_tensorboard.py::TestTensorBoardSummary::test_histogram_doane, test/test_tensorboard.py::TestTensorBoardSummary::test_histogram_fd, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_bool, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_domain_discrete, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_number, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_smoke, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_string, test/test_tensorboard.py::TestTensorBoardSummary::test_hparams_wrong_parameter, test/test_tensorboard.py::TestTensorBoardSummary::test_image_with_3_channel_batched, test/test_tensorboard.py::TestTensorBoardSummary::test_image_with_boxes, test/test_tensorboard.py::TestTensorBoardSummary::test_image_with_one_channel, test/test_tensorboard.py::TestTensorBoardSummary::test_image_with_one_channel_batched, test/test_tensorboard.py::TestTensorBoardSummary::test_image_without_channel, test/test_tensorboard.py::TestTensorBoardSummary::test_list_input, test/test_tensorboard.py::TestTensorBoardSummary::test_mesh, test/test_tensorboard.py::TestTensorBoardSummary::test_scalar_new_style, test/test_tensorboard.py::TestTensorBoardSummary::test_text, test/test_tensorboard.py::TestTensorBoardSummary::test_uint8_image, test/test_tensorboard.py::TestTensorBoardSummary::test_video, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_mlp_graph, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_nested_nn_squential, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_pytorch_graph, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_pytorch_graph_dict_input, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_torchvision_smoke, test/test_tensorboard.py::TestTensorBoardPytorchGraph::test_wrong_input_size, test/test_tensorboard.py::TestTensorBoardFigure::test_figure, test/test_tensorboard.py::TestTensorBoardFigure::test_figure_list, test/test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_np, test/test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_np_expect_fail, test/test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_simple_cnnmodel, test/test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_simple_model, test/test_tensorboard.py::TestTensorBoardNumpy::test_pytorch_np_expect_fail, test/test_tensorboard.py::TestTensorBoardNumpy::test_scalar 2023-03-31T05:53:39.2656472Z 2023-03-31T05:53:39.2656674Z test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_autograd_np PASSED [ 1%] 2023-03-31T05:53:39.2657066Z test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_histogram PASSED [ 3%] 2023-03-31T05:53:39.2657525Z test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_histogram_raw PASSED [ 5%] 2023-03-31T05:53:39.2657914Z test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_np PASSED [ 7%] 2023-03-31T05:53:39.2658367Z test_tensorboard.py::TestTensorBoardPyTorchNumpy::test_pytorch_write PASSED [ 9%] 2023-03-31T05:53:39.2658748Z test_tensorboard.py::TestTensorBoardUtils::test_convert_to_HWC_dtype_remains_same PASSED [ 11%] 2023-03-31T05:53:39.2659122Z test_tensorboard.py::TestTensorBoardUtils::test_numpy_vid_uint8 PASSED [ 13%] 2023-03-31T05:53:39.2659472Z test_tensorboard.py::TestTensorBoardUtils::test_prepare_video PASSED [ 15%] 2023-03-31T05:53:39.2659796Z test_tensorboard.py::TestTensorBoardUtils::test_to_HWC PASSED [ 16%] 2023-03-31T05:53:39.2660127Z test_tensorboard.py::TestTensorBoardWriter::test_writer PASSED [ 18%] 2023-03-31T05:53:39.2660483Z test_tensorboard.py::TestTensorBoardSummaryWriter::test_pathlib PASSED [ 20%] 2023-03-31T05:53:39.2660875Z test_tensorboard.py::TestTensorBoardSummaryWriter::test_summary_writer_close PASSED [ 22%] 2023-03-31T05:53:39.2661261Z test_tensorboard.py::TestTensorBoardSummaryWriter::test_summary_writer_ctx PASSED [ 24%] 2023-03-31T05:53:39.2661647Z test_tensorboard.py::TestTensorBoardEmbedding::test_embedding PASSED [ 26%] 2023-03-31T05:53:39.2662007Z test_tensorboard.py::TestTensorBoardEmbedding::test_embedding_64 PASSED [ 28%] 2023-03-31T05:53:39.2662350Z test_tensorboard.py::TestTensorBoardSummary::test_audio PASSED [ 30%] 2023-03-31T05:53:39.2662779Z test_tensorboard.py::TestTensorBoardSummary::test_custom_scalars PASSED [ 32%] 2023-03-31T05:53:39.2663141Z test_tensorboard.py::TestTensorBoardSummary::test_empty_input PASSED [ 33%] 2023-03-31T05:53:39.2663495Z test_tensorboard.py::TestTensorBoardSummary::test_float32_image PASSED [ 35%] 2023-03-31T05:53:39.2663837Z test_tensorboard.py::TestTensorBoardSummary::test_histogram_auto PASSED [ 37%] 2023-03-31T05:53:39.2664284Z test_tensorboard.py::TestTensorBoardSummary::test_histogram_doane PASSED [ 39%] 2023-03-31T05:53:39.2664641Z test_tensorboard.py::TestTensorBoardSummary::test_histogram_fd PASSED [ 41%] 2023-03-31T05:53:39.2664979Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_bool PASSED [ 43%] 2023-03-31T05:53:39.2665346Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_domain_discrete PASSED [ 45%] 2023-03-31T05:53:39.2665714Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_number PASSED [ 47%] 2023-03-31T05:53:39.2666065Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_smoke PASSED [ 49%] 2023-03-31T05:53:39.2666402Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_string PASSED [ 50%] 2023-03-31T05:53:39.2666766Z test_tensorboard.py::TestTensorBoardSummary::test_hparams_wrong_parameter PASSED [ 52%] 2023-03-31T05:53:39.2667152Z test_tensorboard.py::TestTensorBoardSummary::test_image_with_3_channel_batched PASSED [ 54%] 2023-03-31T05:53:39.2667514Z test_tensorboard.py::TestTensorBoardSummary::test_image_with_boxes PASSED [ 56%] 2023-03-31T05:53:39.2667885Z test_tensorboard.py::TestTensorBoardSummary::test_image_with_one_channel PASSED [ 58%] 2023-03-31T05:53:39.2668274Z test_tensorboard.py::TestTensorBoardSummary::test_image_with_one_channel_batched PASSED [ 60%] 2023-03-31T05:53:39.2668653Z test_tensorboard.py::TestTensorBoardSummary::test_image_without_channel PASSED [ 62%] 2023-03-31T05:53:39.2669005Z test_tensorboard.py::TestTensorBoardSummary::test_list_input PASSED [ 64%] 2023-03-31T05:53:39.2669346Z test_tensorboard.py::TestTensorBoardSummary::test_mesh PASSED [ 66%] 2023-03-31T05:53:39.2669696Z test_tensorboard.py::TestTensorBoardSummary::test_scalar_new_style PASSED [ 67%] 2023-03-31T05:53:39.2670029Z test_tensorboard.py::TestTensorBoardSummary::test_text PASSED [ 69%] 2023-03-31T05:53:39.2670374Z test_tensorboard.py::TestTensorBoardSummary::test_uint8_image PASSED [ 71%] 2023-03-31T05:53:39.2670713Z test_tensorboard.py::TestTensorBoardSummary::test_video PASSED [ 73%] 2023-03-31T05:53:39.2671149Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_mlp_graph PASSED [ 75%] 2023-03-31T05:53:39.2671519Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_nested_nn_squential PASSED [ 77%] 2023-03-31T05:53:39.2671992Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_pytorch_graph PASSED [ 79%] 2023-03-31T05:53:39.2672390Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_pytorch_graph_dict_input PASSED [ 81%] 2023-03-31T05:53:39.2672774Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_torchvision_smoke PASSED [ 83%] 2023-03-31T05:53:39.2673172Z test_tensorboard.py::TestTensorBoardPytorchGraph::test_wrong_input_size PASSED [ 84%] 2023-03-31T05:53:39.2673555Z test_tensorboard.py::TestTensorBoardFigure::test_figure SKIPPED (no matplotlib) [ 86%] 2023-03-31T05:53:39.2673936Z test_tensorboard.py::TestTensorBoardFigure::test_figure_list SKIPPED (no matplotlib) [ 88%] 2023-03-31T05:53:39.2674299Z test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_np SKIPPED (no caffe2) [ 90%] 2023-03-31T05:53:39.2674684Z test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_np_expect_fail SKIPPED (no caffe2) [ 92%] 2023-03-31T05:53:39.2675076Z test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_simple_cnnmodel SKIPPED (no caffe2) [ 94%] 2023-03-31T05:53:39.2675450Z test_tensorboard.py::TestTensorBoardNumpy::test_caffe2_simple_model SKIPPED (no caffe2) [ 96%] 2023-03-31T05:53:39.2675818Z test_tensorboard.py::TestTensorBoardNumpy::test_pytorch_np_expect_fail PASSED [ 98%] 2023-03-31T05:53:39.2676178Z test_tensorboard.py::TestTensorBoardNumpy::test_scalar PASSED [100%] 2023-03-31T05:53:39.2676365Z 2023-03-31T05:53:39.2676778Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorboard/test_tensorboard-1df86ba04af3d664.xml - 2023-03-31T05:53:39.2677161Z ======================== 47 passed, 6 skipped in 25.62s ======================== 2023-03-31T05:53:39.2677488Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:53:39.2677699Z 2023-03-31T05:53:39.2678043Z ##[endgroup] 2023-03-31T05:53:39.2678463Z FINISHED PRINTING LOG FILE of test_tensorboard (/var/lib/jenkins/workspace/test/test-reports/test_tensorboard_led2xc_s.log) 2023-03-31T05:53:39.2678705Z 2023-03-31T05:53:41.3011929Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:53:41.3040127Z Ignoring disabled issues: [] 2023-03-31T05:53:41.3236701Z Running test_transformers ... [2023-03-31 05:53:41.323339] 2023-03-31T05:53:41.3239359Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_transformers.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:53:41.323688] 2023-03-31T05:54:15.4881094Z 2023-03-31T05:54:15.4881545Z Expand the folded group to see the log file of test_transformers 2023-03-31T05:54:15.4882351Z ##[group]PRINTING LOG FILE of test_transformers (/var/lib/jenkins/workspace/test/test-reports/test_transformers__3uqokqt.log) 2023-03-31T05:54:15.5057984Z Test results will be stored in test-reports/python-pytest/test_transformers/test_transformers-38a31af97a106b47.xml 2023-03-31T05:54:15.5060277Z ============================= test session starts ============================== 2023-03-31T05:54:15.5060992Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:54:15.5061426Z cachedir: .pytest_cache 2023-03-31T05:54:15.5061952Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:54:15.5062493Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:54:15.5063246Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:54:15.5066325Z collecting ... collected 20223 items 2023-03-31T05:54:15.9752831Z Running 20223 items in this shard: test/test_transformers.py::TestTransformers::test_bias_is_none, test/test_transformers.py::TestTransformers::test_decoder_only_layer, test/test_transformers.py::TestTransformers::test_decoder_padding_and_src_mask_bool, test/test_transformers.py::TestTransformers::test_encoder_is_causal, test/test_transformers.py::TestTransformers::test_encoder_padding_and_src_mask_bool, test/test_transformers.py::TestTransformers::test_is_causal_gpu, test/test_transformers.py::TestTransformers::test_mask_check_fastpath, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_0_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_2_device_cpu, test/test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_5_device_cpu, test/test_transformers.py::TestTransformers::test_script_mha_in_proj_weight_none, test/test_transformers.py::TestTransformers::test_self_attn_TxT_attn_mask, test/test_transformers.py::TestTransformers::test_train_with_is_causal_device_cpu, test/test_transformers.py::TestTransformers::test_train_with_pad_and_catch_error_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_True_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_True_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_True_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_True_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_12, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_256, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_12, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_256, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_12, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_256, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_12, test/test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_256, test/test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_False_training_False_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_False_training_True_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_True_training_False_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_True_training_True_enable_nested_tensor_False_device_cpu, test/test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_1, test/test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_4, test/test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_8, test/test_transformers.py::TestSDPA::test_dispatch_fails_no_backend, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None, test/test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1, test/test_transformers.py::TestSDPA::test_flash_autocast_fp32_bfloat16, test/test_transformers.py::TestSDPA::test_flash_autocast_fp32_float16, test/test_transformers.py::TestSDPA::test_flash_backward_sm86_headdim128, test/test_transformers.py::TestSDPA::test_flash_fail_fp32, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_error_cases, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True, test/test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_query_dense, test/test_transformers.py::TestSDPA::test_fused_kernels_seq_len_0_inputs_fused_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_fused_kernels_seq_len_0_inputs_fused_kernel_SDPBackend_FLASH_ATTENTION, test/test_transformers.py::TestSDPA::test_fused_kernels_seq_len_1_inputs_fused_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_fused_kernels_seq_len_1_inputs_fused_kernel_SDPBackend_FLASH_ATTENTION, test/test_transformers.py::TestSDPA::test_fused_sdp_choice_type_dense, test/test_transformers.py::TestSDPA::test_fused_sdp_choice_type_nested, test/test_transformers.py::TestSDPA::test_incompatible_mask, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_attn_mask_present_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_broadcast_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_dim_3_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_head_dim_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_head_dim_kernel_SDPBackend_FLASH_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_fused_inputs_invalid_dtype_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_FLASH_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_MATH_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_FLASH_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_MATH_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_FLASH_ATTENTION_device_cpu, test/test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_MATH_device_cpu, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None, test/test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1, test/test_transformers.py::TestSDPA::test_memory_efficeint_sm86_failure, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel_SDPBackend_FLASH_ATTENTION, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel_SDPBackend_EFFICIENT_ATTENTION, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel_SDPBackend_FLASH_ATTENTION, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_True, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_True, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_False_head_dims_match_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_False_head_dims_match_True, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_True_head_dims_match_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_True_head_dims_match_True, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_False_head_dims_match_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_False_head_dims_match_True, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_True_head_dims_match_False, test/test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_True_head_dims_match_True, test/test_transformers.py::TestSDPA::test_sdp_choice_with_determinism_warn_only_False, test/test_transformers.py::TestSDPA::test_sdp_choice_with_determinism_warn_only_True, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_bfloat16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_float16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_bfloat16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_float16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_bfloat16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_float16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_bfloat16, test/test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_float16, test/test_transformers.py::TestSDPA::test_sdp_math_gradcheck_contiguous_inputs_False, test/test_transformers.py::TestSDPA::test_sdp_math_gradcheck_contiguous_inputs_True, test/test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_False, test/test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_True, test/test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_False, test/test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_True, test/test_transformers.py::TestSDPA::test_unaligned_tensors 2023-03-31T05:54:16.3331603Z 2023-03-31T05:54:16.3332007Z test_transformers.py::TestTransformers::test_bias_is_none PASSED [ 0%] 2023-03-31T05:54:16.3332419Z test_transformers.py::TestTransformers::test_decoder_only_layer SKIPPED (Fairseq not found) [ 0%] 2023-03-31T05:54:16.3333053Z test_transformers.py::TestTransformers::test_decoder_padding_and_src_mask_bool SKIPPED (not supported on pre-3.11 Python) [ 0%] 2023-03-31T05:54:16.3333457Z test_transformers.py::TestTransformers::test_encoder_is_causal PASSED [ 0%] 2023-03-31T05:54:16.3333969Z test_transformers.py::TestTransformers::test_encoder_padding_and_src_mask_bool SKIPPED (not supported on pre-3.11 Python) [ 0%] 2023-03-31T05:54:16.3334542Z test_transformers.py::TestTransformers::test_is_causal_gpu SKIPPED (Platform does not supposrt fused SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3334944Z test_transformers.py::TestTransformers::test_mask_check_fastpath PASSED [ 0%] 2023-03-31T05:54:16.3335415Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3335908Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3336356Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3336833Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3337315Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3337797Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3338591Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3339059Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3339513Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3339978Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3340435Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3340910Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3341389Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3341852Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3342299Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3342765Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3343358Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3343813Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3344343Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3344855Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3345330Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3345799Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3346245Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3346695Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3347171Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3347648Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3348124Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3348580Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_0_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3349043Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_2_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3349497Z test_transformers.py::TestTransformers::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_5_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3349939Z test_transformers.py::TestTransformers::test_script_mha_in_proj_weight_none PASSED [ 0%] 2023-03-31T05:54:16.3350510Z test_transformers.py::TestTransformers::test_self_attn_TxT_attn_mask SKIPPED (4D mask not supported yet - activate when 4D mask supported) [ 0%] 2023-03-31T05:54:16.3350929Z test_transformers.py::TestTransformers::test_train_with_is_causal_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3351374Z test_transformers.py::TestTransformers::test_train_with_pad_and_catch_error_device_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 0%] 2023-03-31T05:54:16.3351882Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3352372Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_True_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3352867Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3353371Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_True_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3353859Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3354342Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_True_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3354833Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3355326Z test_transformers.py::TestTransformers::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_True_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3355896Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_12 PASSED [ 0%] 2023-03-31T05:54:16.3356470Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_256 PASSED [ 0%] 2023-03-31T05:54:16.3357026Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_12 PASSED [ 0%] 2023-03-31T05:54:16.3357572Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_256 PASSED [ 0%] 2023-03-31T05:54:16.3358124Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_12 PASSED [ 0%] 2023-03-31T05:54:16.3358688Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_256 PASSED [ 0%] 2023-03-31T05:54:16.3359229Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_12 PASSED [ 0%] 2023-03-31T05:54:16.3359777Z test_transformers.py::TestTransformers::test_transformerencoder_fastpath_device_cpu_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_256 PASSED [ 0%] 2023-03-31T05:54:16.3360312Z test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_False_training_False_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3360832Z test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_False_training_True_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3361373Z test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_True_training_False_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3361939Z test_transformers.py::TestTransformers::test_transformerencoder_square_input_with_no_grad_True_training_True_enable_nested_tensor_False_device_cpu PASSED [ 0%] 2023-03-31T05:54:16.3362406Z test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_1 PASSED [ 0%] 2023-03-31T05:54:16.3362838Z test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_4 PASSED [ 0%] 2023-03-31T05:54:16.3363491Z test_transformers.py::TestTransformers::test_transformerencoderlayer_src_mask_device_cpu_nhead_8 PASSED [ 0%] 2023-03-31T05:54:16.3363952Z test_transformers.py::TestSDPA::test_dispatch_fails_no_backend SKIPPED (Platform does not support fused scaled dot product attention) [ 0%] 2023-03-31T05:54:16.3364714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3365523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3366303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3367097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3367933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3368756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3369542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3370315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3371099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3371885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3372650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3373425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3374280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3375067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3375842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3376618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3377388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3378166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3378939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3379711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3380526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3381316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3382094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3382946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3383741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3384522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3385309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3386071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3386922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3387704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3388478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3389252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3390032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3390800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3391580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3392392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3393197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3393972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3394752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3395516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3396292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3397070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3397840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3398681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3399450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3400230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3401005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3401791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3402580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3403478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3404242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3405076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3405916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3406702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3407482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3408264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3409028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3409811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3410590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3411440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3412217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3412993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3413760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3414531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3415316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3416088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3416906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3417698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3418474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3419244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3420030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3420806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3421585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3422359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3423295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3424074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3424856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3425641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3426428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3427193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3427974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3428753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3429564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3430376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3431154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3431913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3432696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3433474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3434246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3435028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3435856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3436629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3437408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3438189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3438978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3439746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3440516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3441340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3442154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3442927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3443818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3444609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3445379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3446161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3446949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3447841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3448627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3449403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3450171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3450954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3451736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3452509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3453285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3454097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3454911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3455681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3456459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3457241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3458015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3458769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3459546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3460382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3461157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3461935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3462709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3463570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3464345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3465128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3465947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3466756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3467530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3468293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3469073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3469854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3470628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3471396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3472201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 0%] 2023-03-31T05:54:16.3473012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3473785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3474555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3475329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3476108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3476867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3477634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3478443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3479247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3480028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3480793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3481566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3482343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3483240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3484014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3484882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3485669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3486436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3487212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3488000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3488779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3489556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3490310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3491130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3491951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3492728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3493497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3494275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3495036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3495808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3496578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3497416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3498189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3498967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3499729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3500515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3501293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3502070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3502962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3503784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3504550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3505330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3506101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3506879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3507655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3508420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3509250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3510024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3510797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3511565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3512336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3513105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3513876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3514658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3515467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3516272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3517057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3517822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3518604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3519393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3520170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3520945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3521780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3522545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3523428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3524200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3524975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3525742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3526512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3527352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3528170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3528948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3529723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3530499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3531266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3532043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3532820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3533639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3534456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3535241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3536004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3536786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3537562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3538337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3539113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3539954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3540747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3541518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3542293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3543151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3543932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3544710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3545477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3546320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3547099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3547879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3548662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3549430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3550214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3550986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3551764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3552575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3553381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3554146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3554931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3555711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3556490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3557276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3558056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3558884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3559664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3560450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3561234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3562021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3562793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3563644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3564473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3565283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3566061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3566838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3567609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3568386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3569154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3569926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3570780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3571556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3572313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3573086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3573872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3574652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3575430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3576205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3577004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3577814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3578587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3579359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3580143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3580920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3581675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3582460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3583393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3584172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3584949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3585722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3586502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3587281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3588047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3588855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3589665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3590430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3591205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3591997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3592782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3593550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3594327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3595151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3595934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3596706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3597481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3598269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3599050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3599814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3600584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3601398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3602199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3602972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3603830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3604617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3605396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3606173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3606945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3607807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3608580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3609357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3610126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3610918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3611708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3612484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3613290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3614108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3614882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3615656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3616430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3617213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3617975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3618749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3619558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3620375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3621144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3621905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3622674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3623527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3624302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3625073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3625890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3626678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3627462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3628232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3629013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3629787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3630567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3631320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 1%] 2023-03-31T05:54:16.3632164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3632943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3633723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3634495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3635282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3636049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3636823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3637594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3638406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3639205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3639970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3640747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3641519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3642294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3643170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3643948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3644803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3645580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3646351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3647122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3647907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3648680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3649440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3650247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3651057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3651839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3652622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3653405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3654165Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3654950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3655730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3656562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3657345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3658110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3658890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3659676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3660447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3661228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3662004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3662792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3663700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3664478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3665254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3666039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3666815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3667584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3668363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3669202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3669978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3670752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3671511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3672303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3673080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3673853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3674628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3675441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3676223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3676998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3677767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3678541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3679310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3680075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3680831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3681659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3682432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3683300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3684066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3684852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3685610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3686383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3687205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3688024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3688797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3689563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3690348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3691125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3691897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3692679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3693537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3694300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3695072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3695840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3696612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3697387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3698157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3698913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3699711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3700504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3701270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3702038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3702802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3703659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3704421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3705190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3706050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3706832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3707593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3708432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3709286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3710133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3710983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3711960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3712846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3713666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3714510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3715344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3716207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3717037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3717872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3718663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3719613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3720458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3721299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3722134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3722929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3724070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3724904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3725840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3726724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3727595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3728384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3729225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3730070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3730929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3731764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3732626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3733513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3734350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3735211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3736054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3736894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3737676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3738518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3739420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3740280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3741107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3741965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3742761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3743694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3744528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3745361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3746339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3747187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3747992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3748826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3749699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3750545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3751397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3752230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3753098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3753978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3754816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3755657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3756502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3757316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3758159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3758989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3759924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3760791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3761626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3762417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3763367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3764211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3765077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3765954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3766879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3767679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3768587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3769427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3770278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3771127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3771979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3772840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3773744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3774592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3775435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3776306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3777099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3777930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3778771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3779672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3780561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3781400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3782199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3783113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3783989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3784820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3785668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3786602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3787427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3788274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3789107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3789953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3790813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3791657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3792452Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3793327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3794228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3795093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3795925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3796724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3797570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 2%] 2023-03-31T05:54:16.3798426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3799265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3800199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3801041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3801833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3802693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3803639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3804477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3805309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3806234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3807097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3807936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3808771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3809644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3810489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3811283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3812128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3813051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3813949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3814791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3815644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3816449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3817309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3818146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3818989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3819865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3820786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3821577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3822413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3823348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3824181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3825045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3825877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3826733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3827604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3828469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3829309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3830156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3830953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3831806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3832674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3833547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3834448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3835284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3836117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3836967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3837800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3838630Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3839490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3858701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3859706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3860483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3861277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3862053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3862938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3863713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3864544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3865363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3866131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3866888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3867661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3868439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3869212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3869990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3870870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3871652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3872419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3873199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3873988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3874770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3875540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3876317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3877121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3877928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3878700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3879481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3880250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3881022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3881795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3882578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3883545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3884318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3885097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3885871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3886665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3887446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3888222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3889029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3889851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3890628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3891406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3892188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3892972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3893733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3894502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3895283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3896140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3896906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3897690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3898456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3899233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3900011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3900785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3901596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3902397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3903277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3904048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3904831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3905604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3906381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3907142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3907990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3908770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3909544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3910321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3911103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3911872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3912653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3913433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3914241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3915040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3915800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3916569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3917347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3918108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3918879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3919653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3920476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3921252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3922029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3922799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3923718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3924495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3925261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3926041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3926868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3927681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3928459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3929238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3930012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3930779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3931556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3932339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3933194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3933956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3934722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3935490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3936270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3937045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3937818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3938615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3939415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3940181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3940957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3941724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3942497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3943353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3944122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3944938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3945755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3946531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3947292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3948069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3948848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3949623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3950393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3951198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3951978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3952744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3953511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3954286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3955056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3955826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3956588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3957414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3958192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3958970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3959750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3960511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3961285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3962059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3962830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3963804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3964620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3965384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3966156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3966947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3967726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3968499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3969279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3970158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3970940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3971713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3972481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3973258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3974029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3974788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3975593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3976409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3977182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3977947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3978706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 3%] 2023-03-31T05:54:16.3979480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3980247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3981017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3981797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3982627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3983474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3984245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3985033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3985811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3986584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3987364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3988159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3988969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3989747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3990518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3991300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3992052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3992839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3993614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3994441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3995215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3995990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3996749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3997532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3998315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3999101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.3999885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4000700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4001497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4002279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4003197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4003985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4004769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4005542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4006312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4007177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4007956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4008731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4009500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4010268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4011053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4011834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4012615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4013433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4014237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4014999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4015776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4016557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4017328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4018107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4018900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4019737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4020510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4021292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4022076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4022932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4023708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4024492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4025310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4026115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4026893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4027674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4028436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4029218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4029992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4030770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4031623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4032409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4033165Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4033935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4034716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4035489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4036265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4037045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4037841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4038644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4039413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4040188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4040967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4041734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4042515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4043401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4044269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4045051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4045830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4046603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4047383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4048161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4048929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4049744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4050548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4051307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4052077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4052861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4053635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4054408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4055175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4056005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4056779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4057556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4058338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4059130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4059900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4060676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4061441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4062261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4063169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4063950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4064713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4065497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4066265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4067040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4067807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4068655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4069419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4070193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4070962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4071747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4072517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4073288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4074049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4074855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4075672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4076450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4077227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4077999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4078782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4079549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4080319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4081154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4081932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4082698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4083576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4084357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4085130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4085896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4086714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4087514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4088279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4089052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4089947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4090727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4091491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4092265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4093128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4093906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4094683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4095459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4096224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4097007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4097787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4098568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4099382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4100455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4101231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4102017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4102796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4103654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4104426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4105207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4106049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4106823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4107596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4108374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4109153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4109917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4110693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4111472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4112288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4113083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4113858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4114624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4115404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4116188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4116969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4117744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4118595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4119355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4120129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4120903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4121680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4122445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4123315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4124157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4124968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4125745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4126520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4127296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4128055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4128829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4129602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4130459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4131235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4132015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4132772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4133558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4134335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4135109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4135883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4136695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4137486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4138259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4139035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 4%] 2023-03-31T05:54:16.4139816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4140591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4141345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4142121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4143040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4143818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4144584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4145351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4146115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4146887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4147660Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4148424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4149242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4149730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4150203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4150682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4151163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4151617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4152096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4152574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4153118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4153593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4154071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4154546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4155022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4155494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4155967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4156521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4157031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4157494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4157969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4158445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4158920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4159390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4159863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4160392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4160869Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4161338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4161799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4162280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4162758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4163348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4163825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4164356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4164880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4165346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4165824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4166307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4166782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4167248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4167726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4168286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4168770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4169239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4169715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4170189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4170655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4171123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4171637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4172129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4172605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4173069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4173552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4174033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4174506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4174972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4175487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4175990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4176460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4176936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4177418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4177897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4178368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4178842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4179352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4179841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4180318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4180791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4181273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4181742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4182197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4182660Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4183292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4183773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4184249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4184721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4185205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4185677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4186151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4186618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4187127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4187624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4188097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4188568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4189058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4189529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4190000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4190467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4190998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4191467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4191925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4192394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4192872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4193345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4193816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4194313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4194817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4195286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4195754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4196229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4196711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4197185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4197657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4198187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4198665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4199145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4199612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4200086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4200561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4201036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4201503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4201988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4202487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4202955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4203536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4204005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4204474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4204939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4205414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4205967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4206445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4206913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4207388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4207856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4208335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4208808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4209285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4209788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4210305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4210779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4211251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4211727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4212192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4212664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4213134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4213701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4214175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4214643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4215124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4215595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4216063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4216530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4217033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4217528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4217998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4218468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4218945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4219418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4219889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4220364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4220892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4221368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4221825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4222301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4222779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4223337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4223812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4224283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4224793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4225292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4225762Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4226230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4226709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4227174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4227649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4228112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4228646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4229114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4229584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4230056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4230532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4231002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4231480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4231937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4232442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4232935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4233403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4233874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4234349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4234825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4235299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4235776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4236312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4236779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4237245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4237716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4238190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4238651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4239122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4239617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4240114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4240581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 5%] 2023-03-31T05:54:16.4241049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4241506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4241982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4242443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4242915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4243552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4244028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4244497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4244969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4245437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4245911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4246373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4246844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4247346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4247847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4248316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4248781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4249251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4249717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4250180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4250652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4251181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4251635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4252096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4252560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4253037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4253500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4253961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4254433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4254924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4255414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4255886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4256354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4256829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4257303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4257769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4258239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4258769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4259245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4259711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4260174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4260652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4261109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4261570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4262039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4262527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4263098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4263567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4264040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4264514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4264978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4265436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4265908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4266441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4266911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4267382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4267862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4268337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4268813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4269282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4269782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4270283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4270739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4271205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4271678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4272145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4272606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4273065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4273601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4274089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4274697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4275157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4275629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4276098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4276561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4277026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4277537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4278030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4278498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4278965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4279440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4279907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4280361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4280830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4281359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4281829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4282294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4282763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4283469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4283944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4284412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4284875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4285405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4285913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4286375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4286841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4287319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4287792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4288262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4288725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4289287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4289760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4290237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4290696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4291179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4291658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4292135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4292601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4293113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4293611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4294082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4294552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4295027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4295495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4295971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4296438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4296964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4297431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4297902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4298372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4298849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4299318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4299789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4300265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4300766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4301240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4301712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4302187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4302669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4303221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4303691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4304226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4304706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4305182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4305650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4306126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4306605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4307072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4307545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4308047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4308545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4309015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4309481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4309957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4310418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4310892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4311362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4311884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4312358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4312826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4313293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4313765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4314240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4314710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4315181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4315679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4316185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4316661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4317133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4317610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4318084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4318546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4319017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4319539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4320006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4320472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4320947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4321419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4321888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4322351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4322846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4323431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4323899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4324368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4324848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4325322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4325797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4326264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4326825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4327297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4327772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4328237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4328718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4329193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4329665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4330118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4330620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4331155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4331623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4332089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4332567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4333033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4333508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4333971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4334521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4334987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4335462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4335929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4336404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4336875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4337346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4337811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 6%] 2023-03-31T05:54:16.4338310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4338800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4339274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4339743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4340209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4340678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4341147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4341613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4342139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4342613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4343158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4343634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4344110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4344576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4345047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4345549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4346059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4346525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4347001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4347482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4347958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4348424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4348891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4349387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4349877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4350344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4350819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4351294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4351768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4352236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4352708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4353203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4353690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4354157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4354624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4355105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4355576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4356045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4356513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4357037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4357513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4357980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4358455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4358927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4359386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4359852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4360322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4360825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4361322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4361790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4362263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4362737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4363303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4363768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4364243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4364794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4365265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4365734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4366206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4366675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4367143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4367611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4368077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4368586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4369097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4369543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4370012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4370491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4370958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4371426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4371899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4372425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4372899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4373367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4373836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4374309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4374777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4375241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4375739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4376232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4376691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4377148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4377618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4378078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4378544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4378988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4379496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4380001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4380463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4380925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4381402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4381870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4382336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4382804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4383396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4383892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4384359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4384824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4385297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4385764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4386234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4386695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4387219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4387689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4388158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4388611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4389081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4389540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4390001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4390464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4391004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4391492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4391962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4392429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4392912Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4393381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4393850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4394313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4394845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4395314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4395788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4396255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4396731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4397200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4397669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4398144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4398632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4399118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4399583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4400046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4400519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4400982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4401445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4401907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4402426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4402892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4403458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4403923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4404398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4404859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4405323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4405845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4406360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4406827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4407296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4407765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4408233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4408696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4409166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4409681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4410180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4410653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4411113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4411579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4412046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4412501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4412962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4413453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4413946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4414407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4414866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4415333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4415803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4416275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4416738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4417262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4417721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4418185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4418650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4419117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4419586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4420050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4420519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4421010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4421512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4421973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4422434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4422982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4423453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4423916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4424380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4424902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4425380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4425845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4426313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4426790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4427254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4427727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4428204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4428712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4429217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4429689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4430164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4430644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4431121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4431594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4432067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4432597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4433072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4433545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4434024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4434501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 7%] 2023-03-31T05:54:16.4434973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4435446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4435946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4436457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4436934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4437389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4437864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4438337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4438806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4439277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4439809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4440284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4440761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4441234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4441715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4442186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4442660Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4443225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4443746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4444254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4444727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4445204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4445684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4446156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4446635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4447105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4447688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4448161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4448636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4449107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4449584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4450053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4450528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4451030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4451537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4452007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4452486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4452963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4453444Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4453920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4454397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4454930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4455410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4455879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4456353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4456828Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4457304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4457757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4458227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4458729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4459229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4459701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4460172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4460651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4461122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4461589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4462063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4462594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4463150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4463622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4464101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4464581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4465061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4465531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4466036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4466536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4467015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4467483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4467947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4468418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4468890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4469355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4469860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4470362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4470833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4471306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4471782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4472257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4472725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4473193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4473706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4474213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4474691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4475158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4475641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4476116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4476589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4477061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4477595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4478053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4478529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4479005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4479485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4479955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4480430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4480897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4481401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4481902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4482374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4482836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4483401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4483877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4484347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4484822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4485385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4485863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4486335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4486803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4487288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4487761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4488220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4488728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4489243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4489710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4490185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4490666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4491146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4491618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4492092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4492627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4493107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4493573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4494053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4494527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4495002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4495476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4495947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4496442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4496945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4497417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4497890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4498350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4498829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4499303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4499777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4500312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4500795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4501264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4501733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4502208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4502685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4503250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4503759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4504264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4504734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4505204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4505676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4506150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4506625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4507092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4507625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4508120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4508577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4509042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4509525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4509999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4510473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4510941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4511443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4511936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4512408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4512878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4513357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4513825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4514300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4514765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4515301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4515773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4516246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4516716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4517195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4517666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4518138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4518590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4519117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4519610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4520089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4520562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4521045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4521520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4521997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4522476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4523009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4523593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4524071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4524545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4525028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4525498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4525969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4526489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4527014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4527482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4527958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4528428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4528891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4529361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4529834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4530372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4530844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4531311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4531778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4532251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4532720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 8%] 2023-03-31T05:54:16.4533188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4533659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4534157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4534658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4535129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4535608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4536083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4536557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4537023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4537492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4538021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4538493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4538952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4539427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4539909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4540382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4540846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4541347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4541851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4542320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4542791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4543359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4543831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4544308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4544771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4545311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4545782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4546262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4546732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4547211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4547688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4548162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4548628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4549117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4549673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4550173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4550681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4551228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4551732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4552232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4552717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4553430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4553937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4554435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4554934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4555447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4555953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4556449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4556986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4557545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4558050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4558591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4559093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4559605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4560111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4560611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4561148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4561708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4562169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4562671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4563275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4563791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4564295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4564793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4565342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4565886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4566448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4566957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4567463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4567972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4568477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4568983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4569579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4570102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4570633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4571135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4571643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4572148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4572644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4573097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4573637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4574194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4574692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4575221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4575725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4576223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4576725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4577233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4577790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4578289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4578791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4579320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4579822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4580336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4580839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4581404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4581936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4582441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4583026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4583580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4584044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4584550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4585045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4585612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4586119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4586618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4587121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4587628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4588153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4588688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4589186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4589722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4590251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4590753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4591255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4591760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4592281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4592789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4593293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4593859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4594320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4594829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4595334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4595840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4596377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4596906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4597437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4597966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4598462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4598960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4599465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4599970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4600482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4601006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4601540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4602065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4602561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4603210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4603729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4604234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4604730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4605186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4605786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4606318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4606819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4607330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4607837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4608336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4608835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4609339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4609938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4610459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4610994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4611492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4611997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4612495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4612990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4613497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4614056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4614566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4715457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4716186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4716694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4717171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4717641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4718116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4718913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4719386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4719855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4720329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4720805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4721281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4721744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4722267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4722788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4723437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4723904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4724374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4724841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4725307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4725772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4726290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4726786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4727260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4727721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4728194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4728658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4729123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4729585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4730099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4730598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4731069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4731537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4732011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4732479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4732948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4733410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4733953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4734413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4734882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4735345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4735822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 9%] 2023-03-31T05:54:16.4736293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4736760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4737225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4737724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4738228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4738691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4739288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4739925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4740588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4741339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4742033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4742964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4743625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4744273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4744970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4745825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4746629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4747324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4748204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4749146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4750025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4750926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4751822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4752740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4753631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4754517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4755421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4756454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4757353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4758252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4759153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4760041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4760935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4761833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4762782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4763930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4764815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4765690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4766588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4767492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4768392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4769291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4770423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4771323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4772223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4773123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4774029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4774910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4775794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4776671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4777623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4778591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4779474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4780373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4781264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4782161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4783164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4784039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4785049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4785961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4786871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4787774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4788682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4789588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4790480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4791374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4792353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4793298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4794125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4794981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4795882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4796779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4797673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4798568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4799572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4800482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4801374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4802264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4803339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4804252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4805054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4806068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4807023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4807919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4808818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4809727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4810624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4811513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4812406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4813470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4814374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4815284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4816168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4817066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4817966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4818857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4819744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4820727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4821645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4822547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4823541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4824450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4825328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4826122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4827001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4828040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4828933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4829814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4830658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4831567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4832435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4833318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4834260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4835201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4836083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4836990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4837881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4838786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4839639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4840507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4841462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4842415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4843500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4844406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4845300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4846185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4847078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4847922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4848918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4849857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4850734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4851623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4852522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4853406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4854281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4855166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4856284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4857181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4858062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4858960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4859842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4860740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4861621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4862504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4863605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4864546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4865432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4866325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4867220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4868118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4869003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4869886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4870828Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4871626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4872383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4873050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4873827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4874551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4875248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4876022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4876817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4877455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4878135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4878919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4879724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4880551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4881262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4882023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4882661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4883764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4884430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4885080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4885719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4886353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4886986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4887728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4888413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4889051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4889671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4890323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4890949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4891570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4892201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4892966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4893605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4894234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4894867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4895518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4896156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4896788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4897419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4898105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4898776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4899407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4900035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4900670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4901315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4902014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4902645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4903474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4904104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4904735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 10%] 2023-03-31T05:54:16.4905356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4905995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4906626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4907252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4907927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4908612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4909236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4909872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4910511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4911151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4911782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4912420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4913139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4913772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4914442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4915085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4915716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4916361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4916985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4917627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4918308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4919046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4919787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4920455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4921186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4921875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4922636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4923658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4924690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4925553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4926313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4927094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4927909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4928692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4929478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4930268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4931120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4931873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4932507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4933223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4934024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4934733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4935496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4936236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4936941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4937579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4938234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4938913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4939550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4940179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4940805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4941517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4942197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4942969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4943709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4944385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4945076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4945822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4946566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4947511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4948272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4948968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4949669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4950467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4951263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4952123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4952801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4953532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4954235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4954878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4955523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4956185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4956821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4957477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4958127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4958869Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4959500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4960148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4960797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4961459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4962103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4962745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4963564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4964379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4965065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4965695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4966319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4966967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4967595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4968222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4968847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4969583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4970214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4970856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4971487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4972130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4972766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4973404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4974076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4974760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4975388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4976018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4976662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4977295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4977924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4978555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4979387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4980119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4980760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4981449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4982088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4982718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4983448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4984094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4984842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4985634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4986314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4987069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4987784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4988646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4989518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4990404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4991291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4992039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4992823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4993593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4994369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4995143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4995905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4996675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4997401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4998189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4998966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.4999697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5000458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5001182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5001822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5002465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5003468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5004129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5004766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5005419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5006108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5006772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5007527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5008199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5008988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5009762Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5010500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5011278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5012144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5012838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5013609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5014397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5015342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5016127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5016908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5017695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5018461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5019238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5019974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5020758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5021519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5022261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5023071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5023812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5024550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5025275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5026003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5026794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5027518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5028211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5028914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5029592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5030243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5030887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5031526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5032223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5032989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5033625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5034322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5035036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5035746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5036441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5037131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5037949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5038659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5039365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5040007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5040676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5041385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5042085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5042800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5044331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5045045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5045733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5046423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5047143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5047834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5048528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5049347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5050103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 11%] 2023-03-31T05:54:16.5050804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5051497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5052173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5052849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5053484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5054130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5054855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5055586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5056266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5056942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5057633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5058296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5058936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5059681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5060510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5061260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5061949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5062662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5063439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5064187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5064905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5065653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5066413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5067168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5067854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5068526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5069213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5069888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5070568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5071283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5072085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5072765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5073467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5074139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5074840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5075506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5076220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5076924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5077634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5078334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5079014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5079697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5080440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5081170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5081873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5082614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5083633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5084441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5085320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5086220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5087097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5087925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5088668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5089493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5090313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5091060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5091813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5092610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5093390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5094178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5094955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5095854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5096619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5097368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5098146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5098923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5099693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5100483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5101263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5102063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5102788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5103621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5104371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5105070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5105756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5106423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5107148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5108023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5108806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5109606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5110430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5111265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5112094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5112929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5113814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5114710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5115534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5116350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5117176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5118014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5118830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5119661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5120534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5121400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5122222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5123241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5124081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5124905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5125731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5126556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5127523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5128386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5129182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5129997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5130830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5131662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5132477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5133289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5134223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5135061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5135875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5136710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5137533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5138362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5139173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5139996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5140875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5141740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5142548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5143467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5144296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5145091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5145881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5146594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5147416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5148138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5148891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5149696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5150496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5151327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5152149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5153054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5153932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5154771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5155596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5156422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5157257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5158097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5158917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5159815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5160702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5161532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5162349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5163386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5164217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5165053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5165888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5166814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5167689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5168494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5169314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5170160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5170989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5171815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5172641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5173586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5174414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5175247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5176069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5176911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5177722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5178540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5179359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5180247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5181121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5181945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5182745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5183649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5184377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5185067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5185809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5186708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5187560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5188408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5189256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5190104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5190940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5191772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5192680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5193557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5194387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5195221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5196057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5196907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5197743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5198569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5199479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5200317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5201157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5201995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5202817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5203811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5204656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5205488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5206440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5207354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5208181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5209023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5209859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5210699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5211510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5212340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 12%] 2023-03-31T05:54:16.5213299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5214143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5214974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5215795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5216635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5217465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5218263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5219100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5219849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5220594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5221366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5222184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5223142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5223998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5224847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5225779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5226696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5227541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5228366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5229206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5230033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5230863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5231703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5232596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5233467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5234294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5235124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5235960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5236790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5237617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5238417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5239345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5240181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5241035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5241869Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5242731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5243752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5244602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5245489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5246437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5247317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5248153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5248983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5249810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5250642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5251473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5252277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5253208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5253914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5254633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5255370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5256199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5257015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5257867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5258800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5259702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5260537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5261390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5262232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5263195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5264039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5264879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5265775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5266663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5267491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5268311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5269145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5269989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5270807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5271641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5272533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5273385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5274219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5275058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5275899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5276737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5277572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5278391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5279318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5280157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5280982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5281837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5282680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5283659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5284374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5285202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5286138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5286955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5287662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5288411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5289205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5290028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5290834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5291682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5292695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5293552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5294377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5295224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5296059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5296887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5297697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5298586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5299485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5300309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5301123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5301971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5302909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5303760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5304595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5305486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5306379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5307222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5308048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5308907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5309748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5310573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5311378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5312283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5313132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5313957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5314785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5315610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5316432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5317267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5318058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5318897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5319620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5320372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5321138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5322000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5322845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5323907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5324743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5325715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5326593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5327436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5328257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5329104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5329938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5330759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5331573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5332565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5333396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5334215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5335039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5335883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5336713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5337546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5338373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5339275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5340149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5340989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5341831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5342672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5343621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5344453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5345283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5346244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5347083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5347910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5348714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5349472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5350208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5350956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5351834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5352731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5353576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5354434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5355285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5356140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5356970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5357807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5358717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5359597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5360434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5361269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5362093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5362943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5363972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5364813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5365748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5366651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5367485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5368336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5369182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5370027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5370863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5371685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5372631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5373479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5374315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5375148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5375994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5376792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5377571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5378282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5379137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5379976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5380809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5381449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5382077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5382707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5383451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 13%] 2023-03-31T05:54:16.5384074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5384824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5385461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5386088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5386723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5387368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5387995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5388614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5389243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5389908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5390579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5391210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5391829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5392452Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5393083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5393791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5394663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5395609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5396492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5397263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5398041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5398846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5399647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5400433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5401271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5402090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5402867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5403692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5404326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5405124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5405902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5406658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5407400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5408081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5408743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5409422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5410070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5410703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5411334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5411970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5412735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5413454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5414207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5414883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5415629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5416340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5417084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5417561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5418140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5418615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5419095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5419570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5420036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5420508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5420983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5421453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5421960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5422490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5423080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5423549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5424030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5424501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5424968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5425438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5425988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5426462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5426934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5427406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5427887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5428359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5428838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5429334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5429829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5430298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5430778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5431249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5431726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5432193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5432664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5433170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5433668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5434130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5434604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5435079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5435553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5436020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5436490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5436992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5437485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5437949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5438423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5438900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5439379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5439835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5440307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5440838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5441312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5441784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5442256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5442732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5443455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5443989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5444461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5445031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5445539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5446010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5446478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5446957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5447434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5447899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5448370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5448919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5449399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5449856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5450334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5450817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5451294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5451769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5452276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5452787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5453262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5453727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5454207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5454685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5455154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5455622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5456105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5456634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5457102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5457569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5458047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5458517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5458991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5459459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5459951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5460448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5460924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5461395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5461880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5462353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5462904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5463372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5463911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5464384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5464859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5465329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5465812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5466281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5466753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5467220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5467724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5468219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5468687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5469157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5469621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5470097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5470564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5471029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5471560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5472032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5472499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5472965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5473444Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5473919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5474392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5474857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5475369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5475867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5476338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5476806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5477285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5477757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5478225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5478692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5479237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5479692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5480157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5480629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5481102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5481568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5482033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5482564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5483262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5483740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5484207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5484680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5485153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5485619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5486086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5486621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5487136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5487608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5488073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5488542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5489015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5489471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5489943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5490458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5490961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5491427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5491892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5492359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 14%] 2023-03-31T05:54:16.5492822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5493289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5493758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5494289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5494764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5495229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5495700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5496170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5496646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5497121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5497589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5498104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5498609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5499069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5499540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5500018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5500493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5500962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5501437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5501968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5502441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5502985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5503452Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5503920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5504387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5504854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5505317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5505820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5506321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5506786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5507250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5507724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5508193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5508661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5509114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5509639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5510110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5510579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5511052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5511523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5511996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5512465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5512960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5513453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5513919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5514383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5514849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5515312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5515780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5516251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5516752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5517250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5517713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5518178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5518642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5519110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5519578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5520048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5520543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5521035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5521504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5521974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5522448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5522915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5523556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5524028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5524600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5525073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5525544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5526010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5526476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5526940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5527407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5527873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5528366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5528868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5529339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5529805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5530279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5530750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5531214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5531679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5532212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5532680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5533144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5533615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5534091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5534559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5535034Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5535500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5536007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5536501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5536970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5537432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5537888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5538351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5538819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5539286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5539861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5540329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5540801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5541265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5541742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5542204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5542680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5543233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5543746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5544240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5544715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5545190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5545671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5546139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5546610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5547079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5547593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5548064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5548536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5549002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5549476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5549939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5550406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5550903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5551398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5551868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5552341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5552813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5553285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5553750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5554221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5554723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5555226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5555694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5556164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5556643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5557119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5557575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5558049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5558553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5559057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5559524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5559995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5560470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5560939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5561400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5561864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5562391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5562863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5563511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5563991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5564467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5564940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5565405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5565879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5566412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5566915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5567370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5567837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5568312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5568787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5569252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5569720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5570274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5570756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5571219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5571698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5572172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5572636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5573105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5573606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5574103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5574573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5575039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5575514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5575989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5576465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5576931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5577389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5577918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5578389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5578863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5579346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5579823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5580299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5580763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5581265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5581768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5582237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5582709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5583257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5583731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5584196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5584659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5585197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5585670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5586140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5586606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5587072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5587543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5588013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5588482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5588993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5589491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 15%] 2023-03-31T05:54:16.5589960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5590426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5590906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5591382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5591852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5592320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5592864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5593338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5593808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5594275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5594748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5595218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5595689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5596150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5596646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5597138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5597607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5598073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5598555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5599019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5599483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5599949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5600513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5600988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5601460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5601932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5602412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5602880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5603528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5604051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5604569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5605036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5605505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5605970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5606447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5606901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5607369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5607835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5608387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5608863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5609331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5609795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5610273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5610741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5611208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5611697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5612207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5612674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5613142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5613609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5614081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5614549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5615017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5615512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5616011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5616460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5616929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5617400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5617867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5618330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5618791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5619285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5619779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5620249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5620713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5621191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5621661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5622130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5622592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5623208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5623685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5624150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5624616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5625087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5625557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5626010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5626477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5626975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5627474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5627947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5628415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5628884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5629353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5629814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5630274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5630794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5631263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5631734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5632199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5632669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5633135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5633603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5634068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5634558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5635056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5635514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5635985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5636461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5636935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5637404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5637874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5638411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5638885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5639352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5639825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5640298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5640775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5641240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5641735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5642236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5642706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5643372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5643852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5644324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5644793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5645262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5645780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5646288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5646766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5647231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5647711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5648181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5648647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5649121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5649631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5650139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5650607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5651077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5651548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5652020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5652493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5652960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5653490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5653959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5654427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5654895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5655358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5655825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5656292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5656755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5657257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5657785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5658255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5658721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5659198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5659664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5660140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5660610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5661145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5661615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5662085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5662558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5663108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5663583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5664052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5664520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5665009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5665518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5665995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5666463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5667050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5667518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5667988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5668461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5669003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5669474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5669943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5670411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5670885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5671356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5671824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5672326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5672822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5673288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5673760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5674235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5674708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5675168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5675637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5676132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5676639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5677100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5677563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5678041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5678511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5678976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5679442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5679939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5680428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5680898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5681366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5681841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5682307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5682776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5683429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5684010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5684486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5684940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5685413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5685893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5686367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5686841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 16%] 2023-03-31T05:54:16.5687315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5687824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5688324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5688782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5689251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5689728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5690197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5690668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5691134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5691671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5692143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5692606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5693080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5693558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5694024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5694481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5694952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5695453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5695950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5696420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5696891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5697366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5697835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5698302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5698772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5699303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5699773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5700242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5700715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5701191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5701661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5702130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5702628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5703234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5703702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5704169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5704633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5705112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5705580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5706045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5706557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5707059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5707530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5707994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5708473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5708943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5709415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5709884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5710381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5710876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5711342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5711803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5712280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5712746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5713220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5713688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5714209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5714683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5715151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5715619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5716098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5716571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5717046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5717506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5718042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5718543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5719013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5719481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5719960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5720434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5720903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5721368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5721896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5722364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5722827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5723458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5723928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5724403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5724866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5725329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5725855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5726351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5726816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5727275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5727745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5728209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5728669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5729131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5729677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5730155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5730625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5731086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5731564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5732037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5732501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5732961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5733448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5733934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5734396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5734854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5735325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5735789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5736256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5736718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5737237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5737697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5738162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5738618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5739090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5739553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5740016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5740475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5740974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5741467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5741932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5742390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5742935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5743411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5743879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5744352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5744890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5745352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5745815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5746275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5746752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5747210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5747673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5748157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5748645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5749105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5749563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5750023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5750496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5750954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5751418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5751871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5752380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5752848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5753312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5753775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5754250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5754717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5755183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5755676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5756174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5756637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5757101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5757563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5758036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5758503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5758970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5759462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5759952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5760412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5760876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5761331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5761794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5762253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5762717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5763441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5763954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5764417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5764881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5765343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5765807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5766268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5766730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5767229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5767725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5768188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5768661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5769126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5769598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5770066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5770528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5771002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5771507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5771974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5772438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5772904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5773381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5773852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5774319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5774822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5775350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5775823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5776292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5776767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5777245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5777712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5778185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5778693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5779187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5779655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5780119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5780577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5781049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5781518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5781983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5782519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5783071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5783541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5784006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 17%] 2023-03-31T05:54:16.5784483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5784959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5785430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5785897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5794770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5795419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5795909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5796392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5796879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5797363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5797837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5798311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5798841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5799318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5799783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5800252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5800725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5801198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5801669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5802140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5802640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5803312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5803781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5804253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5804729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5805202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5805671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5806144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5806735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5807213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5807687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5808159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5808634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5809092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5809560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5810070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5810586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5811056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5811524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5811989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5812459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5812928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5813398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5813898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5814410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5814881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5815348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5815821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5816294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5816768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5817237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5817738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5818227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5818686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5819154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5819629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5820099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5820575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5821041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5821570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5822038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5822506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5823070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5823545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5824007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5824474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5824944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5825455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5825945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5826419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5826890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5827364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5827830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5828285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5828759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5829278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5829739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5830205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5830673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5831144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5831617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5832086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5832553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5833049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5833546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5834014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5834478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5834944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5835404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5835871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5836333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5836857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5837319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5837770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5838244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5838721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5839185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5839655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5840126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5840629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5841114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5841585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5842051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5842522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5842987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5843631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5844100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5844686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5845156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5845618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5846080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5846549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5847014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5847464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5847973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5848480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5848943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5849409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5849882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5850350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5850808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5851273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5851745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5852271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5852738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5853208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5853678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5854155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5854624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5855099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5855594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5856092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5856557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5857008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5857476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5857944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5858405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5858875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5859380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5859884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5860345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5860810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5861281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5861752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5862211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5862675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5863252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5863752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5864212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5864678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5865153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5865624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5866086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5866539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5867071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5867542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5868004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5868475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5868944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5869411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5869872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5870341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5870865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5871352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5871815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5872278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5872745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5873210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5873665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5874131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5874647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5875115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5875571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5876024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5876502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5876974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5877439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5877912Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5878419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5878918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5879381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5879848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5880316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5880780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5881243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5881711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5882236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5882706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5883321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5883789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5884260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5884727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5885187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5885639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5886159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5886665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5887122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5887587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5888061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5888526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5888986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5889451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 18%] 2023-03-31T05:54:16.5890010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5890482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5890947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5891413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5891877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5892338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5892795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5893266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5893763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5894255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5894715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5895163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5895628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5896091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5896548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5897008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5897535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5898001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5898455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5898923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5899393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5899861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5900319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5900789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5901293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5901827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5902291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5902754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5903302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5903766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5904206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5904676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5905209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5905679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5906136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5906598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5907065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5907530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5907988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5908449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5908944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5909441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5909902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5910368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5910841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5911311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5911770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5912237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5912766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5913231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5913674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5914138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5914602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5915071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5915531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5916027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5916522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5916985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5917440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5917917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5918385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5918846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5919306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5919787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5920317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5920795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5921265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5921748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5922225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5922701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5923423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5923958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5924480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5924953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5925427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5925907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5926385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5926858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5927323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5927884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5928361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5928830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5929296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5929776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5930250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5930721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5931191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5931703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5932195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5932666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5933138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5933611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5934084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5934559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5935029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5935561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5936038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5936513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5936980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5937455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5937923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5938389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5938891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5939400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5939866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5940339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5940808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5941288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5941758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5942226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5942722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5943308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5943771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5944241Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5944711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5945198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5945667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5946139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5946644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5947145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5947620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5948094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5948571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5949047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5949513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5949986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5950510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5950988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5951458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5951925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5952401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5952878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5953346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5953802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5954308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5954809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5955276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5955754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5956233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5956713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5957182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5957656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5958182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5958658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5959125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5959597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5960071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5960543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5961010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5961542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5962044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5962516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5962993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5963614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5964074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5964544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5965015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5965584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5966055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5966531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5967002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5967481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5967946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5968422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5968890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5969403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5969903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5970379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5970850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5971324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5971796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5972263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5972732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5973259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5973719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5974197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5974668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5975142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5975610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5976083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5976548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5977047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5977543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5978019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5978487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5978964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5979434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5979908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5980380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5980909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5981376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5981854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5982327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5982803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5983354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5983815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5984313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5984813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5985279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5985751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5986222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5986700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 19%] 2023-03-31T05:54:16.5987167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5987639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5988168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5988641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5989107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5989580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5990060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5990535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5991004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5991480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5991993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5992501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5992975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5993452Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5993921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5994391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5994863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5995338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5995864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5996337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5996802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5997270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5997745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5998227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5998693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5999165Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.5999669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6000171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6000636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6001121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6001601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6002082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6002554Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6003193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6003739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6004216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6004689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6005166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6005642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6006111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6006586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6007099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6007609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6008076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6008544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6009018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6009490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6009961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6010430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6010958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6011426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6011901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6012368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6012852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6013325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6013786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6014254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6014765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6015261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6015731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6016196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6016681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6017158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6017631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6018096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6018664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6019131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6019601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6020066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6020545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6021017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6021488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6021953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6022459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6023041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6023517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6023976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6024464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6024934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6025401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6025872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6026414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6026882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6027354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6027833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6028310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6028778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6029253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6029763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6030260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6030727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6031196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6031671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6032145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6032613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6033086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6033599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6034079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6034549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6035015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6035491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6035964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6036432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6036900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6037398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6037899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6038368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6038841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6039316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6039796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6040266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6040737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6041263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6041737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6042202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6042676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6043277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6043739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6044211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6044725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6045236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6045705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6046176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6046649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6047119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6047592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6048060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6048611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6049085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6049560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6050032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6050508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6050982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6051457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6051925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6052427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6052921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6053397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6053857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6054332Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6054803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6055273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6055742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6056267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6056737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6057209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6057677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6058152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6058625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6059095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6059565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6060069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6060563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6061037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6061506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6061985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6062449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6063008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6063483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6064003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6064473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6064943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6065414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6065889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6066356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6066825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6067320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6067824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6068287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6068760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6069231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6069704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6070173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6070650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6071196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6071672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6072141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6072622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6073099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6073557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6074025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6074498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6075000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6075494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6075958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6076426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6076901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6077379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6077847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6078314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6078873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6079344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6079812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6080280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6080757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6081230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6081700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6082170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6082671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6083285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6083746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6084223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 20%] 2023-03-31T05:54:16.6084695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6085169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6085632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6086111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6086673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6087145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6087612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6088085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6088557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6089023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6089488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6089999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6090506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6090981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6091445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6091926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6092395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6092867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6093334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6093859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6094332Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6094808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6095273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6095750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6096225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6096697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6097170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6097687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6098188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6098863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6099339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6099819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6100285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6100755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6101225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6101778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6102250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6102717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6103262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6103729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6104199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6104679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6105142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6105654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6106155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6106623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6107089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6107567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6108040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6108513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6108980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6109511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6109980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6110454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6110920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6111398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6111865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6112328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6112819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6113304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6113776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6114246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6114718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6115193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6115657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6116119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6116619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6117119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6117586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6118056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6118528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6119005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6119476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6119945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6120439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6120940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6121417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6121892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6122363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6122836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6123411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6123877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6124439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6124918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6125383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6125856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6126329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6126802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6127271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6127741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6128250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6128755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6129227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6129701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6130175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6130653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6131125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6131597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6132125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6132605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6133065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6133540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6134020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6134491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6134962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6135426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6135924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6136459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6136930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6137397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6137872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6138344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6138816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6139286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6139820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6140298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6140773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6141245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6141720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6142192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6142647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6143225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6143731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6144202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6144674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6145150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6145625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6146091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6146551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6147020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6147543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6148015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6148483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6148957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6149425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6149892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6150359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6150858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6151357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6151829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6152290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6152767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6153245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6153723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6154195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6154702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6155199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6155678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6156152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6156629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6157101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6157574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6158044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6158549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6159043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6159514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6159984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6160462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6160931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6161400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6161863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6162381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6162852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6163459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6163931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6164408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6164884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6165361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6165828Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6166352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6166863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6167339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6167811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6168288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6168760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6169234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6169703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6170264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6170735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6171211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6171679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6172156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6172612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6173084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6173591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6174093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6174561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6175033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6175510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6175983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6176449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6176923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6177428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6177942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6178410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6178885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6179369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6179844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6180314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6180783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6181284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6181780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6182236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 21%] 2023-03-31T05:54:16.6182705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6183251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6183726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6184189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6184661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6185188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6185669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6186134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6186610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6187087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6187568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6188032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6188508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6189013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6189513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6189980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6190451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6190931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6191401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6191872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6192334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6192859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6193327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6193797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6194269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6194741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6195214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6195678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6196209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6196704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6197176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6197645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6198122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6198599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6199069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6199535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6200051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6200551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6201024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6201494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6201965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6202442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6202910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6203502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6204025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6204535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6205007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6205475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6205952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6206424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6206892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6207356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6207902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6208371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6208845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6209311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6209791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6210263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6210735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6211205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6211713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6212192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6212663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6213131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6213612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6214082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6214551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6215014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6215549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6216018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6216482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6216946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6217422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6217893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6218359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6218851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6219347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6219813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6220282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6220747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6221228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6221687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6222163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6222634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6223240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6223707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6224190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6224663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6225141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6225608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6226082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6226579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6227080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6227548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6228011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6228483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6228952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6229419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6229892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6230427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6230896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6231366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6231825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6232297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6232769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6233238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6233705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6234210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6234708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6235174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6235640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6236109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6236576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6237045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6237520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6238044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6238515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6238978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6239442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6239913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6240380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6240846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6241298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6241801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6242293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6242761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6243361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6243832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6244304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6244779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6245249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6245802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6246279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6246747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6247213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6247687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6248160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6248624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6249127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6249631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6250099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6250567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6251021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6251493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6251960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6252428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6252898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6253456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6253923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6254395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6254868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6255342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6255816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6256285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6256787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6257293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6257766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6258236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6258715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6259183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6259657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6260132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6260621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6261121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6261594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6262066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6262541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6263081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6263551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6264019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6264524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6265017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6265486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6265955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6266430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6266899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6267371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6267839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6268370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6268837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6269301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6269768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6270244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6270706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6271179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6271646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6272158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6272649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6273115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6273583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6274052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6274518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6274983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6275453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6275975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6276448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6276915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6277381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6277857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6278320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6278792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6279258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 22%] 2023-03-31T05:54:16.6279758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6280232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6280695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6281161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6281633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6282102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6282570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6283152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6283712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6284182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6284645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6285104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6285571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6286028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6286498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6287012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6287517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6287981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6288447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6288919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6289395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6289845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6290321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6290798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6291329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6291802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6292277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6292756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6293233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6293700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6294179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6294690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6295187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6295654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6296121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6296599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6297070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6297536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6298005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6298544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6299021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6299474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6299943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6300417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6300888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6301353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6301825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6302327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6302926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6303399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6303876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6304352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6304827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6305294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6305772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6306311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6306782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6307254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6307726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6308199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6308672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6309132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6309620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6310130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6310605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6311073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6311542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6312017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6312485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6312955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6313427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6313983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6314456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6314921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6315398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6315876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6316351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6316818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6317325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6317819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6318289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6318760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6319235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6319692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6320161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6320630Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6321156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6321631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6322104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6322569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6323177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6323651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6324122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6324584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6325106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6325609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6326080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6326548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6327024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6327495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6327966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6328435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6328978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6329443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6329913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6330383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6330860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6331330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6331798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6332294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6332793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6333259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6333727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6334191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6334669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6335137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6335611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6336079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6336622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6337092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6337567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6338042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6338514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6338978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6339440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6339936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6340434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6340898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6341372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6341846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6342322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6342788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6343337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6343871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6344343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6344806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6345274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6345742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6346211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6346683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6347151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6347652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6348149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6348616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6349078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6349551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6350023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6350489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6350957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6351480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6351953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6352418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6352886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6353361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6353830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6354294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6354759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6355255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6355748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6356213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6356683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6357153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6357623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6358087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6358557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6359067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6359540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6360013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6360490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6360964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6361437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6361913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6362417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6362921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6363528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6363998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6364470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6364943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6365416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6365882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6366417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6366917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6367389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6367853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6368329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6368784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6369256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6369720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6370223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6370761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6371234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6371699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6372178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6372648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6373121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6373589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6374120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6374587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6375060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6375530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6376006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6376479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 23%] 2023-03-31T05:54:16.6376955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6377419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6377919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6378394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6378862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6379326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6379795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6380264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6380729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6381196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6381729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6382195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6382674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6383227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6383713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6384184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6384658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6385133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6385645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6386143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6386620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6387094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6387570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6388039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6388496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6388970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6389513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6389984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6390457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6390934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6391414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6391879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6392352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6392852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6393353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6393821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6394292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6394766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6395242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6395704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6396181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6396712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6397187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6397659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6398136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6398601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6399073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6399544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6400017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6400515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6401011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6401478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6401949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6402429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6402906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6403500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6403976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6404537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6405009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6405478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6405947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6406424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6406895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6407362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6407872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6408367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6408841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6409310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6409793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6410273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6410740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6411207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6411675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6412203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6412672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6413141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6413620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6414091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6414564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6415033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6415543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6416046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6416514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6416983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6417463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6417933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6418389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6418859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6419384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6419855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6420325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6420797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6421272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6421743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6422214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6422683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6423272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6423777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6424249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6424720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6425199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6425667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6426134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6426599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6427133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6427606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6428063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6428529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6429013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6429484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6429954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6430423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6430961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6431455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6431935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6432403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6432882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6433348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6433824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6434291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6434829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6435295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6435758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6436226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6436707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6437172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6437639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6438119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6438617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6439088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6439561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6440035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6440513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6440977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6441448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6441943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6442449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6442919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6443510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6443989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6444461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6444927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6445399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6445914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6446418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6446884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6447353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6447814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6448279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6448745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6449212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6449753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6450226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6450687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6451151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6451621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6452096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6452566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6453037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6453543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6454041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6454511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6454986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6455459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6455933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6456401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6456869Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6457395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6457854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6458320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6458788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6459265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6459733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6460197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6460696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6461189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6461661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6462130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6462603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6463151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6463623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6464091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6464560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6465088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6465561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6466029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6466503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6466969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6467430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6467897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6468393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6468895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6469364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6469834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6470301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6470763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6471226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6471693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6472215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6472681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6473148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6473614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 24%] 2023-03-31T05:54:16.6474083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6474552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6475021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6475488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6475992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6476481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6476936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6477404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6477879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6478351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6478825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6479294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6479826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6480294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6480758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6481222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6481690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6482151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6482616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6483213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6483726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6484223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6484690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6485163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6485644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6486110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6486571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6487039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6487609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6488077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6488543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6489009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6489485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6489949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6490422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6490895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6491398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6491890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6492358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6492818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6493288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6493748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6494212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6494680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6495205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6495667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6496121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6496582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6497053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6497513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6497979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6498478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6498978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6499446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6499919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6500390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6500860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6501324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6501791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6502261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6502788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6503336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6503804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6504270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6504741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6505206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6505667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6506164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6506662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6507121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6507587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6508063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6508536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6508997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6509468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6509962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6510454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6510918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6511384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6511855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6512334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6512799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6513267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6513766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6514268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6514738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6515190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6515661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6516128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6516591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6517055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6517577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6518043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6518506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6518968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6519432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6519898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6520356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6520832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6521345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6521853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6522327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6522802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6523405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6523889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6524359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6524834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6525373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6525856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6526323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6526794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6527270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6527739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6528208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6528681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6529183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6529693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6530161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6530627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6531100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6531573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6532038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6532510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6533036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6533510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6533978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6534451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6534910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6535381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6535851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6536350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6536848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6537322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6537789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6538261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6538731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6539198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6539664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6540132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6540664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6541135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6541596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6542070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6542539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6543083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6543554Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6544069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6544555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6545026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6545502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6545980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6546446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6546916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6547384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6547967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6548437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6548910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6549379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6549860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6550325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6550791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6551255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6551755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6552247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6552721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6553183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6553664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6554127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6554588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6555053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6555581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6556050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6556517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6556984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6557463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6557933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6558404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6558866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6559369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6559861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6560331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6560798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6561277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6561745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6562211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6562673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6563332Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6563813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6564263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6564734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6565206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6565671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6566140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6566638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6567143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6567607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6568075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6568541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6569019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6569490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6569958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6570430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 25%] 2023-03-31T05:54:16.6570975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6571446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6571919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6572389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6572859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6573322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6573779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6574274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6574778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6575245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6575712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6576184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6576657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6577125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6577589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6578115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6578589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6579063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6579526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6579997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6580467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6580934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6581398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6581900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6582396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6582941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6583419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6583879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6584353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6584817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6585280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6585806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6586274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6586740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6587210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6587686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6588161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6588627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6589096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6589602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6590096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6590562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6591033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6591501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6591976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6592443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6592910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6593421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6593896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6594363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6594838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6595307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6595774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6606450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6607053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6607657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6608182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6608661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6609137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6609617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6610089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6610560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6611041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6611618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6612098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6612574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6613054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6613518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6613991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6614467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6614976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6615510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6615987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6616461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6616948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6617423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6617903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6618370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6618870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6619359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6619830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6620303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6620777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6621245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6621719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6622189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6622692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6623261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6623746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6624229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6624711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6625186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6625667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6626139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6626671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6627151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6627624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6628119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6628601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6629074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6629546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6630020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6630526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6631018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6631492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6631963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6632438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6632921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6633460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6638359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6639187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6639870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6640650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6641448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6642280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6643335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6644149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6644968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6645895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6646788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6647610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6648418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6649250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6649773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6650243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6650707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6651266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6651718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6652186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6652665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6653143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6653615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6654085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6654558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6655070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6655567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6656038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6656517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6657002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6657476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6657943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6658413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6658953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6659430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6659904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6660371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6660850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6661308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6661774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6662265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6662761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6663325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6663790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6664257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6664734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6665193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6665656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6666131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6666668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6667132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6667598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6668065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6668546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6669012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6669486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6669980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6670480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6670925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6671392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6671862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6672335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6672800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6673267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6673761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6674269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6674733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6675201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6675666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6676131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6676591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6677055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6677545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6678037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6678497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6678973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6679451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6679929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6680391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6680868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6681372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6681882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6682360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6682835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6683500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6683984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6684458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6684932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6685495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6685996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6686465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6686938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6687415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6687890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 26%] 2023-03-31T05:54:16.6688360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6688833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6689384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6689862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6690318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6690793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6691274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6691749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6692220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6692697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6693215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6693713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6694182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6694655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6695131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6695604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6696076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6696551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6697086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6697564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6698029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6698506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6698979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6699448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6699915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6700405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6700914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6701383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6701848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6702324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6702873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6703354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6703822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6704329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6704832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6705310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6705779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6706252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6706726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6707200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6707666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6708172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6708666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6709129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6709599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6710068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6710533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6711000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6711462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6711994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6712462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6712927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6713391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6713867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6714338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6714809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6715283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6715788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6716288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6716755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6717219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6717702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6718171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6718643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6719107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6719637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6720094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6720563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6721026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6721495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6721969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6722437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6722905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6723607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6724111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6724575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6725036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6725516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6725983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6726450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6726917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6727478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6727949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6728416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6728886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6729360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6729815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6730285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6730781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6731280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6731749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6732213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6732673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6733148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6733615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6734082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6734551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6735084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6735553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6736020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6736490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6736965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6737436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6737903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6738403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6738899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6739355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6739830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6740303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6740776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6741245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6741713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6742272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6742740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6743284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6743750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6744222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6744689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6745155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6745621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6746125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6746619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6747090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6747555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6748025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6748498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6748953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6749420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6749954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6750426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6750897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6751367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6751834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6752302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6752768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6753242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6753741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6754244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6754709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6755173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6755639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6756110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6756574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6757040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6757566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6758042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6758511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6758962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6759435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6759907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6760374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6760841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6761340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6761832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6762304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6762773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6763450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6763935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6764404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6764872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6765436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6765908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6766370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6766837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6767301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6767767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6768240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6768732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6769229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6769698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6770169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6770641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6771106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6771583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6772053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6772557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6773052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6773529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6773995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6774473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6774950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6775429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6775899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6776405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6776893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6777365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6777831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6778288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6778757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6779231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6779694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6780231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6780698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6781167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6781635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6782112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6782579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6783135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6783604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6784120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6784614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6785082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 27%] 2023-03-31T05:54:16.6785550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6786027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6786500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6786968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6787439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6787970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6788428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6788895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6789363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6789836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6790305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6790771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6791238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6791737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6792231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6792700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6793168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6793642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6794107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6794580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6795052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6795588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6796057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6796527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6796998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6797480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6797936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6798408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6798907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6799403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6799872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6800340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6800815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6801286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6801754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6802223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6802787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6803426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6803896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6804365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6804845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6805315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6805780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6806249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6806770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6807278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6807736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6808203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6808676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6809147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6809609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6810077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6810624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6811102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6811572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6812043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6812522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6812992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6813456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6813922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6814429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6814927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6815401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6815873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6816345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6816812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6817275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6817739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6818267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6818737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6819202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6819676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6820152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6820625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6821093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6821594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6822091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6822563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6823103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6823580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6824059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6824531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6824995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6825470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6826005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6826473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6826939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6827397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6827874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6828342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6828813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6829321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6829817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6830284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6830753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6831234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6831706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6832176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6832641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6833149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6833662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6834132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6834600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6835079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6835547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6836014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6836478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6836966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6837457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6837921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6838383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6838856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6839334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6839798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6840260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6840791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6841260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6841725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6842190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6842665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6843326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6843805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6844272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6844801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6845313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6845786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6846258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6846721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6847190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6847658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6848164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6849007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6850320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6851579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6852708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6854064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6855436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6856877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6858270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6859845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6861230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6862604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6864123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6865571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6866820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6868047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6869054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6869970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6870741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6871515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6872297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6873072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6873830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6874602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6875382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6876472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6877487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6878270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6879050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6879833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6880629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6881403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6882174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6883229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6884216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6885109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6885897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6886684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6887469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6888248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6889086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6889900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6890686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6891587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6892556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6893343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6894100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6894877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6895723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6896537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6897312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6898091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6898861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6899748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6900543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6901330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6902150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6903081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6903848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6904621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6905410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6906206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6906994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6907765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6908604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6909432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6910215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6910990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6911776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6912550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6913323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6914106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 28%] 2023-03-31T05:54:16.6914914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6915712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6916473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6917246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6918016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6918791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6919557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6920319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6921162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6921942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6922712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6923703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6924475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6925254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6926022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6926806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6927663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6928481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6929252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6930017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6930786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6931558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6932319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6933082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6933940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6934696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6935465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6936235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6937011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6937777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6938547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6939294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6940109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6940909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6941677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6942449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6943303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6944075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6944842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6945614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6946447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6947223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6947972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6948737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6949507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6950271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6951034Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6951785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6952582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6953379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6954150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6954911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6955682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6956443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6957214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6957977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6958802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6959576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6960357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6961115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6961898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6962674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6963680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6964455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6965306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6966137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6966912Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6967690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6968470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6969242Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6970000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6970778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6971618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6972393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6973162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6973938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6974711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6975489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6976258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6977087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6977890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6978654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6979423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6980200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6980980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6981758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6982528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6983381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6984221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6985005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6985782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6986562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6987342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6988101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6988876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6989688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6990476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6991248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6992006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6992786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6993562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6994336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6995106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6995905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6996687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6997460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6998233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6999010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.6999779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7000553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7001312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7002118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7002913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7003937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7004710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7005481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7006260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7007035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7007803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7008704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7009472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7010225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7010990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7011770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7012545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7013322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7014093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7014900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7015703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7016467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7017239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7018021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7018780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7019559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7020330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7021164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7021935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7022709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7023549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7024327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7025099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7025875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7026640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7027446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7028226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7028994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7029764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7030541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7031311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7032068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7032838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7033670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7034436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7035211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7035986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7036753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7037520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7038296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7039109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7039916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7040685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7041449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7042221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7043000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7043923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7044691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7045448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7046317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7047085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7047853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7048623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7049394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7050141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7050905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7051718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7052544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7053311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7054078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7054842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7055623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7056387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7057157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7057958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7058756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7059526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7060296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7061069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7061845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7062614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7063453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7064261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7065059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7065824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7066584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7067362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7068124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7068887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7069658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7070523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7071290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7072037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7072814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 29%] 2023-03-31T05:54:16.7073600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7074373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7075136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7075905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7076709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7077513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7078286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7079055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7079830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7080605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7081367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7082129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7082952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7083996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7084757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7085514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7086296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7087064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7087830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7088595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7089440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7090232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7090994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7091761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7092533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7093305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7094070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7094829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7095694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7096469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7097236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7098009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7098781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7099553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7100322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7101092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7101896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7102686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7103521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7104290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7105061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7105821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7106575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7107319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7108141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7108910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7109672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7110442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7111219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7111979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7112742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7113526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7114330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7115125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7115887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7116645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7117419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7118181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7118938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7119701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7120509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7121271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7122031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7122808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7123715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7124485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7125235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7125999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7126820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7127627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7128390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7129161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7129919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7130685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7131453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7132226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7133073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7133835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7134604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7135372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7136148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7136926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7137694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7138481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7139291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7140054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7140811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7141575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7142335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7143183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7143948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7144723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7145556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7146321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7147069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7147838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7148616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7149381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7150150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7150945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7151734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7152501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7153263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7154036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7154814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7155571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7156333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7157132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7157932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7158702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7159458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7160214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7160988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7161753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7162516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7163567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7164378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7165129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7165893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7166658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7167423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7168182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7168930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7169781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7170580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7171351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7172115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7172888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7173646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7174409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7175175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7175970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7176759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7177513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7178276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7179047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7179816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7180580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7181337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7182160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7183007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7183769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7184524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7185317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7186107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7186875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7187655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7188474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7189284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7190065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7190826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7191605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7192378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7193142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7193908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7194750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7195524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7196304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7197077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7197864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7198648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7199422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7200184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7200999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7201794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7202572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7203486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7204273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7205035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7205805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7206588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7207469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7208257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7209038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7209801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7210590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7211367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7212143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7212957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7213761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7214539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7215318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7216101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7216882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7217653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7218410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7219247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7220026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7220801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7221581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7222364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7223212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7223988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7224768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7225583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7226406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7227191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7227955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7228737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7229517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7230292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7231064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7231881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 30%] 2023-03-31T05:54:16.7232656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7233429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7234215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7234999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7235775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7236533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7237337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7238132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7238905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7239686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7240467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7241230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7242012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7242789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7243882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7244667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7245446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7246211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7246988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7247763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7248553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7249329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7250140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7250951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7251731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7252508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7252992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7253463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7253940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7254411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7254961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7255430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7255899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7256372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7256849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7257306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7257783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7258288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7258823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7259295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7259776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7260247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7260727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7261198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7261671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7262220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7262699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7263248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7263718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7264190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7264666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7265135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7265607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7266113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7266609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7267077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7267538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7268013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7268480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7268947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7269423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7269957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7270430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7270898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7271373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7271850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7272324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7272792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7273295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7273798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7274271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7274739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7275211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7275687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7276160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7276629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7277133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7277622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7278095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7278561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7279040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7279510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7279980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7280446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7280951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7281447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7281916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7282383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7282868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7283466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7283941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7284412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7284972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7285446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7285923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7286396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7286871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7287342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7287801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7288268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7288782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7289295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7289770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7290235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7290716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7291188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7291653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7292126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7292663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7293134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7293610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7294082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7294561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7295037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7295509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7296017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7296518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7296989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7297459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7297917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7298395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7298871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7299343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7299865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7300340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7300809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7301280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7301751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7302227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7302693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7303239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7303745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7304252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7304726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7305203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7305680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7306161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7306638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7307116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7307648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7308122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7308597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7309068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7309547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7310025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7310498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7323897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7324593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7325076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7325553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7326028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7326509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7326981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7327447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7327972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7328496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7328974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7329447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7329928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7330402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7330875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7331334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7331872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7332359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7332835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7333308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7333785Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7334262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7334730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7335198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7335736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7336201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7336678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7337146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7337622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7338094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7338569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7339038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7339538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7340035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7340514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7340985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7341451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7341922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7342402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7342961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7343508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7343982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7344460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7344933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7345411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7345879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7346354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7346863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7347362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7347831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7348305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7348780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7349255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7349725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7350199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7350727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 31%] 2023-03-31T05:54:16.7351205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7351664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7352144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7352621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7353100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7353573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7354052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7354555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7355065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7355537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7356014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7356492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7356972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7357442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7357915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7358445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7358922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7359391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7359872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7360350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7360823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7361291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7361787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7362286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7362755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7363374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7363867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7364352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7364827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7365304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7365868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7366340Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7366815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7367283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7367768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7368247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7368728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7369202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7369715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7370219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7370692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7371163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7371648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7372110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7372585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7373050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7373585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7374062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7374528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7375000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7375484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7375956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7376426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7376931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7377439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7377910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7378379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7378849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7379334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7379810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7380283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7380784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7381296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7381765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7382222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7382695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7383260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7383735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7384206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7384711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7385218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7385690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7386161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7386639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7387117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7387589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7388065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7388626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7389105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7389575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7390049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7390531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7391015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7391489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7391952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7392455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7392952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7393423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7393900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7394379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7394854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7395328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7395805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7396328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7396808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7397277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7397755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7398234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7398709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7399180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7399684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7400180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7400651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7401124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7401604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7402060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7402539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7403158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7403826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7404303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7404772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7405245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7405724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7406196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7406664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7407129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7407644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7408157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7408627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7409095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7409576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7410045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7410520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7410996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7411535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7412009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7412466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7412940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7413422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7413895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7414374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7414843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7415351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7415847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7416318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7416793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7417269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7417735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7418210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7418687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7419215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7419681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7420153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7420624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7421108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7421574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7422033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7422535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7423118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7423595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7424069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7424547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7425027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7425496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7425968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7426496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7426973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7427441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7427918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7428392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7428864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7429335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7429806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7430312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7430810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7431275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7431744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7432210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7432681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7433153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7433627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7434166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7434641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7435105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7435578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7436057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7436533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7436997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7437468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7437977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7438470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7438933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7439403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7439878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7440351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7440816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7441294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7441812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7442281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7442745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7443341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7443817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7444287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7444753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7445270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7445779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7446253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7446718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7447195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7447674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7448141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7448607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 32%] 2023-03-31T05:54:16.7449082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7449668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7450146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7450611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7451084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7451546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7452010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7452473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7452977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7453474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7453944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7454411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7454883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7455355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7455825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7456289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7456825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7457302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7457772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7458240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7458728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7459198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7459673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7460142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7460652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7461158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7461620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7462091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7462569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7463121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7463606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7464076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7464615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7465089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7465564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7466035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7466519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7466993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7467464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7467968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7468474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7468946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7469419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7469889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7470375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7470844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7471319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7471815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7472327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7472804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7473278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7473752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7474234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7474702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7475173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7475672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7476177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7476649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7477121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7477599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7478076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7478545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7479017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7479545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7480022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7480492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7480960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7481442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7481908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7482382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7482857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7483624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7484145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7484618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7485092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7485570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7486043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7486509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7486988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7487548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7488021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7488489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7488963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7489441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7489919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7490389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7490889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7491390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7491865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7492325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7492802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7493274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7493750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7494225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7494759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7495230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7495713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7496189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7496667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7497145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7497622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7498097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7498604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7499101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7499574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7500044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7500527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7500996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7501466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7501919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7502451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7503019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7503503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7503973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7504458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7504927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7505407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7505880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7506390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7506929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7507407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7507884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7508364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7508829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7509306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7509776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7510316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7510783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7511252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7511736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7512201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7512668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7513140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7513639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7514147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7514612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7515084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7515562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7516038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7516503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7516977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7517510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7517985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7518450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7518929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7519406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7519875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7520343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7520812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7521309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7521803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7522260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7522739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7523337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7523808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7524277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7524751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7525305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7525777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7526244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7526714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7527188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7527659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7528125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7528638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7529147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7529621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7530091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7530567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7531041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7531518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7531970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7532446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7532970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7533438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7533901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7534375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7534850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7535320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7535783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7536281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7536777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7537248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7537716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7538198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7538676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7539150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7539621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7540168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7540644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7541112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7541579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7542046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7542523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7543094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7543566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7544082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7544576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7545044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7545515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7545997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7546466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 33%] 2023-03-31T05:54:16.7546942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7547406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7547937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7548409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7548879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7549351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7549838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7550309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7550788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7551257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7551754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7552255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7552734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7553207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7553690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7554161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7554631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7555107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7555653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7556125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7556597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7557067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7557541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7558009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7558488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7558995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7559496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7559963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7560434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7560911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7561387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7561847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7562321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7562820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7563574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7564055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7564530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7565004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7565477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7565944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7566414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7566959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7567476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7567945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7568411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7568887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7569365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7569833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7570302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7570834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7571306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7571760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7572230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7572704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7573181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7573650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7574123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7574624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7575118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7575584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7576050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7576525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7577001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7577470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7577940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7578471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7578948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7579415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7579886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7580357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7580826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7581294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7581775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7582275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7582743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7583291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7583766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7584239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7584708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7585173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7585644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7586175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7586648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7587119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7587594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7588061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7588534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7589004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7589505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7590001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7590472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7590941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7591402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7591865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7592330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7592799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7593333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7593804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7594273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7594740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7595213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7595682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7596157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7596632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7597143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7597641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7598121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7598593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7599080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7599555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7600033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7600502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7601031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7601491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7601962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7602435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7602913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7603528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7604006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7604478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7605004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7605509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7605987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7606458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7606933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7607403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7607879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7608352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7608905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7609376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7609854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7610334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7610809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7611268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7611745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7612246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7612743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7613214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7613691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7614164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7614640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7615109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7615580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7616107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7616583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7617056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7617534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7618008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7618479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7618942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7619419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7619927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7620430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7620899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7621368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7621841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7622316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7622787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7623346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7623918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7624395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7624867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7625345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7625822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7626293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7626761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7627262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7627760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7628231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7628700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7629178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7629650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7630121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7630590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7631119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7631578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7632054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7632521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7633000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7633469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7633946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7634415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7634928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7635425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7635899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7636364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7636843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7637315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7637782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7638246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7638779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7639252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7639721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7640190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7640670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7641138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7641595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7642060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7642582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7643249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7643777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7644249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7644734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 34%] 2023-03-31T05:54:16.7645205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7645676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7646149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7646724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7647198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7647671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7648141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7648623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7649088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7649557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7650064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7650571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7651042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7651499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7651975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7652449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7652915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7653382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7653907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7654381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7654851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7655328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7655800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7656277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7656746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7657219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7657720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7658224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7658697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7659163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7659637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7660115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7660582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7661052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7661557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7662028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7662499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7663049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7663527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7664005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7664473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7664944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7665449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7665946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7666413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7666888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7667358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7667834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7668310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7668788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7669319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7669800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7670272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7670752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7671230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7671688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7672161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7672662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7673164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7673637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7674111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7674588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7675059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7675532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7676002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7676528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7677000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7677471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7677944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7678427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7678899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7679371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7679843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7680354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7680849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7681306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7681776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7682255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7682726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7683312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7683778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7684365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7684843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7685317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7685786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7686263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7686734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7687205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7687678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7688198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7688701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7689185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7689661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7690142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7690615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7691100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7691561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7692095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7692566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7693044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7693524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7694008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7694486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7694958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7695462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7695965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7696430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7696904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7697380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7697852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7698321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7698801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7699338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7699805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7700270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7700742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7701217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7701678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7702145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7702624Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7703209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7703708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7704177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7704653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7705134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7705609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7706080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7706551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7707080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7707562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7708033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7708506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7708981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7709447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7709910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7710415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7710909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7711377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7711827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7712309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7712787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7713260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7713728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7714239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7714741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7715218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7715691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7716172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7716647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7717116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7717587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7718088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7718587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7719058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7719527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7720007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7720478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7720950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7721413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7721933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7722406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7722873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7723569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7724056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7724530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7725003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7725475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7726010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7726526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7726996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7727468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7727951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7728421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7728894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7729371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7729925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7730399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7730866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7731339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7731810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7732274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7732744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7733244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7733749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7734215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7734684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7735151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7735630Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7736099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7736568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7737099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7737579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7738052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7738525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7738999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7739476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7739943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7740411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7740912Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7741442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7741898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7742370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 35%] 2023-03-31T05:54:16.7742907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7743386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7743854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7744327Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7744866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7745338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7745806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7746278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7746753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7747228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7747698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7748171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7748673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7749174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7749645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7750118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7750591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7751068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7751538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7752002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7752542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7753015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7753484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7753959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7754430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7754897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7755367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7755868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7756362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7756831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7757296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7757779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7758247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7758719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7759185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7759716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7760191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7760666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7761139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7761608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7762080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7762554Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7763153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7763678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7764178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7764653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7765120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7765592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7766062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7766530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7767000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7767545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7768024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7768495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7768968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7769442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7769911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7770385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7770857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7771361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7771844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7772316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7772786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7773268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7773734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7774208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7774681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7775210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7775676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7776155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7776626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7777097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7777565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7778037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7778541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7779047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7779514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7779981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7780456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7780930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7781382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7781851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7782377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7782923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7783406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7783877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7784356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7784826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7785293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7785758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7786266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7786768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7787235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7787706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7788178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7788641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7789097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7789566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7790093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7790565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7791034Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7791488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7791958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7792427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7792890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7793354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7793855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7794348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7794814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7795281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7795756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7796226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7796688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7797162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7797694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7798171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7798642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7799108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7799589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7800055Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7800517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7801070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7801625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7802120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7802589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7803224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7803745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7804213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7804677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7805151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7805737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7806208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7806677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7807149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7807626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7808100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7808564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7809076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7809590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7810061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7810529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7810991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7811466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7811933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7812391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7812892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7813386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7813859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7814325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7814798Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7815268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7815737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7816205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7816707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7817202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7817669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7818135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7818615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7819089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7819559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7820029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7820555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7821031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7821501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7821965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7822443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7822983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7823450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7823915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7824413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7824906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7825371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7825835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7826311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7826784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7827259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7827729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7828267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7828746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7829227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7829696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7830168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7830641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7831115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7831583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7832091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7832592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7833062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7833532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7834010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7834480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7834948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7835413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7835941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7836415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7836884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7837355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7837832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7838302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7838772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7839265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7839762Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 36%] 2023-03-31T05:54:16.7840219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7840688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7841160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7841639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7842107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7842580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7843239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7843726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7844200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7844670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7845144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7845618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7846084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7846557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7847065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7847572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7848040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7848514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7848987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7849462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7849926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7850383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7850915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7851390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7851853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7852323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7852799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7853275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7853743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7854214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7854707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7855203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7855671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7856139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7856609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7857081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7857544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7858017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7858583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7859052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7859525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7859985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7860458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7860926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7861396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7861906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7862401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7862953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7863431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7863907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7864375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7864847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7865316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7865853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7866324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7866792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7867264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7867745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7868214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7868693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7869159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7869657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7870138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7870611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7871078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7871556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7872027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7872494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7872963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7873490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7873959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7874432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7874898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7875382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7875853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7876321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7876789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7877288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7877784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7878254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7878721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7879197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7879657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7880126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7880589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7881108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7881572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7882037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7882508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7882983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7883737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7884211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7884729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7885245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7885707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7886177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7886650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7887131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7887604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7888079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7888592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7889103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7889573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7890031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7890502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7890973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7891435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7891906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7892406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7892896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7893363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7893831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7894305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7894777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7895240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7895704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7896227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7896696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7897161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7897638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7898109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7898581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7899048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7899504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7900006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7900501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7900970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7901438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7901915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7902385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7902926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7903399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7903928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7904400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7904871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7905341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7905813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7906278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7906743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7907211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7907706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7908210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7908671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7909138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7909601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7910072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7910541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7911012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7911550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7912018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7912485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7912957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7913431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7913896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7914364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7914860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7915356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7915823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7916286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7916759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7917233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7917697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7918162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7918698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7919181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7919648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7920117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7920592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7921062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7921530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7921998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7922506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7922996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7923600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7924064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7924542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7925015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7925486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7925954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7926514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7926981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7927444Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7927902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7928377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7928836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7929304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7929770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7930279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7930786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7931253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7931723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7932200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7932673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7933139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7933603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7934138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7934614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7935086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7935552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7936030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7936504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7936982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 37%] 2023-03-31T05:54:16.7937451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7937950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7938436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7938902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7939364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7939837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7940307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7940775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7941243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7941776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7942249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7942716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7943266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7943745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7944213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7944683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7945185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7945700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7946166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7946632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7947102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7947584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7948057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7948517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7949018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7949517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7949986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7950448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7950914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7951384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7951860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7952330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7952827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7953323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7953792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7954258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7954728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7955199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7955665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7956138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7956655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7957129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7957599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7958053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7958527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7959001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7959467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7959932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7960434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7960928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7961389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7961857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7962325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7962794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7963479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7963954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7964523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7964999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7965466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7965934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7966403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7966884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7967359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7967819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7968333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7968849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7969321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7969796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7970271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7970752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7971224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7971699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7972231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7972710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7973185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7973657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7974128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7974603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7975071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7975569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7976102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7976575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7977048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7977529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7977992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7978468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7978935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7979465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7979939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7980417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7980892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7981376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7981849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7982334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7982803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7983397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7983899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7984378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7984852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7985332Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7985800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7986276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7986745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7987275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7987746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7988208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7988688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7989171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7989643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7990119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7990613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7991127Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7991598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7992073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7992553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7993030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7993495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7993970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7994509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7994990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7995458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7995933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7996410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7996879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7997346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7997820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7998308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7998804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7999275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.7999746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8000226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8000699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8001171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8001645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8002181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8002661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8003269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8003749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8004224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8004697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8005166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8005721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8006234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8006704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8007174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8007646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8008123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8008584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8009049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8009570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8010081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8010561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8011028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8011502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8011974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8012441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8012908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8013423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8013922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8014394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8014866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8015351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8015829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8016304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8016773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8017315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8017792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8018263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8018719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8019194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8019671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8020138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8020604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8021105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8021599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8022069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8022539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8023104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8023584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8024049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8024518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8025057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8025532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8026010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8026479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8026962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8027433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8027903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8028409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8028903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8029376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8029855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8030330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8030806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8031275Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8031747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8032282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8032760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8033228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8033695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8034171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8034649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 38%] 2023-03-31T05:54:16.8035118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8035585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8036120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8036625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8037100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8037575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8038056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8038531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8038984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8039453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8039996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8040475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8040947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8041422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8041900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8042370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8042837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8043609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8044143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8044619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8045089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8045570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8046045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8046523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8046992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8047507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8048010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8048489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8048959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8049424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8049903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8050376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8050848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8051355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8051851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8052328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8052801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8053279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8053750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8054225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8054691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8055217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8055684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8056157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8056625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8057106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8057574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8058047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8058513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8059008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8059513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8059993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8060468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8060957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8061436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8061913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8062391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8063022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8063500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8063978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8064455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8064936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8065405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8065880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8066391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8066887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8067362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8067842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8068317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8068794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8069253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8069729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8070266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8070748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8071221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8071694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8072178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8072656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8073131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8073607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8074116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8074615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8075083Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8075568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8076050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8076528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8077000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8077476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8078013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8078488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8078961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8079442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8079905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8080377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8080846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8081356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8081861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8082338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8082803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8083437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8083920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8084391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8084861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8085431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8085914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8086390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8086858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8087342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8087817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8088292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8088761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8089277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8089789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8090244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8090717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8091199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8091665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8092135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8092603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8093168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8093642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8094114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8094584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8095066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8095540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8096010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8096509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8097027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8097499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8097976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8098451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8098941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8099412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8099880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8100367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8100867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8101335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8101807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8102283Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8102763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8103334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8103804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8104317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8104816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8105286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8105760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8106236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8106718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8107191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8107663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8108194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8108675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8109140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8109612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8110073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8110550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8111021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8111493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8111995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8112493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8112961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8113437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8113913Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8114388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8114853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8115328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8115861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8116337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8116806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8117276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8117753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8118228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8118695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8119197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8119690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8120156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8120626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8121104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8121572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8122046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8122516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8123219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8123777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8124251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8124723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8125205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8125677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8126154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8126620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8127148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8127658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8128130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8128595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8135214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8135781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8136277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8136749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8137369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8137850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8138333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8138810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8139296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8139768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 39%] 2023-03-31T05:54:16.8140243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8145402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8149608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8150075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8150552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8151071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8151572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8152049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8152529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8153049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8153554Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8154014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8154490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8154970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8155450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8155929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8156411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8156930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8157431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8157906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8158385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8158865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8159344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8159814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8160297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8160824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8161302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8161774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8162247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8162728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8163355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8163829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8164291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8164819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8165329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8165795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8166265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8166740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8167211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8167679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8168158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8168700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8169179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8169646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8170121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8170599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8171072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8171543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8172039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8172547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8173018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8173488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8173953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8174426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8174895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8175362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8175867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8176378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8176853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8177317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8177793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8178264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8178736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8179203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8179715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8180212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8180677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8181147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8181627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8182100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8182573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8183140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8183690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8184149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8184625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8185092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8185572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8186044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8186518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8186985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8187494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8187988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8188459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8188922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8189399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8189873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8190341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8190808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8191339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8191813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8192282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8192747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8193227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8193689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8194164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8194635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8195134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8195637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8196107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8196571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8197045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8197518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8197990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8198454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8198982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8199451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8199919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8200388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8200867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8201333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8201807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8202305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8202808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8203524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8203994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8204470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8204950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8205418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8205885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8206453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8206936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8207402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8207875Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8208350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8208823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8209292Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8209763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8210276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8210782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8211253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8211720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8212199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8212674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8213140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8213592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8214120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8214598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8215074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8215547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8216029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8216509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8216978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8217457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8217966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8218470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8218943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8219420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8219901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8220373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8220844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8221323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8221851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8222333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8222804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8223373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8223839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8224316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8224793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8225302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8225800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8226278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8226753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8227235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8227713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8228190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8228662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8229201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8229674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8230153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8230629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8231109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8231581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8232060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8232529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8233041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8233537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8234001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8234469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8234950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8235420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8235894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8236361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8236899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8237374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8237850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8238323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8238806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8239281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8239757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8240225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8240739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8241244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8241726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8242203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8242692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8243302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8243778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8244299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8244810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8245287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8245765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8246244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 40%] 2023-03-31T05:54:16.8246722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8247189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8247666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8248183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8248690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8249158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8249631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8250113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8250587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8251057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8251534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8252068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8252544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8253015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8253493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8253963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8254441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8254912Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8255384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8255888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8256396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8256864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8257339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8257819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8258300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8258769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8259245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8259771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8260243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8260705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8261179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8261658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8262132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8262601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8263199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8263710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8264168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8264647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8265128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8265602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8266075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8266543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8267081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8267550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8268025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8268488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8268963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8269432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8269901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8270374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8270878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8271374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8271843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8272308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8272783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8273259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8273731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8274188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8274724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8275199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8275671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8276138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8276626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8277100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8277574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8278042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8278576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8279075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8279546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8280013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8280494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8280964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8281432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8281898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8282437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8282910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8283625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8284085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8284566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8285042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8285514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8286041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8286556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8287022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8287491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8287967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8288448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8288917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8289392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8289942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8290421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8290889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8291357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8291834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8292308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8292779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8293248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8293751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8294234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8294703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8295173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8295647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8296120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8296587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8297061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8297591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8298064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8298536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8299009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8299485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8299959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8300432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8300904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8301403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8301903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8302367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8302914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8303394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8303853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8304319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8304793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8305321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8305797Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8306270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8306738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8307214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8307694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8308161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8308667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8309174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8309647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8310116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8310588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8311065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8311538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8312002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8312504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8313001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8313471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8313926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8314402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8314871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8315342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8315809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8316313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8316810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8317281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8317747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8318227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8318695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8319168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8319636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8320184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8320654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8321131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8321598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8322077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8322546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8323144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8323599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8324122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8324629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8325103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8325569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8326041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8326515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8326982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8327452Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8328013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8328483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8328952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8329422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8329900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8330366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8330839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8331306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8331816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8332306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8332777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8333246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8333710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8334183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8334653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8335126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8335658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8336128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8336600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8337072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8337543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8338007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8338474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8338973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8339467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8339931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8340403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8340877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8341352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8341816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8342287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8342889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8343355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8343826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 41%] 2023-03-31T05:54:16.8344298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8344774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8345247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8345713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8346184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8346685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8347183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8347651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8348122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8348594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8349061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8349521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8349986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8350517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8350989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8351457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8351932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8352405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8352859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8353324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8353803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8354310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8354812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8355277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8355756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8356235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8356713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8357182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8357656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8358193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8358671Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8359137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8359607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8360085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8360562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8361028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8361528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8362040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8362510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8362966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8363689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8364169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8364639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8365107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8365683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8366159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8366635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8367106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8367591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8368065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8368535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8369004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8369523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8370029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8370500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8370968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8371450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8371920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8372394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8372860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8373389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8373864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8374334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8374801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8375285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8375759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8376228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8376692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8377206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8377706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8378180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8378651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8379135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8379604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8380078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8380584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8381087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8381560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8382030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8382512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8383060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8383517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8383986Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8384495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8384993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8385466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8385940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8386413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8386889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8387362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8387830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8388375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8388856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8389330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8389808Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8390290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8390772Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8391246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8391716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8392217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8392720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8393191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8393649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8394130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8394604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8395071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8395536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8396066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8396543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8397008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8397478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8397948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8398418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8398889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8399385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8399878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8400352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8400820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8401300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8401770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8402246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8402717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8403344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8403863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8404343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8404811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8405290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8405758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8406234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8406700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8407216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8407716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8408187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8408656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8409131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8409598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8410073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8410540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8411077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8411545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8412017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8412485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8412959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8413420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8413891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8414355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8414858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8415368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8415841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8416308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8416787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8417257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8417732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8418208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8418742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8419209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8419677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8420144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8420629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8421099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8421565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8422058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8422559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8423098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8423566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8424040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8424515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8424983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8425460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8425971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8426472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8426941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8427415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8427890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8428366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8428833Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8429305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8429806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8430301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8430765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8431233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8431718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8432188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8432655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8433111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8433643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8434110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8434574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8435044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8435520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8435994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8436458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8436927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8437424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8437922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8438390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8438861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8439333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8439810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8440279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8440753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8441285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 42%] 2023-03-31T05:54:16.8441753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8442222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8442690Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8443346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8443841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8444305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8444831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8445336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8445804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8446266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8446741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8447218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8447693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8448162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8448687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8449193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8449677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8450146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8450629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8451101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8451572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8452046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8452553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8453047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8453506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8453971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8454454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8454920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8455386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8455856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8456396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8456867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8457334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8457806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8458287Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8458756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8459230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8459700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8460212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8460707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8461177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8461646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8462133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8462603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8463159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8463623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8464157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8464626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8465103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8465579Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8466058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8466529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8467002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8467538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8468051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8468517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8468987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8469462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8469943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8470422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8470891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8471423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8471900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8472370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8472847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8473311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8473788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8474258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8474731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8475237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8475730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8476200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8476674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8477153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8477631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8478095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8478569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8479095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8479572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8480041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8480513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8480987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8481462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8481933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8482435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8482939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8483524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8483994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8484473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8484949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8485429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8485903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8486423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8486930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8487401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8487870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8488347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8488823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8489299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8489771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8490280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8490784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8491253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8491720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8492197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8492665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8493137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8493591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8494136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8494608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8495088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8495561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8496040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8496513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8496988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8497459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8497967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8498463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8498936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8499402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8499882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8500349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8500820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8501286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8501823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8502294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8502769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8503315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8503777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8504252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8504722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8505225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8505725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8506195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8506669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8507147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8507622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8508087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8508558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8509068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8509563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8510033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8510507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8510987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8511459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8511929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8512397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8512897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8513380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8513844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8514315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8514788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8515260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8515729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8516201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8516728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8517205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8517670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8518145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8518623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8519093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8519558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8520035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8520565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8521076Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8521538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8522009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8522478Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8522944Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8523616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8524096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8524699Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8525181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8525648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8526124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8526599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8527069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8527535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8528042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8528555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8529030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8529496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8529975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8530446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8530918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8531387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8531892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8532394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8532872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8533328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8533805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8534280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8534742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8535209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8535710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8536201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8536669Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8537133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8537608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8538078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8538544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8539010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8539553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 43%] 2023-03-31T05:54:16.8540030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8540497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8540966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8541443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8541914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8542381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8542921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8543421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8543916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8544389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8544860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8545335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8545802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8546266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8546728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8547261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8547732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8548201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8548666Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8549141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8549612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8550079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8550541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8551044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8551540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8552009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8552477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8552941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8553413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8553883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8554349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8554881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8555351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8555825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8556291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8556764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8557232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8557695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8558191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8558687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8559152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8559620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8560091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8560564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8561030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8561492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8561985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8562474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8562939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8563607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8564085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8564564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8565037Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8565510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8566030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8566539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8567006Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8567479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8567954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8568427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8568895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8569364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8569910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8570380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8570850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8571318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8571787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8572257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8572714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8573179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8573674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8574175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8574645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8575117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8575591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8576062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8576531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8576997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8577526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8578000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8578471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8578939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8579411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8579884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8580346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8580806Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8581296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8581789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8582244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8582710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8583266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8583734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8584204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8584674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8585244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8585726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8586199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8586670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8587147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8587623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8588098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8588576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8589078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8589580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8590050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8590526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8590998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8591468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8591923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8592423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8592921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8593393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8593860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8594338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8594816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8595286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8595754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8596250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8596741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8597214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8597686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8598164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8598637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8599112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8599576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8600113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8600581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8601056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8601527Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8601995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8602470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8602939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8603635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8604169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8604672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8605149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8605618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8606096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8606564Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8607042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8607508Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8608057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8608530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8609004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8609469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8609946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8610414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8610888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8611358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8611856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8612350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8612826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8613302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8613782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8614247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8614723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8615217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8615721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8616188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8616657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8617135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8617611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8618081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8618552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8619051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8619542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8620009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8620482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8620961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8621436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8621890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8622365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8622968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8623453Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8623925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8624398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8624874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8625348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8625815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8626291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8626791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8627291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8627757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8628227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8628700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8629175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8629639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8630113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8630641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8631113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8631568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8632047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8632519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8632996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8633463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8633934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8634438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8634940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8635411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8635882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8636357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8636834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 44%] 2023-03-31T05:54:16.8637307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8637781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8638307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8638775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8639239Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8639713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8640190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8640661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8641125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8641615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8642154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8642627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8643214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8643702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8644174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8644641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8645110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8645673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8646150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8646620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8647086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8647561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8648031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8648506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8648976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8649494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8650000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8650467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8650937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8651401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8651868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8652338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8652799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8653342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8653814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8654280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8654744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8655225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8655693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8656162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8656627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8657131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8657627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8658098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8658570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8659047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8659519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8659987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8660449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8660983Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8661438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8661905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8662371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8662908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8663382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8663855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8664355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8664856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8665319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8665790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8666260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8666736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8667209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8667679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8668185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8668687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8669159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8669628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8670101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8670578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8671033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8671504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8672003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8672496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8672967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8673432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8673901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8674373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8674840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8675304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8675828Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8676297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8676763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8677230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8677705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8678183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8678648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8679113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8679615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8680123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8680584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8681060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8681530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8682004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8682466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8682937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8683715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8684193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8684659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8685122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8685588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8686064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8686532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8687002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8687509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8688013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8688484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8688950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8689421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8689894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8690360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8690811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8691343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8691813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8692284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8692755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8693227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8693697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8694163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8694636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8695133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8695625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8696087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8696555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8697023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8697492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8697963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8698461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8698958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8699429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8699895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8700356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8700824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8701288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8701753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8702290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8702786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8703375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8703847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8704326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8704800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8705272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8705740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8706288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8706753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8707227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8707689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8708157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8708614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8709078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8709544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8710031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8710524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8710991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8711456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8711932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8712394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8712863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8713334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8713856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8714330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8714803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8715271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8715742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8716212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8716688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8717151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8717646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8718135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8718605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8719070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8719531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8719996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8720464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8720931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8721469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8721934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8722399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8722867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8723474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8723943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8724419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8724891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8725419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8725922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8726392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8726862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8727339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8727810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8728281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8728746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8729295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8729755Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8730229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8730693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8731173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8731637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8732104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8732601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8733108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8733573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8734043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 45%] 2023-03-31T05:54:16.8734519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8734997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8735464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8735935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8736436Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8736933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8737398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8737872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8738344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8738816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8739269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8739741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8740245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8740741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8741201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8741672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8742140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8742610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8743157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8743631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8744166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8744637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8745100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8745573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8746047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8746519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8746985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8747456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8747956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8748455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8748911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8749379Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8749859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8750331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8750800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8751270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8751803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8752278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8752741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8753213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8753684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8754155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8754622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8755090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8755588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8756081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8756545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8757019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8757492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8757957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8758423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8758881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8759450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8759920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8760386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8760864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8761339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8761809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8762277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8762778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8763495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8763975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8764445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8764918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8765387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8765851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8766315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8766880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8767358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8767823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8768291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8768758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8769225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8769694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8770166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8770681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8771189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8771661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8772130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8772611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8773084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8773556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8774025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8774551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8775024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8775495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8775963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8776439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8776905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8777374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8777831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8778317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8778811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8779281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8779747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8780223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8780698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8781165Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8781627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8782151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8782622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8783164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8783631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8784112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8784587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8785064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8785528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8786031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8786532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8787001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8787468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8787943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8788397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8788860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8789321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8789845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8790313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8790776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8791243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8791718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8792182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8792648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8793142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8793645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8794115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8794583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8795056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8795533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8796004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8796475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8796974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8797474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8797932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8798405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8798877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8799348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8799815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8800281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8800770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8801260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8801729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8802200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8802676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8803264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8803734Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8804199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8804764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8805240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8805711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8806182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8806658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8807136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8807591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8808067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8808578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8809084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8809555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8810026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8810496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8810967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8811435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8811902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8812429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8812900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8813369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8813836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8814305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8814770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8815240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8815706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8816202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8816698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8817153Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8817626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8818104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8818575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8819044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8819514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8820078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8820553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8821025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8821495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8821969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8822445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8822990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8823463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8823962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8824460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8824928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8825402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8825874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8826346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8826811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8827300Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8827796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8828271Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8828740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8829219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8829695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8830170Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8830640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8831145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 46%] 2023-03-31T05:54:16.8831637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8832107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8832571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8833048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8833519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8833993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8834456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8834976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8835448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8835915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8836383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8836846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8837310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8837783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8838248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8838746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8839255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8839730Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8840201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8840682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8841149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8841619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8842090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8842626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8843272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8843800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8844274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8844750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8845217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8845692Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8846159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8846691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8847176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8847641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8848103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8848583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8849049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8849525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8849992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8850546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8851019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8851489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8851966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8852443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8852907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8853376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8853879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8854373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8854842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8855316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8855789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8856268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8856724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8857191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8857698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8858196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8858660Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8859129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8859603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8860069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8860538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8861011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8861512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8862001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8862464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8863008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8863487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8863960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8864430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8864904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8865448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8865930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8866382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8866856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8867331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8867802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8868267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8868740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8869237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8869726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8870184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8870654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8871126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8871594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8872060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8872525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8873057Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8873528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8873995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8874465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8874939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8875408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8875862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8876329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8876837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8877372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8877840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8878305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8878777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8879247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8879711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8880176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8880709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8881178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8881640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8882106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8882582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8883181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8883648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8884122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8884642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8885144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8885609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8886068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8886544Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8887014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8887480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8888005Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8888514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8888987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8889449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8889925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8890398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8890872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8891342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8891842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8892338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8892802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8893259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8893733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8894199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8894668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8895134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8895653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8896123Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8896587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8897049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8897525Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8898000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8898465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8898928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8899433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8899924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8900393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8900858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8901335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8901804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8902268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8902737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8903352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8903821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8904286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8904744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8905201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8905673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8906138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8906600Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8907102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8907599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8908067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8908532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8909003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8909473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8909934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8910392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8910919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8911391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8911858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8912319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8912788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8913264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8913731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8914192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8914676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8915163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8915628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8916091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8916563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8917025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8917485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8917942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8918476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8918935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8919391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8919845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8920318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8920781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8921240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8921697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8922200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8922694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8923365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8923849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8924312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8924784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8925252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8925716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8926293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8926759Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8927225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8927681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8928158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8928623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 47%] 2023-03-31T05:54:16.8929086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8929546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8930052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8930547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8931011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8931472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8931945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8932408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8932877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8933336Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8933850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8934322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8934786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8935246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8935719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8936186Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8936655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8937177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8937674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8938136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8938596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8939056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8939531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8939992Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8940455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8940949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8941445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8941904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8942364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8942909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8943368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8950802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8951359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8951925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8952461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8952935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8953408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8953859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8954330Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8954792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8955257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8955760Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8956255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8956710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8957178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8957647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8958125Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8958590Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8959044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8959526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8960018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8960483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8960948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8961410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8961894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8962365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8962832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8963490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8964026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8964500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8964981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8965458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8965931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8966400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8966869Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8967386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8967891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8968358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8968824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8969290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8969758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8970230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8970696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8971249Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8971720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8972191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8972657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8973115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8973589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8974056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8974521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8975025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8975518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8975993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8976465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8976938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8977415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8977880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8978352Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8978889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8979358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8979824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8980286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8980752Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8981224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8981694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8982159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8982662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8983243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8983716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8984179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8984655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8985128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8985598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8986063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8986604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8987078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8987547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8988013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8988486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8988955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8989420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8989922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8990422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8990891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8991355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8991816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8992284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8992733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8993201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8993706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8994203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8994674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8995142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8995611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8996080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8996547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8997017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8997517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8998023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8998491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8998963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8999437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.8999909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9000382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9000851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9001420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9001890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9002345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9002813Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9003542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9004019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9004490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9004963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9005487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9005991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9006456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9006925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9007399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9007862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9008328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9008799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9009344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9009809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9010278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9010748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9011225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9011693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9012150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9012617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9013142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9013640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9014109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9014565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9015039Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9015497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9015966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9016434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9016959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9017423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9017888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9018354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9018823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9019281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9019749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9020213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9020709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9021197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9021655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9022121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9022591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9023136Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9023606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9024068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9024602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9025071Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9025539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9025998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9026473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9026934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9027399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9027860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9028356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9028842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9029306Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9029776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9030251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9030713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9031172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9031639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9032184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9032652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 48%] 2023-03-31T05:54:16.9033124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9033594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9034074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9034539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9035010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9035501Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9036001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9036468Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9036935Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9037396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9037866Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9038328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9038796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9039296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9039791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9040251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9040700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9041167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9041633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9042089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9042550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9043165Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9043674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9044134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9044595Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9045070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9045543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9046007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9046473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9047023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9047495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9047956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9048420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9048888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9049353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9049812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9050266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9050764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9051253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9051714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9052179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9052643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9053109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9053571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9054041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9054560Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9055028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9055488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9055954Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9056429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9056898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9057362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9057830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9058333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9058835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9059295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9059750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9060218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9060682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9061137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9061605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9062162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9062623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9063157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9063621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9064085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9064546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9065003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9065464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9065964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9066448Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9066905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9067371Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9067846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9068316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9068778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9069233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9069767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9070237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9070701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9071169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9071632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9072098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9072557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9073022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9073522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9074013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9074472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9074934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9075396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9075859Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9076316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9076779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9077296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9077764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9078223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9078673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9079149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9079615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9080069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9080534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9081046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9081545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9082013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9082481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9082947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9083658Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9084124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9084592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9085161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9085628Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9086088Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9086549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9087014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9087474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9087915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9088372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9088867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9089362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9089819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9090289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9090758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9091225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9091687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9092152Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9092673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9093134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9093588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9094050Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9094522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9094994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9095457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9095925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9096417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9096906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9097355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9097825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9098290Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9098753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9099211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9099693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9100227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9100705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9101176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9101659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9102138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9102749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9103307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9103834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9104335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9104811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9105281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9105756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9106232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9106708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9107174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9107701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9108172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9108644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9109111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9109591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9110062Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9110529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9111000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9111520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9112027Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9112502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9112972Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9113457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9113929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9114400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9114867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9115405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9115881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9116351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9116822Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9117304Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9117763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9118236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9118702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9119206Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9119743Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9120213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9120689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9121167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9121636Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9122106Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9122602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9123258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9123735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9124205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9124683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9125164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9125631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9126102Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9126615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9127128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9127602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9128067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9128545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9129020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9129484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 49%] 2023-03-31T05:54:16.9129958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9130500Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9130978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9131445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9131916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9132388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9132863Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9133332Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9133801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9134320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9134819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9135291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9135762Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9136240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9136716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9137187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9137647Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9138179Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9138661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9139130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9139604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9140080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9140550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9141017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9141513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9142014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9142484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9143035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9143518Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9143989Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9144464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9144928Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9145460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9145926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9146401Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9146871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9147348Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9147804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9148280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9148747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9149255Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9149757Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9150229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9150703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9151183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9151654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9152131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9152599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9153129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9153597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9154066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9154537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9155016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9155485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9155955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9156427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9156929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9157419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9157871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9158342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9158819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9159289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9159765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9160233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9160768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9161236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9161711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9162185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9162660Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9163422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9163900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9164430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9164947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9165414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9165883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9166357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9166831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9167298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9167754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9168303Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9168782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9169247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9169716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9170190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9170661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9171130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9171610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9172115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9172614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9173078Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9173557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9174031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9174505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9174969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9175441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9175966Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9176440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9176906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9177374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9177843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9178313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9178782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9179252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9179790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9180281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9180746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9181216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9181694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9182172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9182638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9183184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9183715Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9184191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9184657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9185134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9185610Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9186077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9186545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9187049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9187529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9187996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9188462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9188940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9189414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9189883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9190353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9190892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9191366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9191834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9192297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9192783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9193260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9193732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9194333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9194864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9195362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9195836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9196309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9196788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9197264Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9197726Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9198198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9198724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9199197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9199667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9200134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9200615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9201085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9201559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9202023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9202531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9203167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9203638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9204107Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9204588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9205061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9205533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9206001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9206570Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9207048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9207520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9207980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9208460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9208931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9209406Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9209916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9210423Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9210895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9211369Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9211844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9212320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9212793Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9213261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9213791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9214268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9214733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9215202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9215673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9216154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9216627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9217097Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9217602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9218099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9218565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9219041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9219517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9219999Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9220471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9220943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9221484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9221957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9222428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9222974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9223458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9223932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9224399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9224906Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9225411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9225889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9226361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9226837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9227312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 50%] 2023-03-31T05:54:16.9227784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9228243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9228775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9229246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9229722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9230195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9230675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9231148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9231626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9232099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9232602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9233099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9233569Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9234035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9234512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9234987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9235461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9235930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9236491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9236964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9237438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9237903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9238367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9238834Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9239307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9239777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9240280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9240774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9241250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9241724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9242202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9242677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9243365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9243847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9244424Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9244896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9245367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9245832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9246311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9246780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9247248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9247754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9248247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9248718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9249189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9249665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9250137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9250609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9251079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9251604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9252081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9252549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9253021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9253491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9253967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9254435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9254914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9255417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9255910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9256377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9256850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9257326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9257794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9258246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9258717Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9259246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9259725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9260192Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9260667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9261142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9261617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9262081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9262552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9263139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9263641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9264109Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9264586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9265066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9265537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9266010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9266482Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9267012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9267489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9267955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9268421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9268894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9269364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9269826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9270328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9270839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9271315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9271781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9272260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9272736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9273204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9273672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9274205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9274679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9275150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9275616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9276093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9276566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9277040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9277505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9277998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9278487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9278959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9279425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9279910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9280380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9280847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9281317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9281847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9282320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9282789Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9283380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9283855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9284326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9284794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9285261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9285779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9286280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9286753Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9287220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9287686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9288162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9288634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9289096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9289646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9290120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9290591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9291056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9291529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9291996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9292456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9292915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9293413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9293909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9294376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9294845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9295317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9295783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9296247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9296777Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9297301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9297764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9298232Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9298705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9299184Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9299654Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9300124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9300621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9301122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9301585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9302054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9302531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9303101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9303572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9304042Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9304577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9305048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9305517Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9305984Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9306462Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9306934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9307387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9307856Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9308356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9308862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9309331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9309800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9310276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9310747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9311215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9311689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9312212Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9312693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9313164Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9313639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9314115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9314588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9315053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9315520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9316009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9316504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9316958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9317428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9317902Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9318375Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9318841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9319310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9319832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9320308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9320774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9321245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9321719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9322187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9322655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9323309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9323879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9324399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9324874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 51%] 2023-03-31T05:54:16.9325345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9325818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9326286Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9326733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9327193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9327723Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9328195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9328661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9329121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9329588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9330058Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9330531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9331032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9331528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9332001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9332473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9332950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9333422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9333896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9334365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9334872Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9335374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9335854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9336322Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9336788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9337261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9337736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9338211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9338718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9339217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9339688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9340156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9340634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9341108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9341577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9342048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9342575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9343133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9343608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9344075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9344552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9345018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9345491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9345953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9346447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9346943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9347422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9347889Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9348364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9348838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9349310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9349780Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9350314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9350781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9351257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9351725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9352205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9352677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9353147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9353646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9354166Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9354631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9355100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9355571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9356049Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9356504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9356977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9357477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9357976Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9358442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9358917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9359398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9359870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9360335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9360807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9361307Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9361800Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9362265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9362739Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9363356Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9363835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9364305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9364775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9365325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9365803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9366266Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9366724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9367205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9367674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9368145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9368611Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9369117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9369623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9370089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9370565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9371041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9371515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9371981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9372451Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9372977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9373445Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9373914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9374387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9374862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9375331Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9375907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9376402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9376914Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9377390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9377860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9378342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9378817Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9379296Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9379763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9380268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9380763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9381237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9381705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9382180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9382655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9383210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9383681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9384188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9384681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9385149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9385617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9386099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9386555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9387026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9387488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9388017Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9388488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9388960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9389427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9389907Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9390376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9390848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9391314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9391824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9392321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9392790Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9393256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9393735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9394208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9394677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9395141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9395677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9396133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9396597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9397064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9397541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9398009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9398477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9398940Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9399438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9399930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9400398Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9400870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9401353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9401825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9402293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9402761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9403540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9404029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9404504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9404979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9405450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9405921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9406378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9406884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9407395Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9407864Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9408333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9408799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9409274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9409736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9410198Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9410733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9411209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9411672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9412140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9412616Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9413086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9413556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9414029Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9414568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9415061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9415531Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9415985Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9416457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9416929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9417397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9417865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9418384Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9418855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9419325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9419786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9420257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9420725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9421195Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9421659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9422158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9422652Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 52%] 2023-03-31T05:54:16.9423193Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9423657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9424131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9424608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9425082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9425536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9426061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9426535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9427007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9427473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9427945Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9428416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9428884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9429353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9429848Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9430351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9430819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9431282Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9431747Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9432215Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9432682Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9433148Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9433667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9434135Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9434607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9435067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9435537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9436004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9436470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9436969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9437465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9437938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9438408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9438877Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9439351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9439820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9440285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9440791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9441285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9441756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9442223Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9442693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9443285Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9443750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9444214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9444729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9445218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9445684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9446149Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9446621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9447087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9447557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9448022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9448567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9449036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9449511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9449980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9450455Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9450921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9451390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9451860Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9452361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9452849Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9453316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9453779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9454247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9454696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9455160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9455626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9456159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9456625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9457091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9457557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9458026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9458493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9458963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9459431Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9459930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9460421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9460888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9461357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9461831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9462298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9462768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9463310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9463847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9464305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9464776Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9465244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9465713Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9466177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9466651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9467120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9467622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9468111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9468580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9469046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9469513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9469975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9470449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9470920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9471492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9471961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9472435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9472911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9473385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9473840Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9474317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9474818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9475316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9475782Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9476257Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9476732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9477200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9477664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9478133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9478659Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9479132Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9479599Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9480069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9480546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9481019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9481487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9481961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9482466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9482967Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9483665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9484128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9484603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9485074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9485540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9486018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9486588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9487060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9487529Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9487998Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9488475Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9488947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9489418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9489887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9490394Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9490899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9491367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9491841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9492316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9492787Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9493252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9493727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9494237Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9494707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9495172Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9495653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9496121Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9496589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9497059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9497572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9498063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9498533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9499004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9499488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9499959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9500426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9500893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9501425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9501894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9502362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9502909Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9503388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9503844Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9504316Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9504781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9505293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9505791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9506267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9506740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9507220Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9507695Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9508167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9508638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9509178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9509649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9510118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9510584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9511063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9511533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9511997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9512469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9512970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9513461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9513915Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9514380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9514855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9515323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9515792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9516261Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9516810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9517277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9517745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9518218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9518691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9519162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9519632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 53%] 2023-03-31T05:54:16.9520130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9520629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9521095Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9521563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9522041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9522515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9522978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9523539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9524092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9524566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9525034Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9525499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9525969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9526439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9526905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9527377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9527881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9528392Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9528857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9529329Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9529805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9530276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9530746Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9531221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9531794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9532272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9532735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9533207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9533663Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9534134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9534598Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9535065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9535568Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9536068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9536533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9537002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9537472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9537943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9538414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9538882Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9539403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9539874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9540347Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9540820Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9541293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9541768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9542236Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9542737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9543302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9543771Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9544234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9544705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9545180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9545649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9546112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9546614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9547103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9547571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9548036Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9548514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9548979Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9549446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9549911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9550416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9550908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9551374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9551837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9552312Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9552781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9553234Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9553703Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9554225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9554688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9555157Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9555621Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9556096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9556566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9557030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9557493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9557987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9558470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9558934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9559397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9559868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9560341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9560812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9561280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9561809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9562279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9562737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9563410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9563901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9564373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9564847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9565323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9565855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9566359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9566835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9567301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9567775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9568247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9568721Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9569189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9569736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9570200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9570668Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9571144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9571620Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9572091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9572547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9573046Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9573546Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9574013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9574480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9574956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9575430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9575895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9576368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9576904Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9577377Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9577845Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9578324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9578802Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9579272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9579736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9580204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9580708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9581201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9581665Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9582141Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9582601Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9583140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9583609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9584080Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9584608Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9585082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9585550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9586018Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9586492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9586965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9587581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9588155Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9588676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9589221Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9589686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9590162Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9590639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9591112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9591576Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9592051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9592580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9593038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9593502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9593981Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9594458Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9594930Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9595400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9595905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9596405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9596873Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9597339Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9597823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9598295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9598768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9599235Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9599761Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9600233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9600708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9601176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9601656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9602124Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9602584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9603151Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9603679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9604188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9604651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9605120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9605596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9606069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9606532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9606997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9607547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9608021Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9608490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9608961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9609443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9609911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9610387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9610853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9611353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9611852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9612324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9612778Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9613258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9613728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9614199Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9614670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9615207Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9615679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9616147Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9616614Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9617092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 54%] 2023-03-31T05:54:16.9617558Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9618022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9618526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9619126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9619596Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9620064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9620537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9621010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9621479Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9621949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9622442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9623026Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9623502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9623975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9624447Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9624920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9625390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9625858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9626366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9626876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9627341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9627803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9628279Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9628750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9629217Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9629685Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9630211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9630684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9631154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9631619Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9632091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9632553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9633022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9633495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9633993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9634487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9634955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9635426Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9635905Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9636373Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9636839Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9637309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9637841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9638313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9638773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9639238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9639711Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9640181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9640650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9641114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9641615Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9642111Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9642575Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9643175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9643709Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9644177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9644645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9645114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9645681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9646150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9646618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9647090Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9647563Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9648038Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9648504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9649053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9649561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9650032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9650503Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9650970Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9651432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9651883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9652345Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9652876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9653346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9653815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9654277Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9654751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9655219Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9655689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9656161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9656662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9657145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9657613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9658082Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9658556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9659023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9659498Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9659969Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9660492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9660961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9661428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9661878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9662351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9662892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9663357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9663825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9664335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9664821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9665289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9665754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9666224Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9666686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9667150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9667623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9668154Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9668626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9669091Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9669559Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9670031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9670502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9670974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9671429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9671929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9672420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9672883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9673344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9673812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9674270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9674736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9675202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9675720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9676181Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9676643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9677112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9677584Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9678054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9678522Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9678993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9679489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9679978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9680442Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9680895Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9681364Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9681830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9682299Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9682766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9683407Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9683874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9684335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9684796Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9685259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9685716Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9686178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9686686Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9687188Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9687644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9688110Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9688581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9689052Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9689511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9689971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9690456Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9690946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9691405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9691865Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9692343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9692812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9693276Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9693749Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9694243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9694750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9695216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9695677Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9696145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9696606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9697064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9697532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9698060Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9698528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9698988Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9699450Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9699903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9700467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9700943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9701414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9701923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9702422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9702955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9703427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9703903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9704376Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9704846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9705321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9705897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9706378Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9706846Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9707318Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9707791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9708262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9708731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9709200Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9709689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9710182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9710649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9711119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9711592Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9712065Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9712530Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9713003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9713542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9714016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 55%] 2023-03-31T05:54:16.9714480Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9714946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9715417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9715890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9716357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9716830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9717333Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9717829Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9718294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9718764Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9719228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9719700Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9720171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9720643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9721173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9721646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9722113Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9722588Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9723205Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9723714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9724178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9724708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9725213Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9725689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9726159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9726642Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9727119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9727591Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9728061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9728609Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9729085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9729543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9730012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9730487Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9730955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9731421Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9731887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9732383Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9732876Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9733343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9733804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9734281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9734751Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9735225Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9735688Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9736222Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9736691Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9737158Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9737625Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9738099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9738571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9739028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9739493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9740002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9740494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9740965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9741428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9741897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9742368Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9742898Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9743362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9743892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9744365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9744830Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9745298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9745779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9746246Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9746712Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9747175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9747681Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9748183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9748634Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9749100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9749580Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9750047Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9750519Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9750991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9751521Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9757965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9758511Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9759007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9759486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9759961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9760432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9760963Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9761463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9761938Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9762408Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9762881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9763502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9763968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9764435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9765014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9765493Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9765960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9766433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9766911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9767389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9767855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9768328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9768842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9769349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9769821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9770294Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9770768Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9771240Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9771706Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9772178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9772731Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9773203Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9773673Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9774140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9774604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9775074Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9775539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9776001Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9776504Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9777004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9777473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9777939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9778418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9778891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9779359Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9779832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9780365Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9780838Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9781302Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9781773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9782355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9782929Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9783402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9783862Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9784358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9784852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9785320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9785786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9786262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9786732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9787201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9787674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9788209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9788684Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9789150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9789617Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9790085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9790550Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9791020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9791510Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9792016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9792492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9792962Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9793437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9793892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9794358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9794823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9795315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9795801Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9796272Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9796737Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9797209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9797675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9798144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9798618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9799116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9799606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9800077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9800547Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9801023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9801488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9801959Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9802430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9802957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9803629Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9804103Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9804567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9805048Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9805512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9805973Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9806439Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9806960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9807460Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9807927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9808397Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9808867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9809334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9809803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9810270Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9810816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9811280Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9811750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9812226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9812705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9813161Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9813632Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9814101Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9814603Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9815092Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9815566Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9816033Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9816507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9816975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9817440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9817903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 56%] 2023-03-31T05:54:16.9818422Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9818883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9819349Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9819814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9820291Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9820754Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9821218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9821683Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9822178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9822648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9823190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9823661Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9824139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9824605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9825068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9825533Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9826067Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9826537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9827004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9827472Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9827947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9828412Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9828880Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9829341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9829831Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9830344Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9830811Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9831284Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9831758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9832208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9832675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9833144Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9833674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9834139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9834607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9835068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9835537Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9835996Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9836465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9836958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9837454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9837919Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9838385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9838857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9839328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9839788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9840250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9840742Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9841233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9841678Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9842145Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9842607Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9843209Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9843670Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9844130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9844640Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9845137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9845594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9846070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9846545Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9847025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9847491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9847965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9848489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9848995Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9849469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9849943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9850419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9850890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9851358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9851815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9852313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9852804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9853269Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9853735Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9854211Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9854680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9855142Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9855612Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9856131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9856606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9857069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9857538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9858015Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9858488Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9858955Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9859428Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9859926Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9860430Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9860900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9861361Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9861836Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9862309Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9862775Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9863338Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9863867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9864337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9864803Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9865274Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9865744Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9866216Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9866675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9867138Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9867639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9868133Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9868597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9869077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9869553Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9870020Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9870486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9870974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9871476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9871950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9872420Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9872896Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9873370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9873841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9874305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9874804Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9875293Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9875756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9876218Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9876689Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9877159Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9877623Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9878085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9878613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9879081Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9879549Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9880013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9880494Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9880953Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9881419Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9881885Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9882389Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9882887Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9883586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9884061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9884540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9885012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9885485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9885948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9886520Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9886994Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9887464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9887927Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9888402Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9888870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9889337Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9889805Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9890351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9890843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9891313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9891779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9892258Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9892733Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9893202Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9893667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9894197Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9894667Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9895143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9895606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9896086Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9896556Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9897025Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9897490Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9897991Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9898486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9898951Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9899418Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9899894Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9900350Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9900814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9901281Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9901810Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9902278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9902748Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9903321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9903794Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9904262Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9904727Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9905226Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9905724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9906191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9906655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9907130Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9907606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9908069Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9908541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9909061Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9909534Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9910003Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9910454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9910918Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9911385Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9911850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9912315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9912814Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9913320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9913784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9914248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9914719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 57%] 2023-03-31T05:54:16.9915196Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9915664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9916128Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9916655Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9917137Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9917605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9918072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9918543Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9919016Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9919485Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9919943Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9920438Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9920933Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9921400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9921861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9922326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9922791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9923391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9923861Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9924409Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9924883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9925354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9925819Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9926289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9926765Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9927233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9927701Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9928201Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9928705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9929173Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9929631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9930108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9930581Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9931043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9931512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9932044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9932516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9932978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9933446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9933910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9934381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9934847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9935341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9935837Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9936301Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9936770Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9937238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9937705Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9938174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9938646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9939139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9939641Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9940117Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9940587Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9941063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9941536Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9942007Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9942476Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9943077Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9943697Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9944174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9944638Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9945112Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9945578Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9946045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9946516Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9947072Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9947541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9948010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9948477Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9948952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9949403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9949874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9950346Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9950850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9951351Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9951826Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9952289Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9952763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9953230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9953702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9954167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9954694Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9955163Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9955635Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9956100Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9956572Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9957040Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9957509Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9957980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9958481Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9958952Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9959425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9959892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9960367Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9960835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9961313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9961779Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9962311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9962784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9963464Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9963948Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9964427Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9964893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9965362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9965892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9966413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9966878Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9967354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9967823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9968295Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9968758Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9969214Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9969756Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9970231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9970698Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9971169Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9971643Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9972116Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9972582Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9973053Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9973548Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9974044Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9974507Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9974975Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9975449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9975917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9976382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9976854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9977380Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9977857Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9978319Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9978773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9979247Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9979714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9980174Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9980637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9981143Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9981633Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9982104Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9982577Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9983131Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9983602Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9984068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9984540Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9985070Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9985541Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9986012Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9986486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9986961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9987443Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9987910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9988366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9988892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9989390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9989855Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9990323Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9990795Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9991265Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9991725Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9992190Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9992718Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9993187Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9993649Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9994120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9994597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9995064Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9995528Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9996023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9996514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9996980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9997446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9997908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9998386Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9998852Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9999317Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:16.9999818Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0000314Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0000786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0001253Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0001722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0002191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0002656Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0003260Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0003774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0004278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0004745Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0005210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0005687Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0006160Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0006627Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0007093Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0007675Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0008140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0008604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0009075Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0009551Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0010019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0010489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0010956Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0011465Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0011958Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 58%] 2023-03-31T05:54:17.0012429Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0012893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0013372Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0013842Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0014313Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0014773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0015297Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0015767Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0016233Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0016704Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0017175Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0017630Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0018094Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0018561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0019063Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0019567Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0020032Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0020499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0020978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0021449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0021916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0022381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0022993Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0023469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0023934Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0024403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0024881Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0025343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0025809Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0026273Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0026773Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0027251Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0027720Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0028177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0028653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0029122Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0029594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0030056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0030585Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0031051Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0031515Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0031980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0032454Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0032921Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0033390Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0033874Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0034370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0034843Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0035315Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0035783Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0036252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0036702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0037167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0037657Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0038150Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0038613Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0039073Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0039535Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0040002Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0040461Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0040916Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0041400Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0041892Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0042353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0042815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0043502Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0043982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0044449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0044917Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0045437Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0045946Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0046405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0046871Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0047341Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0047807Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0048268Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0048729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0049229Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0049728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0050191Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0050653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0051120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0051583Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0052041Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0052499Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0053013Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0053486Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0053942Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0054403Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0054870Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0055334Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0055791Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0056254Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0056741Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0057227Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0057696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0058171Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0058645Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0059115Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0059573Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0060035Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0060542Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0061011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0061469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0061932Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0062399Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0062950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0063405Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0063867Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0064363Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0064868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0065326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0065788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0066252Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0066722Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0067183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0067646Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0068167Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0068637Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0069096Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0069562Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0070028Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0070489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0070947Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0071415Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0071908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0072416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0072879Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0073343Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0073812Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0074267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0074729Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0075189Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0075702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0076183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0076648Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0077118Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0077594Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0078066Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0078538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0079010Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0079506Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0080000Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0080471Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0080941Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0081416Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0081888Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0082353Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0082816Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0083491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0083960Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0084425Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0084891Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0085362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0085832Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0086298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0086763Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0087267Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0087769Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0088238Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0088708Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0089178Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0089650Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0090120Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0090589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0091119Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0091593Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0092059Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0092532Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0093004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0093470Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0093920Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0094411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0094950Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0095606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0096079Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0096554Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0097031Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0097505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0097971Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0098484Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0098987Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0099466Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0099936Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0100413Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0100886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0101362Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0101827Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0102324Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0102893Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0103366Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0103823Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0104305Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0104774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0105243Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0105710Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0106230Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0106693Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0107156Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0107626Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0108099Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0108565Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 59%] 2023-03-31T05:54:17.0109030Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0109497Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0110014Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0110505Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0110974Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0111446Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0111924Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0112388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0112858Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0113311Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0113841Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0114310Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0114781Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0115245Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0115719Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0116194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0116662Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0117126Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0117618Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0118105Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0118574Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0119043Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0119514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0119977Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0120440Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0120908Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0121433Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0121903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0122374Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0122825Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0123491Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0123961Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0124434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0124903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0125457Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0125964Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0126435Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0126903Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0127382Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0127851Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0128321Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0128784Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0129328Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0129792Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0130256Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0130732Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0131204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0131664Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0132129Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0132604Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0133098Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0133557Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0134023Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0134495Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0134968Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0135432Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0135900Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0136396Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0136890Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0137355Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0137824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0138298Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0138766Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0139231Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0139696Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0140185Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0140672Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0141134Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0141597Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0142056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0142524Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0143056Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0143523Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0144054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0144526Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0144990Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0145463Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0145939Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0146404Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0146868Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0147335Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0147835Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0148326Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0148788Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0149259Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0149736Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0150210Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0150674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0151140Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0151651Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0152114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0152571Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0153034Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0153513Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0153980Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0154441Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0154911Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0155411Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0155899Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0156358Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0156824Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0157288Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0157750Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0158208Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0158676Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0159204Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0159674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0160139Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0160606Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0161068Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0161538Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0161997Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0162459Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0162949Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0163589Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0164045Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0164512Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0164982Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0165449Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0165910Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0166370Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0166923Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0167388Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0167850Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0168320Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0168786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0169250Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0169707Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0170176Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0170680Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0171182Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0171644Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0172108Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0172586Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0173054Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0173514Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0173978Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0174496Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0174957Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0175414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0175886Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0176354Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0176821Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0177278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0177738Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0178228Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0178724Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0179180Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0179639Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0180089Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0180552Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0181009Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0181474Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0182019Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0182489Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0183022Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0183492Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0183965Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0184434Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0184897Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0185357Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0185853Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0186342Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0186799Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0187263Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0187728Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0188194Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0188653Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0189114Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0189622Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0190084Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0190539Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0191004Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0191467Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0191931Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0192387Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0192884Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0193381Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0193847Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0194308Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0194774Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0195248Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0195714Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0196177Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0196674Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0197168Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0197631Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0198087Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0198561Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0199011Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0199469Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0199925Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0200414Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0200901Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0201360Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0201815Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0202278Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0202740Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0203417Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0203883Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0204410Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0204922Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0205391Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 60%] 2023-03-31T05:54:17.0205854Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0206325Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0206786Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0207244Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0207702Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0208183Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0208679Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0209146Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0209605Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0210085Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0210555Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0211024Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0211483Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0212008Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0212473Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0212937Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0213393Z test_transformers.py::TestSDPA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0213708Z test_transformers.py::TestSDPA::test_flash_autocast_fp32_bfloat16 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0214020Z test_transformers.py::TestSDPA::test_flash_autocast_fp32_float16 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0214242Z test_transformers.py::TestSDPA::test_flash_backward_sm86_headdim128 SKIPPED (Does not support fused SDPA or not SM86 hardware) [ 61%] 2023-03-31T05:54:17.0214541Z test_transformers.py::TestSDPA::test_flash_fail_fp32 SKIPPED (Does not support fused SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0214760Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_error_cases SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0215169Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0215610Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0216041Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0216446Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0216857Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0217263Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0217662Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0218119Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0218528Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0218923Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0219331Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0219731Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0220127Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0220552Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0220975Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0221366Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0221774Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0222175Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0222572Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0223079Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0223502Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0223897Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0224296Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0224695Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0225094Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0225516Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0225953Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0226346Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0226735Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0227130Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0227521Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0227910Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0228361Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0228764Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0229160Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0229564Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0229959Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0230351Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0230776Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0231191Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0231585Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0231991Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0232387Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0232779Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0233249Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0233645Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0234034Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0234431Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0234828Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0235223Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0235644Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0236063Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0236452Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0236843Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0237237Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0237622Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0238012Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0238462Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0238851Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0239236Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0239628Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0240014Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0240397Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0240799Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_EFFICIENT_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0241217Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0241610Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0242003Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0242393Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0242777Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0243373Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0243763Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0244145Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0244549Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0244953Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0245342Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0245770Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0246195Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0246591Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0246985Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0247381Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0247775Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0248169Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0248623Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0249012Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0249404Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0249805Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0250195Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0250581Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0251004Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0251419Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0251806Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0252202Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0252594Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0252985Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0253436Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0253825Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0254220Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0254619Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0255013Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0255399Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0255824Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0256244Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0256627Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0257021Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0257416Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0257803Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0258195Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0258636Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0259030Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0259419Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0259814Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0260201Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0260600Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0261021Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0261443Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0261834Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0262230Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0262618Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0263085Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0263540Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0263930Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0264319Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0264714Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0265106Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0265492Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0265894Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0266310Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0266699Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_kernel_SDPBackend_FLASH_ATTENTION_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0266930Z test_transformers.py::TestSDPA::test_fused_kernels_nested_broadcasting_query_dense SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0267196Z test_transformers.py::TestSDPA::test_fused_kernels_seq_len_0_inputs_fused_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0267454Z test_transformers.py::TestSDPA::test_fused_kernels_seq_len_0_inputs_fused_kernel_SDPBackend_FLASH_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0267710Z test_transformers.py::TestSDPA::test_fused_kernels_seq_len_1_inputs_fused_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0267962Z test_transformers.py::TestSDPA::test_fused_kernels_seq_len_1_inputs_fused_kernel_SDPBackend_FLASH_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 61%] 2023-03-31T05:54:17.0268126Z test_transformers.py::TestSDPA::test_fused_sdp_choice_type_dense PASSED [ 61%] 2023-03-31T05:54:17.0268286Z test_transformers.py::TestSDPA::test_fused_sdp_choice_type_nested PASSED [ 61%] 2023-03-31T05:54:17.0268471Z test_transformers.py::TestSDPA::test_incompatible_mask PASSED [ 61%] 2023-03-31T05:54:17.0268781Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_attn_mask_present_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0269056Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_broadcast_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0269307Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_dim_3_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0269576Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_head_dim_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0269839Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_head_dim_kernel_SDPBackend_FLASH_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0270118Z test_transformers.py::TestSDPA::test_invalid_fused_inputs_invalid_dtype_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Does not support fused scaled dot product attention) [ 61%] 2023-03-31T05:54:17.0270351Z test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0270576Z test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_FLASH_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0270790Z test_transformers.py::TestSDPA::test_invalid_inputs_1_dimensional_inputs_kernel_SDPBackend_MATH_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0271021Z test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0271247Z test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_FLASH_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0271487Z test_transformers.py::TestSDPA::test_invalid_inputs_different_datatypes_kernel_SDPBackend_MATH_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0271770Z test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_EFFICIENT_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0271978Z test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_FLASH_ATTENTION_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0272185Z test_transformers.py::TestSDPA::test_invalid_inputs_different_devices_kernel_SDPBackend_MATH_device_cpu PASSED [ 61%] 2023-03-31T05:54:17.0272724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0273222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0273713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0274198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0274691Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0275235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0275735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0276225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0276715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0277203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0277698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0278180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0278671Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0279190Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0279707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0280192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0280685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0281176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0281667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0282135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0282688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0283414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0283915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0284401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0284899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0285382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0285877Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0286361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0286914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 61%] 2023-03-31T05:54:17.0287433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0287929Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0288420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0288911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0289397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0289880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0290438Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0290930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0291414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0291901Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0292390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0292882Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0293353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0293848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0294361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0294868Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0295354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0295833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0296323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0296817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0297307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0297827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0298350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0298840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0299325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0299820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0300308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0300798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0301279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0301798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0302304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0302798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0303361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0303855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0304339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0304814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0305304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0305862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0306346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0306830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0307317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0307811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0308291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0308783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0309303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0309818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0310302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0310788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0311280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0311770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0312252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0312740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0313279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0313769Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0314253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0314742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0315231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0315723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0316192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0316713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0317223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0317713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0318196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0318698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0319185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0319673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0320152Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0320702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0321188Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0321670Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0322152Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0322641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0323258Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0323742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0324269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0324794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0325277Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0325761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0326248Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0326738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0327211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0327702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0328261Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0328747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0329230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0329715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0330208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0330692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0331174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0331656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0332186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0332707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0333196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0333688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0334173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0334664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0335150Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0335706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0336201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0336694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0337180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0337678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0338163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0338642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0339126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0339651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0340155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0340642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0341127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0341621Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0342106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0342594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0343203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0343723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0344209Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0344697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0345190Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0345685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0346168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0346661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0347176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0347696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0348181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0348666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0349151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0349645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0350113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0350606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0351144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0351636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0352115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0352616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0353103Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0353592Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0354071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0354592Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0355100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0355587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0356070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0356566Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0357052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0357541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0358029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0358578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0359066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0359554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0360047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0360539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0361008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0361491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0362008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0362536Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0363162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0363747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0364242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0364733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0365216Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0365709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0366291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0366782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0367268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0367751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0368239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0368730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0369215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0369738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0370260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0370747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0371229Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0371718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0372201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0372673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0373153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0373695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0374180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0374671Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0375156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0375648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0376137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0376623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0377136Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0377650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0378136Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0378623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0379113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0379603Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0380085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0380568Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0381109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0381600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0382083Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0382568Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0383135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0383614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0384095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0384579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0385098Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0385614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0386093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0386585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0387080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0387569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 62%] 2023-03-31T05:54:17.0388048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0388637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0389122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0389604Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0390083Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0390578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0391063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0391543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0392019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0392537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0393045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0393534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0394018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0394513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0394989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0395475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0395960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0396498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0396983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0397470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0397955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0398449Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0398929Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0399418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0399933Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0400441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0400923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0401412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0401898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0402389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0402871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0403476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0404051Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0404547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0405032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0405518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0405992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0406475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0406956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0407475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0407991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0408477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0408957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0409445Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0409930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0410415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0410896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0411437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0411920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0412401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0412883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0413371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0413856Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0414333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0414818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0415331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0415842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0416325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0416792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0417289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0417777Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0418264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0418784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0419302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0419782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0420261Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0420745Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0421223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0421694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0422184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0422697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0423293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0423778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0424266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0424761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0425258Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0425745Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0426232Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0426783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0427278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0427748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0428245Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0428738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0429223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0429709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0430226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0430732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0431217Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0431696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0432192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0432677Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0433163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0433642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0434186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0434673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0435160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0435646Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0436136Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0436620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0437104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0437618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0438131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0438612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0439087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0439577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0440066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0440548Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0441037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0441584Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0442076Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0442557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0443283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0443823Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0444313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0444795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0445283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0445825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0446350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0446835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0447330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0447815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0448302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0448784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0449376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0449859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0450332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0450811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0451297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0451778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0452271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0452758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0453280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0453793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0454285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0454775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0455271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0455753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0456241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0456755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0457266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0457751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0458242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0458736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0459227Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0459708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0460195Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0460717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0461212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0461697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0462185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0462676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0463238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0463718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0464206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0464762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0465256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0465741Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0466233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0466723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0467205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0467687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0468207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0468718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0469205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0469684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0470178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0470665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0471154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0471636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0472181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0472653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0473138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0473623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0474111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0474594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0475078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0475589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0476104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0476584Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0477069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0477560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0478045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0478520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0479007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0479543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0480032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0480509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0480991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0481483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0481978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0482464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0482956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0483601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0484123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0484602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0485097Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0485590Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0486073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0486550Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0487071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0487594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0488084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 63%] 2023-03-31T05:54:17.0488567Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0489055Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0489538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0490021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0490499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0491015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0491522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0492006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0492487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0492983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0493469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0493953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0494423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0494969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0496619Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0497106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0497588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0498081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0498595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0499082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0499600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0500109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0500592Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0501081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0501566Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0502052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0502535Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0503131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0503655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0504145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0504674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0505157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0505645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0506134Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0506616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0507109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0507650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0508148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0508632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0509122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0509612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0510100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0510577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0511087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0511569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0512082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0512558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0513057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0513549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0514042Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0514534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0515052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0515563Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0516063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0516547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0517041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0517531Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0518017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0518501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0519012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0519511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0519998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0520489Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0520982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0521468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0521955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0522473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0522992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0523607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0524093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0524576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0525070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0525558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0526045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0526582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0527112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0527596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0528084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0528575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0529063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0529543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0530071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0530576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0531067Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0531549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0532040Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0532528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0533016Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0533495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0534011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0534501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0535018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0535498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0535987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0536471Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0536965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0537481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0537994Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0538481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0538969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0539457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0539944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0540431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0540915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0541422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0541894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0542404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0542974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0543467Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0543954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0544441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0544929Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0545444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0545960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0546444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0546934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0547424Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0547913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0554400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0555052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0555569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0556104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0556597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0557093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0557584Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0558072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0558558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0559065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0559575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0560063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0560542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0561034Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0561522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0562009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0562497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0563009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0563709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0564194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0564681Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0565171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0565656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0566142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0566661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0567190Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0567677Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0568160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0568647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0569132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0569618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0570091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0570613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0571101Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0571647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0572134Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0572616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0573105Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0573590Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0574119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0574626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0575114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0575601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0576095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0576576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0577065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0577549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0578062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0578548Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0579077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0579567Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0580060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0580546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0581036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0581547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0582065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0582552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0583135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0583627Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0584117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0584601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0585097Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0585621Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0586114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0586628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0587112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0587599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0588087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0588576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0589062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0589578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0590089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0590571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0591062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0591547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0592038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0592508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0593027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0593513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0594024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0594505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0594997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0595480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0595963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 64%] 2023-03-31T05:54:17.0596446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0596964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0597474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0597964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0598452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0598939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0599421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0599912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0600395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0600910Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0601419Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0601902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0602387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0602878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0603507Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0603979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0604515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0605034Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0605517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0606005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0606495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0606982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0607465Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0607954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0608481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0608979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0609503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0609990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0610481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0610970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0611459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0611974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0612483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0612976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0613456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0613946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0614433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0614910Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0615396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0615919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0616403Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0616918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0617400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0617886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0618371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0618858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0619366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0619876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0620353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0620845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0621335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0621825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0622310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0622794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0623384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0623871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0624381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0624867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0625351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0625824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0626309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0626804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0627313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0627827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0628309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0628797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0629282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0629767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0630245Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0630731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0631257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0631784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0632266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0632758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0633240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0633726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0634207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0634726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0635230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0635721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0636207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0636695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0637169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0637653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0638130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0638648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0639160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0639647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0640137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0640631Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0641115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0641600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0642115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0642625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0643234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0643723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0644211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0644702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0645185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0645668Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0646197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0646689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0647212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0647695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0648170Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0648658Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0649143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0649619Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0650127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0650652Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0651137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0651622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0652110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0652596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0653080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0653600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0654083Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0654593Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0655074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0655559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0656037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0656523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0657008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0657522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0658028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0658501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0658978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0659469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0659951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0660441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0660920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0661438Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0661944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0662432Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0662988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0663484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0663970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0664447Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0664954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0665458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0665935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0666413Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0666892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0667371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0667846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0668334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0668851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0669338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0669850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0670339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0670827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0671312Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0671800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0672288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0672797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0673310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0673791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0674283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0674772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0675260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0675744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0676260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0676745Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0677256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0677738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0678226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0678705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0679185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0679660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0680161Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0680667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0681155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0681634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0682123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0682607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0683215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0683700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0684238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0684751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0685235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0685720Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0686210Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0686694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0687178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0687701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0688248Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0688731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0689218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0689698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0690187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0690667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0691147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0691647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0692140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0692651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0693142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0693629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0694115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0694596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0695106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0695611Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0696099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0696582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 65%] 2023-03-31T05:54:17.0697075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0697556Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0698047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0698532Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0699048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0699541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0700060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0700545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0701036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0701520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0702009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0702476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0703068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0703581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0704066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0704552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0705038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0705526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0706008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0706517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0707008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0707513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0708002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0708486Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0708971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0709454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0709940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0710453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0710965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0711447Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0711930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0712416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0712899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0713369Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0713848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0714366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0714878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0715357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0715842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0716328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0716818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0717299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0717814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0718316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0718800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0719278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0719766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0720249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0720727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0721203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0721717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0722199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0722704Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0723307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0723795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0724266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0724751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0725277Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0725793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0726274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0726749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0727226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0727720Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0728205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0728694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0729215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0729705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0730227Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0730718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0731206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0731693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0732179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0732659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0733165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0733675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0734162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0734646Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0735114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0735609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0736095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0736583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0737096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0737603Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0738080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0738565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0739044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0739534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0740015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0740538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0741041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0741521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0742000Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0742488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0743044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0743527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0744001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0744504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0744978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0745495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0745966Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0746457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0746935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0747416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0747945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0748458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0748945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0749418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0749895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0750373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0750845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0751331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0751845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0752334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0752841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0753317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0753798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0754280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0754757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0755241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0755746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0756240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0756717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0757214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0757708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0758201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0758687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0759184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0759696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0760226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0760710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0761209Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0761699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0762191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0762676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0763344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0763868Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0764363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0764852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0765349Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0767248Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0767783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0768276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0768765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0769363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0769850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0770336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0770832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0771306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0771799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0772313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0772832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0773327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0773819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0774308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0774794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0775329Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0775816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0776300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0776798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0777307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0777804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0778293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0778787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0779269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0779787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0780294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0780786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0781274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0781761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0782234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0782765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0783342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0783837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0784327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0784854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0785340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0785833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0786325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0786812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0787323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0787838Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0788325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0788822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0789315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0789810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0790333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0790828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0791315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0791809Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0792324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0792811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0793299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0793774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0794259Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0794790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0795304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0795800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0796288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0796781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0797271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0797792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0798282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 66%] 2023-03-31T05:54:17.0798770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0799254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0799771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0800253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0800741Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0801227Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0801726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0802208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0802724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0803386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0803877Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0804365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0804857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0805383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0805873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0806356Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0806844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0807395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0807883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0808373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0808863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0809347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0809829Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0810351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0810879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0811363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0811848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0812329Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0812851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0813337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0813826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0814311Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0814830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0815316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0815803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0816279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0816766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0817249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0817762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0818264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0818766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0819257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0819747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0820230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0820772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0821262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0821756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0822245Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0822761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0823335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0823823Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0824310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0824806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0825328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0825845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0826334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0826828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0827316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0827796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0828310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0828797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0829280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0829767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0830271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0830766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0831249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0831742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0832223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0832740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0833250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0833745Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0834225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0834712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0835193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0835710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0836189Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0836688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0837174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0837689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0838171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0838664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0839135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0839623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0840128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0840641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0841131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0841614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0842099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0842596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0843268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0843761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0844253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0844744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0845275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0845760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0846251Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0846738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0847222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0847704Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0848236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0848770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0849255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0849746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0850220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0850743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0851227Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0851718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0852205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0852727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0853212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0853698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0854183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0854676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0855162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0855679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0856192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0856680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0857159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0857649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0858164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0858647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0859130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0859613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0860127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0860614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0861095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0861571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0862058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0862551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0863148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0863663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0864155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0864638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0865125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0865672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0866160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0866648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0867132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0867649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0868130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0868618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0869104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0869596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0870080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0870599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0871104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0871596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0872081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0872561Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0873045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0873570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0874054Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0874545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0875027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0875550Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0876035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0876521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0877005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0877484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0877996Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0878508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0878991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0879480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0879971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0880461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0880974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0881463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0881953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0882440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0882948Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0883575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0884047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0884541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0885026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0885513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0886037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0886557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0887037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0887528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0888012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0888542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0889026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0889517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0890001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0890531Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0891014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0891503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0891990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0892476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0892953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0893464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0893965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0894452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0894917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0895400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0895928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0896423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0896902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0897395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0897907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0898393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0898876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 67%] 2023-03-31T05:54:17.0899363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0899847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0900333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0900844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0901352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0901832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0902320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0902804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0903372Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0903887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0904372Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0904857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0905344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0905845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0906333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0906814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0907301Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0907781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0908298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0908805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0909299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0909787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0910279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0910766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0911292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0911783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0912276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0912767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0913283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0913767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0914262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0914752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0915250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0915740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0916256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0916750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0917242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0917728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0918215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0918723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0919215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0919700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0920195Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0920711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0921197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0921688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0922182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0922674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0923278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0923806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0924357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0924846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0925334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0925823Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0926360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0926849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0927340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0927830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0928343Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0928823Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0929314Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0929803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0930292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0930774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0931290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0931802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0932295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0932780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0933263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0933747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0934267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0934750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0935241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0935755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0936242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0936721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0937201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0937686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0938179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0938687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0939188Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0939679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0940167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0940659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0941142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0941661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0942149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0942634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0943197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0943715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0944203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0944695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0945193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0945675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0946188Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0946700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0947195Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0947680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0948168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0948653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0949166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0949644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0950125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0950613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0951128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0951614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0952102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0952587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0953074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0953554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0954070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0954581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0955066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0955551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0956044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0956559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0957042Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0957529Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0958016Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0958525Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0959009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0959497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0959982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0960462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0960945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0961441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0961951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0962442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0962925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0963552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0964080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0964564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0965056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0965542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0966068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0966547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0967031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0967510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0968012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0968499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0969023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0969537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0970036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0970524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0971013Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0971496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0972015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0972486Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0972975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0973462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0973981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0974460Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0974949Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0975435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0975913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0976418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0976927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0977414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0977897Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0978378Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0978858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0979371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0979862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0980344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0980830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0981339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0981819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0982296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0982776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0983327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0983811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0984344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0984861Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0985347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0985836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0986322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0986838Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0987320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0987805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0988288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0988803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0989281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0989767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0990252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0990738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0991213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0991724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0992232Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0992722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0993202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0993692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0994165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0994676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0995158Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0995638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0996120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0996637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0997111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0997601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0998087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0998575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0999086Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 68%] 2023-03-31T05:54:17.0999607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1000093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1000588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1001077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1001565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1002077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1002562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1003164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1003659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1004194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1004679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1005149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1005642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1006125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1006653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1007166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1007657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1008140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1008632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1009113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1009636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1010121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1010610Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1011102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1011617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1012108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1012597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1013080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1013572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1014053Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1014557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1015069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1015559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1016041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1016522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1017029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1017518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1018001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1018488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1019004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1019487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1019968Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1020459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1020946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1021432Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1021941Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1022456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1023017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1023505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1023989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1024513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1024995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1025481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1025964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1026476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1026956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1027433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1027921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1028413Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1028902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1029419Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1029932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1030421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1030906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1031394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1031880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1032398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1032882Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1033371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1033892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1034375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1034861Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1035344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1035830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1036316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1036827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1037337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1037821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1038304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1038774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1039265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1039776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1040263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1040752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1041243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1041780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1042263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1042752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1043363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1043848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1044374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1044892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1045380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1045866Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1046359Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1046844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1047368Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1047857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1048350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1048832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1049370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1049835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1050321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1050802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1051297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1051782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1052294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1052808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1053288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1053772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1054265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1054773Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1055263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1055750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1056240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1056757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1057246Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1057737Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1058233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1058726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1059221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1059751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1060268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1060745Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1061230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1061716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1062235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1062726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1063291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1063778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1064301Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1064785Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1065272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1065758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1066250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1066732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1067247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1067757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1068247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1068727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1069214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1069699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1070214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1070699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1071186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1071672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1072172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1072659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1073149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1073638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1074117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1074627Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1075142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1075633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1076119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1076606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1077095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1077604Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1078087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1078571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1079061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1079571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1080059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1080535Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1081030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1081518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1082025Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1082530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1083006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1083605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1084091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1084575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1085113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1085598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1086084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1086565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1087089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1087570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1088059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1088542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1089036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1089520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1090038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1090554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1091044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1091527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1092016Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1092527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1093020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1093505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1093993Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1094492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1094986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1095466Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1095954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1096437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1096922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1097423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1097928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1098417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1098903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1099382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1099922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 69%] 2023-03-31T05:54:17.1100406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1100890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1101375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1101890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1102369Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1103062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1103579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1104065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1104545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1105075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1105572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1106066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1106549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1107042Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1107528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1108044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1108522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1109011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1109489Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1110002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1110481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1110971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1111458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1111946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1112458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1112964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1113446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1113932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1114409Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1114894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1115402Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1115885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1116352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1116840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1117358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1117842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1118329Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1118813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1119300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1119785Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1120292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1120811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1121294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1121782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1122265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1122775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1123389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1123876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1124361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1124909Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1125386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1125878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1126361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1126851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1127313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1127825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1128332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1128821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1129302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1129788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1130267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1130789Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1131266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1131753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1132238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1132742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1133219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1133701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1134178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1134664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1135173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1135688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1136179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1136666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1137141Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1137623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1138146Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1138625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1139106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1139584Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1140084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1140567Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1141056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1141544Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1142031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1142509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1143127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1143632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1144114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1144591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1145072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1145586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1146060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1146554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1147041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1147565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1148051Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1148545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1149017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1149515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1149998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1150520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1151030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1151519Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1152001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1152491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1152977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1153491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1153972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1154462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1154973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1155459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1155943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1156430Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1156915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1157396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1157902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1158447Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1158936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1159428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1159898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1160393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1160905Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1161390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1161875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1162369Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1162873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1163486Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1163973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1164469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1164951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1165492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1166010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1166502Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1166983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1167476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1167955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1168479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1168964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1169448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1169927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1170457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1170923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1171414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1171899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1172385Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1172865Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1173377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1173886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1174376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1174860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1175347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1175860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1176352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1176837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1177332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1177856Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1178342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1178832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1179321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1179812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1180298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1180811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1181321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1181792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1182275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1182765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1183328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1183843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1184332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1184815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1185304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1185818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1186308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1186790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1187280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1187766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1188284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1188795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1189287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1189765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1190254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1190732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1191254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1191737Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1192224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1192706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1193204Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1193686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1194179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1194662Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1195154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1195633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1196144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1196661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1197148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1197633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1198116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1198626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1199109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1199588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1200074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 70%] 2023-03-31T05:54:17.1200587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1201068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1201549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1202034Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1202524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1203011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1203640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1204144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1204624Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1205106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1205579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1206104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1206596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1207085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1207570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1208100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1208583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1209071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1209552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1210043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1210522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1211032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1211542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1212031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1212508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1212993Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1213475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1213988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1214450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1214939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1215420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1215932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1216408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1216886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1217364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1217850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1218333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1218871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1219379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1219860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1220337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1220821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1221332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1221812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1222289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1222767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1223348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1223831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1224318Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1224803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1225272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1225759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1226269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1226780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1227269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1227746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1228222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1228701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1229200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1229687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1230173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1230685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1231163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1231644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1232123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1232605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1233084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1233599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1234095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1234578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1235046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1235537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1236014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1236530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1237018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1237505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1237994Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1238508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1238991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1239474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1239952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1240441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1240918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1241428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1241944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1242431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1242912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1243523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1244052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1244543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1245030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1245519Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1246037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1246520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1246993Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1247487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1247972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1248458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1248974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1249490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1249973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1250456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1250935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1251448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1251932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1252416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1252899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1253416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1253895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1254376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1254863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1255347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1255827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1256339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1256845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1257331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1257797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1258280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1258764Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1259288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1259767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1260252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1260740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1261257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1261734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1262222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1262713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1263268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1263775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1264275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1264756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1265247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1265735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1266225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1266744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1267231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1267712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1268203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1268698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1269185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1269673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1270164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1270650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1271135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1271648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1272163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1272650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1273132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1273609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1274128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1274608Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1275094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1275573Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1276105Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1276577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1277063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1277548Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1278035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1278518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1279027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1279517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1280006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1280487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1280969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1281445Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1281949Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1282423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1282908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1283518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1284049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1284524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1285011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1285495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1285981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1286495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1287004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1287480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1287952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1288426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1288912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1289437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1289925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1290397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1290879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1291383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1291864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1292338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1292820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1293302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1293781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1294278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1294791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1295282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1295770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1296258Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1296784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1297272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1297762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1298242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1298759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1299242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1299729Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 71%] 2023-03-31T05:54:17.1300211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1300701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1301169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1301679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1302178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1302665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1303221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1303716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1304201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1304728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1305211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1305699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1306181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1306698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1307185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1307673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1308162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1308648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1309161Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1309670Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1310152Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1310642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1311127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1311615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1312113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1312606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1313087Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1313574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1314091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1314574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1315064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1315556Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1316041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1316525Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1317035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1317549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1318033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1318522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1319010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1319533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1320019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1320510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1320990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1321512Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1321993Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1322484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1322952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1323553Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1324027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1324571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1325091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1325585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1326070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1326564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1327049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1327571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1328056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1328540Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1329023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1329541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1330024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1330517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1331002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1331483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1331997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1332507Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1332991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1333472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1333943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1334431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1334959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1335436Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1335914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1336400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1336914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1337399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1337877Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1338363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1338848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1339332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1339844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1340348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1340826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1341303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1341773Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1342286Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1342766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1343321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1343804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1344314Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1344789Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1345276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1345762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1346246Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1346720Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1347223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1347722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1348211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1348693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1349179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1349658Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1350172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1350649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1351128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1351602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1352110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1352583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1353059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1353534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1354015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1354524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1355023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1355511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1355990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1356472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1356955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1357458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1357937Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1358413Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1358888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1359408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1359886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1360365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1360843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1361323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1361802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1362296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1362792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1363380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1363856Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1364336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1364821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1365328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1365812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1366293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1366779Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1367293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1367773Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1368250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1368738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1369208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1369715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1370221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1370705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1371180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1371664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1372144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1372653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1373132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1373609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1374088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1374593Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1375062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1375541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1376007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1376491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1376969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1383653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1384247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1384749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1385233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1385722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1386241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1386733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1387216Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1387703Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1388216Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1388701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1389183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1389678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1390167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1390655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1391168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1391678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1392164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1392639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1393123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1393612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1394126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1394612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1395098Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1395591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1396104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1396588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1397071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1397557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1398043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1398526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1399031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1399562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1400043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1400521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1401002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1401519Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1402008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1402498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1402979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1403647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1404129Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1404616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1405097Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1405583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 72%] 2023-03-31T05:54:17.1406057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1406594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1407099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1407588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1408070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1408554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1409034Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1409573Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1410057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1410546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1411033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1411554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1412032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1412506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1412981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1413465Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1413977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1414476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1414959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1415440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1415915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1416394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1416894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1417374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1417854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1418332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1418835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1419319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1419800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1420288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1420772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1421252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1421757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1422265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1422748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1423309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1423788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1424293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1424753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1425239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1425726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1426236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1426712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1427197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1427674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1428157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1428634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1429140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1429637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1430114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1430587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1431074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1431555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1432062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1432536Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1433014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1433487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1433992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1434462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1434940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1435408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1435885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1436386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1436890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1437381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1437866Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1438346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1438826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1439325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1439808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1440282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1440766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1441273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1441746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1442219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1442700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1443302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1443782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1444300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1444815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1445293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1445771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1446232Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1446708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1447224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1447706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1448187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1448680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1449199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1449688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1450165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1450648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1451122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1451626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1452125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1452609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1453082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1453557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1454033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1454541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1455018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1455497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1455969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1456458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1456926Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1457402Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1457873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1458352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1458820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1459353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1459864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1460346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1460826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1461308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1461822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1462302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1462772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1463322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1463828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1464301Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1464773Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1465255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1465740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1466215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1466708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1467215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1467700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1468173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1468650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1469130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1469632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1470110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1470582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1471066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1471576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1472060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1472541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1473014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1473491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1473970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1474472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1474976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1475455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1475938Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1476421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1476924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1477392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1477892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1478380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1478899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1479384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1479881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1480371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1480870Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1481352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1481869Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1482382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1482878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1483469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1483963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1484453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1484984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1485472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1485958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1486491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1486977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1487464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1487950Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1488425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1488908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1489434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1489955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1490439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1490923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1491413Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1491903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1492418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1492905Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1493389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1493878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1494390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1494874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1495357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1495852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1496338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1496849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1497360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1497847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1498334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1498820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1499308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1499820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1500291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1500777Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1501259Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1501774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1502262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1502750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1503321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1503810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1504296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1504811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1505324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 73%] 2023-03-31T05:54:17.1505808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1506293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1506780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1507293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1507787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1508273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1508762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1509284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1509767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1510260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1510746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1511223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1511715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1512222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1512728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1513211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1513703Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1514190Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1514710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1515193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1515683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1516164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1516705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1517184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1517673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1518151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1518637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1519111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1519625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1520132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1520623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1521104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1521594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1522075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1522577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1523177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1523666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1524154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1524682Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1525163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1525656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1526142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1526625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1527141Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1527667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1528151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1528636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1529121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1529611Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1530125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1530618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1531100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1531589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1532104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1532589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1533069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1533547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1534032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1534544Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1535047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1535536Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1536017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1536503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1536984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1537501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1537992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1538486Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1538973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1539495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1539980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1540480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1540972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1541465Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1541955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1542476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1543060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1543558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1544047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1544526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1545049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1545537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1546022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1546509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1547022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1547508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1547992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1548482Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1548967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1549454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1549975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1550481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1550971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1551454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1551942Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1552454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1552941Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1553429Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1553914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1554431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1554915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1555405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1555873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1556365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1556853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1557370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1557875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1558370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1558854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1559350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1559834Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1560346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1560827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1561320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1561808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1562320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1562806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1563411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1563899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1564387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1564907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1565429Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1565913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1566394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1566863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1567357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1567893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1568384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1568870Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1569359Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1569879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1570364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1570852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1571345Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1571833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1572348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1572858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1573350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1573835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1574323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1574806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1575347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1575833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1576321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1576808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1577321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1577799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1578275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1578756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1579247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1579732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1580243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1580753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1581241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1581723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1582207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1582719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1583286Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1583765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1584247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1584758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1585244Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1585725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1586214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1586698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1587183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1587695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1588201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1588685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1589156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1589640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1590155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1590639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1591125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1591609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1592124Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1592602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1593085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1593565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1594056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1594541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1595056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1595560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1596051Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1596533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1597032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1597516Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1598035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1598515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1599004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1599492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1599995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1600477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1600963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1601449Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1601936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1602442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1602957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1603557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1604043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1604522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1605009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1605534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 74%] 2023-03-31T05:54:17.1606020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1606500Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1606983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1607503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1607980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1608455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1608936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1609422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1609908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1610421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1610919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1611403Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1611891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1612374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1612879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1613354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1613831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1614304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1614818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1615302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1615790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1616269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1616756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1617235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1617751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1618255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1618737Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1619206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1619685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1620156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1620674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1621159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1621632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1622115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1622628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1623185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1623672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1624152Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1624638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1625116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1625620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1626125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1626618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1627103Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1627599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1628140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1628630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1629117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1629602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1630116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1630603Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1631088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1631577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1632062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1632534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1633052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1633586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1634075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1634559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1635047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1635562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1636046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1636539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1637022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1637535Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1638010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1638501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1638985Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1639474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1639954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1640465Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1640974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1641468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1641952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1642439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1642916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1643563Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1644042Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1644528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1645019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1645546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1646030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1646513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1646998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1647483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1648008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1648528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1649014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1649498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1649979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1650463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1650975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1651459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1651942Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1652425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1652936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1653418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1653898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1654374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1654856Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1655343Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1655848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1656366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1656851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1657347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1657831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1658352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1658836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1659323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1659807Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1660322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1660796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1661279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1661759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1662252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1662732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1663327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1663836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1664324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1664804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1665285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1665804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1666285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1666776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1667261Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1667767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1668253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1668739Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1669228Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1669718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1670199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1670702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1671210Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1671691Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1672174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1672656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1673139Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1673658Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1674146Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1674626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1675112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1675625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1676095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1676575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1677062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1677538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1678045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1678549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1679033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1679510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1680002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1680485Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1681002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1681484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1681969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1682453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1682965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1683572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1684050Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1684532Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1685011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1685487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1686023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1686544Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1687020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1687508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1687998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1688513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1688998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1689487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1689971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1690490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1690972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1691458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1691944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1692429Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1692911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1693445Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1693954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1694435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1694923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1695409Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1695893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1696408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1696884Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1697365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1697843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1698352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1698837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1699323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1699813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1700290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1700805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1701310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1701791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1702271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1702748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1703323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1703845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1704326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1704813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1705296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 75%] 2023-03-31T05:54:17.1705814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1706292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1706777Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1707259Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1707746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1708221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1708709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1709203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1709695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1710177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1710663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1711173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1711653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1712132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1712610Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1713116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1713594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1714068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1714545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1715019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1715506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1716014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1716534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1717025Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1717514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1718001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1718492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1719002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1719481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1719965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1720461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1720967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1721458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1721944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1722440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1722917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1723649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1724169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1724657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1725135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1725626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1726109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1726636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1727119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1727609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1728097Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1728622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1729104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1729595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1730078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1730551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1731030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1731546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1732058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1732538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1733022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1733513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1734021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1734508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1734989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1735475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1735987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1736474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1736957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1737441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1737921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1738401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1738904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1739417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1739904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1740386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1740864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1741361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1741843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1742334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1742882Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1743425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1743912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1744399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1744881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1745377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1745860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1746381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1746899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1747391Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1747879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1748367Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1748851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1749373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1749853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1750336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1750818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1751358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1751844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1752327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1752800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1753283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1753798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1754307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1754792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1755276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1755754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1756238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1756747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1757238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1757723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1758208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1758719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1759199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1759678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1760166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1760648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1761131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1761642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1762153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1762631Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1763278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1763750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1764285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1764772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1765252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1765733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1766254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1766736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1767218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1767697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1768182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1768659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1769184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1769694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1770183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1770663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1771148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1771631Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1772145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1772629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1773108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1773617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1774080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1774555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1775045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1775533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1776023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1776539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1777058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1777547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1778029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1778520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1779005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1779523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1780010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1780494Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1780987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1781501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1781986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1782467Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1783068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1783554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1784071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1784585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1785070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1785530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1786017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1786497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1787017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1787501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1787988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1788474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1788991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1789474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1789964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1790446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1790935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1791417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1791929Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1792437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1792928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1793411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1793896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1794415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1794902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1795384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1795869Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1796366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1796851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1797330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1797812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1798294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1798781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1799285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1799793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1800278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1800759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1801240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1801757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1802242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1802724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1803339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1803879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1804360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1804854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1805340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 76%] 2023-03-31T05:54:17.1805835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1806317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1806850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1807357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1807847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1808327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1808812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1809287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1809836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1810321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1810810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1811297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1811814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1812292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1812773Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1813258Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1813744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1814250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1814760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1815244Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1815733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1816316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1816805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1817321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1817803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1818272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1818755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1819262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1819743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1820224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1820707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1821184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1821668Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1822172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1822685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1823249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1823731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1824213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1824729Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1825212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1825698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1826176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1826691Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1827172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1827663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1828144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1828631Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1829094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1829612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1830116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1830603Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1831085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1831574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1832053Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1832570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1833046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1833523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1833999Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1834511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1834995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1835489Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1835972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1836456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1836964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1837472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1837957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1838433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1838912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1839388Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1839876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1840361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1840850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1841336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1841845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1842324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1842804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1843407Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1843887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1844370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1844893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1845405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1845880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1846365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1846849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1847367Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1847850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1848331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1848809Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1849326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1849796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1850278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1850739Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1851221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1852080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1852917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1853752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1854560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1855356Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1856143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1856940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1857767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1858562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1859360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1860163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1860975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1862024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1863092Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1863908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1864711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1865506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1866340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1867169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1867966Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1868760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1869653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1870798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1871679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1872485Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1873297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1874148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1874978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1875788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1876593Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1877406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1878506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1879387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1880265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1881089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1881885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1882696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1883631Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1884530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1885358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1886428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1887315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1888123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1888942Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1889753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1890559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1891376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1892211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1893096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1894061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1894925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1895743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1896540Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1897394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1898200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1899009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1899816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1900659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1901537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1902347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1903235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1904041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1904888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1905719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1906510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1907319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1908121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1908930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1909778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1910583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1911375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1912175Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1912998Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1913793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1914600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1915403Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1916197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1916996Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1917843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1918667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1919474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1920275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1921064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1921890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1922688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1923634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1924446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1925332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1926123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1926927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1927731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1928530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1929331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1930166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1930981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1931781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1932577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1933377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1934180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1935005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1935805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1936599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1937393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1938218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1939017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1939805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1940599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1941393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1942210Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 77%] 2023-03-31T05:54:17.1943106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1943918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1944700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1945495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1946292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1947121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1947925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1948727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1949512Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1950338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1951128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1951924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1952723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1953526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1954313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1955166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1955992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1956788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1957582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1958379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1959192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1959987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1960776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1961565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1962390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1963317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1964107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1964907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1965706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1966492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1967335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1968147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1968935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1969724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1970524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1971312Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1972155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1972955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1973762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1974563Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1975418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1976218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1977017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1977803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1978605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1979396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1980238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1981061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1981855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1982645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1983529Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1984364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1985157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1985950Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1986740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1987553Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1988344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1989139Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1989925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1990712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1991508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1992326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1993142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1993940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1994735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1995523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1996299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1997129Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1997925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1998714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.1999506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2000330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2001107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2001911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2002710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2003634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2004484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2005319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2006096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2006887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2007684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2008478Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2009318Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2010107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2010883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2011687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2012534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2013323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2014111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2014898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2015671Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2016452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2017280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2018093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2018888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2019662Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2020474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2021315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2022123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2023005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2023810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2024636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2025437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2026241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2027047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2027846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2028645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2029464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2030292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2031102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2031903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2032699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2033491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2034315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2035122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2035919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2036719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2037546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2038340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2039120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2039925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2040729Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2041569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2042394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2043348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2044135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2044931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2045734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2046595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2047390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2048184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2048964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2049818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2050618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2051416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2052213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2053017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2053797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2054641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2055476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2056274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2057068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2057849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2058680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2059476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2060279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2061077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2061906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2062689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2063579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2064376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2065177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2065976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2066809Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2067617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2068411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2069208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2070017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2070822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2071659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2072441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2073243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2074044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2074895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2075699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2076499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2077288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2078085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2078919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2079752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2080552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2081349Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2082130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2082926Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2083920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2084723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2085525Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2086318Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2087137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2087929Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2088728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2089524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2090320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2091110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2091933Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2092762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2093559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2094360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2095155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2095965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2096754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2097551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2098345Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2099170Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2099969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2100752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2101554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2102493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2103380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2104225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2105046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2105831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2106617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 78%] 2023-03-31T05:54:17.2107414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2108213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2109043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2109837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2110621Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2111411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2112231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2113020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2113814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2121817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2122747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2123848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2124688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2125483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2126301Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2127109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2127915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2128765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2129579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2130370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2131173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2132014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2132820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2133624Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2134413Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2135203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2136009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2136857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2137681Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2138479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2139285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2140075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2140903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2141699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2142499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2143392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2144225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2145006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2145808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2146612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2147415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2148205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2149033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2149847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2150638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2151433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2152234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2153028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2153838Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2154639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2155439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2156241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2157077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2157874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2158657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2159458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2160253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2161099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2161921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2162719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2163638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2164440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2165243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2166094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2166894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2167695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2168481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2169323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2170125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2170916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2171716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2172505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2178004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2182515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2183457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2184260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2185102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2185919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2186779Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2187580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2188377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2189171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2189999Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2190794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2191579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2192381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2193188Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2193990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2194833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2195665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2196455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2197259Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2198057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2198854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2199684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2200470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2201237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2202056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2202861Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2203813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2204607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2205399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2206202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2207074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2207914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2208711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2209498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2210270Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2211052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2211887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2212697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2213496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2214291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2215114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2215913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2216713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2217504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2218295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2219082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2219885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2220694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2221490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2222294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2223170Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2224006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2224788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2225578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2226372Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2227208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2227995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2228784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2229560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2230350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2231149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2231988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2232813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2233618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2234400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2235195Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2235985Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2236814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2237605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2238374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2239157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2239980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2240770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2241569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2242368Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2243273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2244110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2244934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2245719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2246501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2247282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2248048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2248867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2249654Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2250442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2251238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2252069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2252841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2253633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2254427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2255214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2255994Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2256805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2257598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2258399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2259183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2259973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2260768Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2261579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2262367Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2263237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2264031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2264854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2265640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2266408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2267192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2267992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2268813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2269640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2270441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2271221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2272011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2272801Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2273624Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2274405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2275193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2275659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2276168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2276641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2277117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2277609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2278097Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2278588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2279099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2279610Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2280095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2280582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2281064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2281581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 79%] 2023-03-31T05:54:17.2282064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2282551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2283167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2283713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2284191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2284683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2285160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2285644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2286128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2286640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2287162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2287647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2288121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2288602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2289075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2289598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2290083Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2290574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2291051Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2291566Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2292046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2292533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2293015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2293497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2294008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2294506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2294982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2295468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2295955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2296437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2296946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2297415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2297896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2298382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2298904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2299386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2299860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2300339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2300815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2301296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2301816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2302326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2302912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2303404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2303883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2304399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2304873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2305355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2305829Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2306345Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2306821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2307312Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2307788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2308262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2308741Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2309263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2309765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2310254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2310736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2311220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2311698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2312214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2312688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2313172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2313656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2314167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2314641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2315120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2315593Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2316075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2316544Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2317052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2317556Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2318027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2318487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2318970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2319483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2319969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2320451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2320925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2321431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2321903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2322382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2322865Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2323474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2323950Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2324475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2325004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2325490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2325974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2326453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2326930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2327444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2327921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2328394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2328876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2329383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2329861Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2330331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2330814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2331290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2331802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2332302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2332782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2333255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2333729Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2334200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2334711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2335180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2335666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2336147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2336668Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2337154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2337640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2338122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2338599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2339072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2339578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2340083Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2340571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2341051Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2341523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2342001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2342515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2343081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2343566Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2344044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2344557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2345036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2345514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2345989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2346469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2347012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2347524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2348004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2348490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2348974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2349453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2349964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2350428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2350901Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2351380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2351896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2352374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2352848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2353326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2353803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2354281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2354787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2355293Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2355778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2356250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2356721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2357228Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2357702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2358182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2358659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2359173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2359649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2360137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2360605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2361084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2361558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2362070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2362570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2363174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2363649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2364131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2364605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2365133Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2365614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2366111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2366599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2367133Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2367613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2368107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2368594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2369084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2369605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2370132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2370623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2371107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2371577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2372070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2372586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2373070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2373554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2374044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2374570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2375055Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2375541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2376030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2376518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2377009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2377523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2378049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2378532Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2379026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2379514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2380033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2380516Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2381008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 80%] 2023-03-31T05:54:17.2381491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2382003Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2382482Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2383040Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2383528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2384020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2384500Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2385018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2385528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2386023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2386506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2386994Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2387507Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2387995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2388479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2388967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2389477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2389968Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2390455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2390945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2391433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2391916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2392434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2392946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2393436Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2393922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2394398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2394886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2395398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2395897Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2396383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2396908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2397392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2397880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2398361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2398863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2399353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2399864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2400376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2400864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2401346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2401837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2402324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2402842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2403454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2403945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2404431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2404977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2405451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2405944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2406432Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2406918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2407438Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2407966Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2408458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2408943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2409427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2409915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2410458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2410946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2411431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2411920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2412435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2412914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2413397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2413889Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2414378Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2414901Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2415410Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2415902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2416382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2416857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2417339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2417858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2418339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2418831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2419310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2419829Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2420317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2420808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2421294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2421785Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2422264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2422780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2423368Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2423859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2424342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2424823Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2425335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2425836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2426328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2426818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2427330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2427812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2428300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2428798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2429287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2429776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2430290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2430815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2431305Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2431800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2432285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2432806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2433295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2433785Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2434269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2434785Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2435267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2435750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2436237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2436727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2437212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2437733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2438239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2438739Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2439204Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2439692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2440202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2440693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2441179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2441666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2442180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2442663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2443311Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2443806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2444296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2444781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2445305Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2445828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2446318Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2446807Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2447292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2447774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2448295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2448786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2449274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2449807Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2450292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2450771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2451256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2451748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2452231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2452746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2453254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2453742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2454223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2454707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2455187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2455713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2456201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2456695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2457184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2457715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2458200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2458691Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2459182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2459666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2460174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2460684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2461171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2461661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2462135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2462620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2463219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2463708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2464194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2464682Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2465194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2465679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2466168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2466656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2467140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2467651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2468161Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2468654Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2469134Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2469621Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2470100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2470615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2471100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2471591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2472070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2472587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2473055Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2473547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2474032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2474520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2475007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2475517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2476028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2476513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2476997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2477483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2477991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2478476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2478964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2479452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2479964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2480445Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2480932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2481420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 81%] 2023-03-31T05:54:17.2481904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2482389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2482897Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2483560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2484034Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2484519Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2485002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2485539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2486028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2486514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2486995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2487526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2488011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2488504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2488989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2489473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2489957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2490474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2490982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2491473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2491958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2492443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2492927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2493434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2493913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2494399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2494883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2495384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2495864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2496352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2496841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2497332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2497839Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2498343Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2498830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2499309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2499794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2500279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2500786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2501273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2501754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2502239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2502752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2503321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2503803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2504287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2504776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2505262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2505770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2506275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2506755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2507236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2507717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2508228Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2508707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2509197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2509683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2510225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2510710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2511197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2511675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2512168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2512649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2513164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2513671Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2514157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2514629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2515120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2515607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2516125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2516609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2517088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2517577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2518100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2518590Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2519082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2519572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2520059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2520567Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2521104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2521591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2522078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2522568Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2523195Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2523725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2524218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2524701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2525184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2525712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2526196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2526679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2527171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2527648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2528158Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2528681Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2529173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2529651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2530143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2530626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2531144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2531629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2532119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2532597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2533125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2533612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2534098Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2534580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2535072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2535581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2536099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2536589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2537075Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2537560Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2538044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2538555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2539043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2539518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2540003Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2540518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2541002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2541488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2541970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2542457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2543012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2543525Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2544032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2544514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2545007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2545493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2546016Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2546503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2546997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2547480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2547997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2548484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2548974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2549460Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2549949Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2550417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2550947Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2551457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2551947Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2552428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2552916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2553401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2553920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2554403Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2554895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2555402Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2555881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2556362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2556849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2557334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2557817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2558325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2558835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2559318Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2559806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2560295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2560782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2561290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2561757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2562240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2562729Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2563371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2563863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2564351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2564839Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2565324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2565853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2566362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2566854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2567332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2567816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2568294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2568818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2569300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2569786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2570263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2570781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2571262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2571751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2572235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2572709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2573188Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2573698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2574201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2574694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2575180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2575674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2576187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2576675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2577161Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2577649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2578164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2578648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2579127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2579611Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2580093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2580579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2581112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2581622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2582106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 82%] 2023-03-31T05:54:17.2582588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2583135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2583614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2584124Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2584615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2585098Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2585607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2586085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2586572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2587056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2587547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2588029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2588562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2589078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2589565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2590038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2590522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2591006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2591514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2591993Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2592481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2592967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2593478Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2593964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2594439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2594922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2595406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2595907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2596415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2596896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2597379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2597864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2598354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2598868Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2599355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2599839Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2600323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2600827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2601312Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2601797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2602280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2602767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2603446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2603965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2604490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2604980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2605458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2605944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2606470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2606958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2607452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2607936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2608470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2608958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2609451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2609934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2610427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2610909Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2611424Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2611933Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2612422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2612906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2613393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2613877Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2614387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2614875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2615364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2615886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2616373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2616846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2617340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2617824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2618313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2618824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2619332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2619821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2620315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2620800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2621288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2621795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2622290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2622780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2623349Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2623873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2624362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2624844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2625328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2625812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2626325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2626830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2627319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2627788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2628278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2628763Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2629278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2629760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2630247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2630738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2631262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2631741Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2632231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2632715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2633202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2633678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2634194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2634705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2635191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2635677Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2636165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2636675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2637159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2637639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2638131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2638652Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2639132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2639617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2640104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2640594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2641079Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2641585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2642102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2642584Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2643211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2643693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2644226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2644708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2645193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2645676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2646204Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2646686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2647173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2647657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2648143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2648628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2649149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2649665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2650143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2650621Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2651101Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2651579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2652100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2652581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2653069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2653554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2654071Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2654555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2655044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2655529Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2656008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2656515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2657015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2657498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2657986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2658470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2658953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2659460Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2659940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2660424Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2660897Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2661400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2661881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2662366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2662931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2663421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2663943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2664452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2664952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2665439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2665931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2666410Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2666935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2667423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2667914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2668404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2668927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2669411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2669901Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2670390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2670880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2671364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2671866Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2672383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2672873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2673355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2673842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2674358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2674841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2675322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2675817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2676337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2676820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2677307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2677791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2678279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2678759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2679267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2679776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2680261Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2680746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2681235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2681754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2682241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 83%] 2023-03-31T05:54:17.2682734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2683341Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2683882Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2684363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2684852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2685334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2685824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2686304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2686824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2687341Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2687840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2688320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2688808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2689291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2689816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2690298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2690787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2691271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2691784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2692270Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2692753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2693236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2693725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2694222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2694741Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2695228Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2695713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2696196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2696678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2697212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2697698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2698180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2698668Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2699183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2699669Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2700155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2700645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2701126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2701639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2702147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2702634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2703190Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2703680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2704162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2704676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2705156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2705633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2706120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2706643Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2707122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2707605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2708089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2708577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2709057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2709587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2710104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2710586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2711067Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2711555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2712068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2712553Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2713040Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2713527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2714039Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2714520Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2715003Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2715489Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2715970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2716439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2716944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2717456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2717944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2718423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2718902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2719386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2719891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2720379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2720863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2721381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2721858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2722342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2722826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2723442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2723923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2724458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2724971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2725457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2725938Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2726426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2726906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2727414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2727899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2728380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2728859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2729384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2729862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2730351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2730831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2731315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2731822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2732328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2732809Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2733288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2733769Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2734249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2734750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2735233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2735716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2736201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2736717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2737199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2737679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2738155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2738640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2739121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2739623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2740119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2740591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2741072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2741554Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2742043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2742549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2743123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2743607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2744126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2744604Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2745085Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2745564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2746043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2746515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2747031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2747545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2748038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2748523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2748988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2749464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2749974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2750450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2750935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2751414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2751924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2752401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2752894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2753383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2753878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2754390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2754928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2755420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2755907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2756396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2756886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2757399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2757888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2758373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2758862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2759381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2759855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2760341Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2760828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2761314Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2761797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2762304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2762813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2763429Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2763916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2764398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2764927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2765412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2765896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2766380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2766908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2767387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2767876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2768361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2768847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2769324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2769844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2770355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2770831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2771315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2771805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2772292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2772828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2773314Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2773799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2774285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2774806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2775289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2775772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2776253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2776738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2777247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2777757Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2778241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2778721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2779200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2779686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2780191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2780675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2781156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2781630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2782145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 84%] 2023-03-31T05:54:17.2782635Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2783191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2783679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2784166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2784686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2785194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2785686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2786165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2786652Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2787130Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2787645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2788122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2788612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2789094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2789601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2790078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2790558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2791035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2791523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2792000Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2792496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2793001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2793487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2793965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2794451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2794961Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2795448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2795931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2796414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2796924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2797406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2797890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2798377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2798863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2799344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2799849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2800366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2800848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2801335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2801818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2802304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2802816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2803435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2803902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2804388Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2804924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2805406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2805885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2806375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2806859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2807380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2807889Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2808381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2808862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2809347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2809829Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2810352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2810832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2811319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2811795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2812315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2812802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2813290Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2813774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2814249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2814731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2815267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2815767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2816256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2816738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2817224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2817734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2818225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2818710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2819191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2819702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2820185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2820667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2821150Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2821638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2822121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2822626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2823198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2823677Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2824169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2824650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2825125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2825642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2826121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2826597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2827080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2827600Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2828086Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2828563Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2829040Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2829517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2830029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2830534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2831019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2831499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2831980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2832459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2832972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2833455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2833939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2834421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2834927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2835401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2835873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2836357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2836842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2837322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2837832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2838326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2838805Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2839279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2839758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2840262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2840743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2841218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_128_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2841714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2842230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2842716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2843332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2843834Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2844327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2844819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2845344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2845873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2846363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2846837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2847323Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2847822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2848341Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2848841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2849324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2849853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2850337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2850826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2851308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2851803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2852286Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2852802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2853322Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2853817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2854298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2854790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2855273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2855790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2856273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2856764Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2857250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2857762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2858233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2858721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2859204Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2859699Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2860210Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2860721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2861209Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2861694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2862177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2862665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2863250Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2863736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2864217Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2864702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2865218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2865715Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2866201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2866694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2867176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2867687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2868200Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2868689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2869170Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2869641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2870128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2870641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2871126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2871617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2872106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2872660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2873144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2873637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2874117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2874613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2875126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2875641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2876133Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2876619Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2877104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2877597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2878111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2878601Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2879093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2879583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2880101Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2880577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2881063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2881549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2882031Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 85%] 2023-03-31T05:54:17.2882515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2883149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2883683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2884168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2884657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2885143Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2885664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2886151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2886637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2887121Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2887646Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2888126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2888606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2889088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2889575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2890057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2890581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2891095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2891582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2892049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2892538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2893061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2893549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2894039Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2894526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2895036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2895524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2896009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2896496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2896981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2897472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2897982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2898491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2898976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2899468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2899954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2900438Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2900948Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2901442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2901934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2902427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2903024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2903522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2904009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2904505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2904991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2905509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2906027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2906521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2907002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2907493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2907977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2908499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2908986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2909475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2909958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2910478Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2910966Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2911448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2911939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2912426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2912939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2913450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2913934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2914412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2914898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2915389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2915908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2916396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2916878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2917363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2917875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2918358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2918845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2919338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2919821Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2920339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2920852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2921342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2921827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2922320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2922800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2923464Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2923952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2931045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2931615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2932220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2932712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2933212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2933706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2934207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2934698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2935218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2935752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2936246Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2936733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2937226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2937771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2938272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2938760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2939253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2939782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2940278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2940772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2941263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2941753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2942236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2942742Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2943394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2943882Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2944376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2944867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2945385Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2945875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2946354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2946844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2947362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2947849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2948339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2948825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2949311Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2949791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2950308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2950814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2951307Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2951786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2952274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2952782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2953276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2953759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2954246Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2954754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2955240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2955727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2956218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2956705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2957191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2957692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2958210Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2958698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2959178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2959666Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2960152Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2960664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2961150Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2961634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2962128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2962638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2963255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2963744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2964237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2964726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2965255Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2965775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2966262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2966743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2967228Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2967711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2968225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2968696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2969186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2969672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2970206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2970686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2971174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2971659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2972151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2972656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2973165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2973651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2974137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2974622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2975104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2975613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2976099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2976577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2977062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2977569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2978058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2978537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2979018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2979490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2979976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2980488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2981000Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2981488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2981968Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2982450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2983056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2983541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2984029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2984514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2985033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2985511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2986001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2986487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2986976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2987459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2987971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2988479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2988965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2989448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2989931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 86%] 2023-03-31T05:54:17.2990400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2990904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2991386Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2991871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2992352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2992874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2993355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2993847Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2994339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2994815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2995321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2995852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2996336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2996816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2997297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2997790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2998303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2998793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2999285Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.2999778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3000299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3000787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3001260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3001750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3002236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3002730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3003503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3004035Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3004521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3005010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3005498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3006023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3006510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3007000Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3007488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3008017Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3008499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3008984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3009476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3009976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3010463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3010975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3011477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3011974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3012456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3012932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3013441Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3013928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3014409Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3014898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3015410Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3015899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3016389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3016875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3017358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3017842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3018354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3018884Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3019373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3019861Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3020342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3020827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3021333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3021825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3022316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3022898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3023385Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3023874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3024352Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3024837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3025317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3025835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3026341Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3026825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3027303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3027792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3028276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3028791Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3029275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3029768Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3030249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3030764Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3031245Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3031733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3032214Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3032701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3033207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3033722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3034208Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3034693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3035162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3035653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3036162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3036645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3037126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3037612Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3038125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3038607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3039090Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3039578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3040063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3040556Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3041067Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3041583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3042063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3042551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3043159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3043692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3044173Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3044658Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3045142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3045682Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3046151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3046644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3047127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3047618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3048101Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3048617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3049129Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3049615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3050094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3050580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3051088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3051576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3052057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3052542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3053046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3053526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3054009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3054494Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3054977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3055456Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3055978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3056484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3056956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3057448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3057936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3058426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3058939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3059422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3059910Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3060397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3060911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3061399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3061887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3062376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3062919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3063436Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3063940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3064428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3064911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3065394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3065872Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3066383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3066862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3067351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3067831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3068331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3068811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3069300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3069786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3070280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3070762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3071265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3071780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3072263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3072744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3073230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3073740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3074221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3074704Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3075197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3075712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3076192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3076674Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3077159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3077642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3078124Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3078656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3079157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3079640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3080123Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3080606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3081091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3081611Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3082096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3082586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3083197Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3083721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3084207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3084694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3085179Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3085657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3086168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3086683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3087180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3087664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3088159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3088648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3089177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3089663Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3090148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 87%] 2023-03-31T05:54:17.3090633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3091149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3091630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3092115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3092598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3093094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3093607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3094118Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3094607Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3095090Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3095573Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3096057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3096569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3097056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3097542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3098026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3098546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3099032Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3099514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3100001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3100489Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3100959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3101467Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3101979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3102463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3103057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3103660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3104193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3104672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3105163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3105648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3106166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3106654Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3107138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3107625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3108111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3108595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3109109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3109620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3110111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3110589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3111077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3111595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3112082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3112555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3113044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3113586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3114066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3114553Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3115033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3115519Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3116001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3116505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3117021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3117510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3118002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3118485Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3118978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3119491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3119981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3120466Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3120956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3121473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3121963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3122448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3122940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3123530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3124065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3124589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3125078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3125562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3126050Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3126530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3127048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3127528Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3128007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3128485Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3129013Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3129494Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3129986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3130472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3130959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3131467Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3131978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3132462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3132946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3133424Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3133908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3134405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3134891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3135378Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3135859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3136367Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3136853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3137334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3137817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3138302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3138788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3139299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3139810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3140291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3140783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3141264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3141784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3142260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3142749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3143306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3143828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3144308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3144790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3145275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3145747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3146226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3146743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3147252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3147748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3148232Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3148723Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3149205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3149733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3150220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3150709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3151230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3151713Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3152196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3152685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3153167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3153653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3154163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3154675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3155160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3155650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3156131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3156606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3157113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3157599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3158080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3158576Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3159088Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3159571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3160058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3160547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3161030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3161540Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3162046Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3162531Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3163012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3163635Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3164119Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3164648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3165132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3165623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3166107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3166633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3167110Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3167599Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3168073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3168559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3169072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3169588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3170068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3170552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3171037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3171523Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3172058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3172543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3173024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3173507Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3174015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3174492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3174969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3175452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3175935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3176426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3176937Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3177458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3177946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3178439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3178912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3179425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3179906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3180396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3180881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3181398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3181879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3182370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3182930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3183427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3183907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3184423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3184932Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3185418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3185896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3186379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3186886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3187370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3187846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3188333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3188844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3189334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3189807Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3190292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 88%] 2023-03-31T05:54:17.3190778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3191260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3191770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3192276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3192756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3193239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3193721Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3194213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3194732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3195220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3195705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3196192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3196706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3197191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3197673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3198158Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3198633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3199146Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3199655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3200140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3200624Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3201100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3201581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3202091Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3202569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3203168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3203649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3204189Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3204668Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3205155Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3205640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3206134Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3206626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3207153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3207673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3208160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3208646Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3209132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3209643Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3210127Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3210613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3211100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3211608Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3212079Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3212559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3213041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3213521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3214002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3214509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3215014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3215501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3215984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3216466Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3216942Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3217440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3217923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3218407Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3218898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3219402Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3219883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3220361Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3220850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3221328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3221835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3222332Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3222801Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3223354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3223843Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3224326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3224840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3225320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3225801Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3226280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3226789Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3227272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3227748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3228222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3228698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3229166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3229678Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3230218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3230703Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3231187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3231664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3232168Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3232650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3233120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3233604Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3234109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3234586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3235062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3235552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3236045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3236533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3237043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3237550Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3238033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3238514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3238999Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3239487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3239999Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3240482Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3240969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3241461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3241979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3242463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3242956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3243561Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3244041Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3244549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3245062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3245549Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3246029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3246521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3247001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3247527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3248011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3248497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3248983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3249506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3249983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3250473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3250956Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3251443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3251951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3252463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3252946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3253434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3253921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3254408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3254922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3255394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3255880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3256368Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3256887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3257366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3257845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3258331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3258819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3259306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3259812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3260329Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3260812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3261295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3261778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3262286Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3262772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3263330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3263810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3264328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3264809Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3265310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3265794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3266270Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3266754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3267267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3267772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3268265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3268744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3269230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3269705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3270212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3270697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3271192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3271702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3272187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3272672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3273157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3273639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3274122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3274628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3275132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3275614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3276099Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3276577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3277056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3277569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3278056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3278539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3279026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3279535Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3280014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3280492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3280979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3281462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3281966Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3282472Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3282963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3283660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3284212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3284698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3285238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3285717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3286205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3286682Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3287201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3287685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3288157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3288635Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3289120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3289596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3290140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 89%] 2023-03-31T05:54:17.3290650Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3291138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3291625Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3292109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3292616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3293100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3293577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3294057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3294558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3295044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3295533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3296023Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3296508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3296990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3297497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3298005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3298491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3298959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3299439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3299918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3300430Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3300917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3301400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3301886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3302392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3302949Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3303433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3303921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3304408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3304920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3305422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3305901Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3306372Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3306858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3307339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3307850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3308328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3308813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3309291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3309803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3310281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3310760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3311233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3311712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3312185Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3312691Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3313199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3313686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3314163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3314647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3315153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3315640Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3316120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3316596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3317094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3317568Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3318040Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3318522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3319013Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3319494Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3320005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3320496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3320970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3321443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3321924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3322401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3322903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3323511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3323987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_256_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3324476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3325011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3325501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3325988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3326481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3326965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3327483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3328003Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3328488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3328963Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3329448Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3329926Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3330449Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3330935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3331406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3331889Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3332399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3332873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3333362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3333845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3334337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3334816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3335331Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3335845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3336334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3336815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3337297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3337804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3338284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3338762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3339248Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3339753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3340231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3340710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3341191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3341672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3342142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3342652Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3343230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3343724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3344207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3344688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3345172Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3345683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3346166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3346645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3347125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3347658Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3348149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3348629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3349114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3349596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3350107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3350609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3351100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3351581Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3352063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3352543Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3353036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3353511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3354001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3354484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3355005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3355490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3355976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3356459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3356945Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3357421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3357937Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3358443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3358936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3359417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3359907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3360417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3360898Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3361383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3361865Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3362384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3362872Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3363480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3363952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3364433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3364909Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3365426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3365943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3366428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3366908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3367392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3367881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3368400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3368884Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3369363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3369845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3370357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3370834Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3371308Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3371796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3372275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3372784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3373284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3373770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3374234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3374719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3375199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3375705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3376182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3376659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3377131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3377644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3378124Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3378609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3379096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3379579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3380057Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3380563Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3381065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3381550Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3382027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3382504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3383102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3383602Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3384089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3384585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3385086Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3385577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3386062Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3386546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3387036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3387521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3388024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3388540Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3389021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3389509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 90%] 2023-03-31T05:54:17.3389987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3390470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3390978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3391458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3391947Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3392435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3392957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3393442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3393925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3394414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3394892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3395409Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3395907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3396396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3396878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3397364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3397845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3398362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3398841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3399326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3399811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3400319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3400801Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3401287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3401766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3402257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3402738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3403396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3403918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3404407Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3404886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3405373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3405915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3406398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3406867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3407353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3407873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3408358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3408839Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3409324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3409812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3410294Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3410804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3411309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3411787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3412262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3412734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3413226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3413738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3414225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3414710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3415198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3415714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3416202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3416685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3417175Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3417636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3418144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3418648Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3419136Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3419617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3420102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3420589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3421103Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3421579Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3422064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3422541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3423141Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3423620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3424100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3424572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3425061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3425542Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3426060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3426568Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3427052Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3427534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3428018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3428518Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3428995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3429473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3429950Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3430453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3430938Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3431421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3431904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3432389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3432871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3433376Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3433883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3434365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3434848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3435328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3435804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3436309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3436795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3437281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3437774Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3438281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3438761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3439224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3439709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3440182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3440685Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3441194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3441672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3442140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3442628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3443238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3443771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3444251Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3444737Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3445213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3445730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3446206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3446693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3447174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3447655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3448126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3448638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3449151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3449624Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3450095Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3450577Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3451073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3451552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3452027Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3452504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3453009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3453487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3453968Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3454451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3454930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3455412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3455921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3456426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3456903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3457382Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3457856Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3458334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3458838Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3459319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3459795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3460265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3460772Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3461247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3461725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3462206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3462683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3463236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3463743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3464274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3464748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3465236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3465716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3466225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3466707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3467193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3467670Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3468187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3468665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3469145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3469629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3470112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3470574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3471078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3471580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3472072Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3472559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3473048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3473534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3474047Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3474525Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3475011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3475493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3476005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3476484Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3476969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3477454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3477946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3478451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3478959Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3479443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3479924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3480406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3480892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3481391Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3481875Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3482354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3482832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3483473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3483955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3484437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3484924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3485405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3485888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3486406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3486923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3487410Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3487888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3488363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3488981Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3489510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 91%] 2023-03-31T05:54:17.3490002Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3490483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3491014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3491496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3491977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3492449Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3492934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3493411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3493920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3494428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3494906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3495380Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3495874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3496359Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3496874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3497355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3497836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3498315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3498818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3499295Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3499778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3500254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3500728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3501196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3501706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3502219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3502690Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3503277Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3503768Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3504282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3504763Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3505243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3505730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3506237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3506712Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3507187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3507670Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3508157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3508638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3509140Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3509647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3510131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3510606Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3511080Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3511562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3512069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3512544Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3513018Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3513487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3513995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3514479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3514958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3515438Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3515913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3516415Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3516912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3517390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3517864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3518339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3518815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3519326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3519814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3520304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3520780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3521297Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3521770Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3522247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3522718Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3523338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3523802Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3524343Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3524845Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3525326Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3525800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3526281Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3526794Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3527277Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3527747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3528224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3528735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3529212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3529687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3530165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3530647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3531134Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3531641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3532148Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3532630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3533108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3533588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3534067Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3534552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3535029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3535501Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3535975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3536477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3536958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3537435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3537919Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3538392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3538864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3539355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3539857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3540333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3540811Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3541283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3541790Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3542269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3542756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3543325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3543837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3544310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3544778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3545248Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3545732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3546203Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3546701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3547202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3547688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3548162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3548647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3549125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3549633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3550108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3550585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3551058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3551565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3552033Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3552511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3552980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3553468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3553944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3554450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3554952Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3555419Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3555893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3556374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3556883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3557359Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3557842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3558324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3558830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3559303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3559778Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3560279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3560769Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3561257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3561767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3562291Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3562779Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3563406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3563892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3564383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3564917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3565410Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3565895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3566374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3566902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3567391Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3567879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3568371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3568855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3569383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3569899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3570385Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3570864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3571346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3571827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3572347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3572828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3573313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3573798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3574330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3574812Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3575298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3575788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3576278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3576780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3577278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3577762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3578253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3578738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3579222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3579738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3580222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3580707Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3581194Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3581732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3582215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3582701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3583267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3583752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3584274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3584788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3585278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3585759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3586247Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3586733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3587254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3587736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3588209Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3588692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 92%] 2023-03-31T05:54:17.3589213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3589693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3590191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3590681Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3591175Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3591655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3592176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3592679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3593169Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3593656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3594149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3594661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3595149Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3595632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3596118Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3596632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3597113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3597592Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3598084Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3598564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3599045Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3599538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3600056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3600536Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3601021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3601503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3601996Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3602506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3602992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3603594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3604082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3604618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3605108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3605589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3606079Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3606555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3607076Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3607586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3608079Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3608564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3609049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3609533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3610060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3610531Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3611015Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3611499Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3612014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3612487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3612967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3613450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3613934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3614442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3614953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3615437Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3615920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3616404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3616892Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3617401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3617888Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3618368Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3618848Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3619357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3619849Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3620340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3620833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3621310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3621804Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3622319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3622988Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3623479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3623965Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3624450Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3624976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3625461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3625953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3626439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3626970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3627452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3627938Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3628421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3628911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3629389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3629903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3630407Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3630896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3631375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3631867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3632383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3632857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3633340Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3633825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3634335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3634820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3635309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3635793Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3636276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3636762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3637279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3637792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3638278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3638765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3639252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3639738Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3640272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3640762Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3641251Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3641740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3642249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3642735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3643339Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3643836Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3644309Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3644844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3645362Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3645854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3646334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3646822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3647302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3647824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3648304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3648789Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3649274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3649814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3650296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3650786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3651275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3651759Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3652276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3652796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3653284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3653771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3654254Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3654740Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3655240Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3655731Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3656221Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3656710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3657226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3657709Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3658196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3658687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3659180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3659667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3660174Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3660683Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3661163Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3661652Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3662138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3662653Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3663224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3663714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3664202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3664722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3665202Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3665686Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3666154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3666641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3667118Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3667636Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3668147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3668632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3669114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3669598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3670113Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3670598Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3671081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3671565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3672073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3672555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3673039Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3673524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3674012Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3674496Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3675006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3675511Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3675992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3676473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3676941Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3677427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3677940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3678424Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3678905Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3679398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3679925Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3680411Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3680894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3681385Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3681868Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3682404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3682914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3683536Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3684019Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3684508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3684987Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3685526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3686007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3686492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3686971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3687498Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3687970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3688455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3688940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 93%] 2023-03-31T05:54:17.3689425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3689903Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3690412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3690913Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3691399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3691878Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3692366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3692873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3693357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3693840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3694321Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3694827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3695302Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3695782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3696256Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3696728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3697215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3697725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3698263Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3698732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3699215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3699698Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3700181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3700689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3701165Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3701639Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3702109Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3702609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3703167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3703659Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3704146Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3704628Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3705144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3705661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3706144Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3706619Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3707102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3707582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3708094Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3708570Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3709061Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3709534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3710055Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3710541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3711030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3711514Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3712005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3712488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3713011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3713521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3714013Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3714497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3714991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3715508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3715996Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3716480Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3716967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3717477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3717960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3718443Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3718924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3719408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3719891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3720390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3720902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3721388Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3721870Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3722356Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3722841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3723513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3724005Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3724487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3724975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3725491Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3725977Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3726459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3726954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3727435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3727957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3728468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3728955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3729433Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3729920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3730401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3730917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3731398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3731872Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3732357Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3732883Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3733365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3733855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3734342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3734825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3735330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3735835Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3736319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3736798Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3737279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3737761Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3738280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3738771Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3739260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3739752Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3740264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3740749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3741230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3741716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3742205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3742676Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3743272Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3743788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3744271Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3744765Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3745249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3745763Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3746244Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3746732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3747213Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3747737Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3748212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3748701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3749180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3749665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3750142Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3750654Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3757160Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3757733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3758229Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3758730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3759219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3759767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3760251Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3760744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3761233Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3761749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3762237Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3762728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3763397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3763886Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3764461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3765004Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3765487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3765973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3766454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3766939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3767459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3767946Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3768428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3768915Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3769442Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3769924Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3770405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3770881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3771363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3771850Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3772356Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3772866Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3773346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3773827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3774305Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3774827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3775313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3775803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3776292Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3776808Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3777288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3777780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3778264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3778754Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3779231Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3779749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3780252Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3780744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3781223Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3781695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3782180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3782689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3783266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3783758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3784243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3784756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3785235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3785714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3786192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3786675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3787192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3787704Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3788191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3788672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3789154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3789642Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3790153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3790633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3791112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3791591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3792089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3792569Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3793054Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3793537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3794022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3794503Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3795011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 94%] 2023-03-31T05:54:17.3795526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3796006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3796490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3796962Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3797465Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3797933Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3798416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3798899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3799418Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3799995Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3800481Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3800955Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3801431Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3801906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3802416Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3802906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3803506Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3803974Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_512_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3804468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3804954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3805490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3805979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3806478Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3806970Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3807490Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3807980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3808468Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3808961Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3809446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3809964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3810500Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3810989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3811474Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3811960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3812446Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3812962Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3813449Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3813921Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3814407Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3814917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3815400Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3815880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3816371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3816858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3817374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3817887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3818381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3818867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3819358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3819841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3820353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3820838Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3821325Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3821803Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3822343Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3822902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3823396Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3823879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3824369Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3824854Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3825354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3825864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3826348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3826832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3827315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3827824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3828312Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3828796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3829284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3829792Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3830280Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3830764Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3831246Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3831733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3832219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3832725Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3833234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3833719Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3834211Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3834703Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3835193Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3835708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3836182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3836665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3837157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3837675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3838164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3838647Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3839133Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3839614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3840126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3840637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3841128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3841610Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3842093Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3842578Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3843232Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3843717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3844206Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3844688Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3845217Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3845695Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3846182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3846667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3847147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3847660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3848177Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3848661Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3849151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3849632Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3850118Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3850630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3851112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3851589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3852073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3852585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3853069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3853551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3854038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3854524Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3855006Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3855512Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3856020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3856502Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3856989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3857473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3857979Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3858463Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3858954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3859436Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3859951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3860429Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3860916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3861399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3861884Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3862360Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3862948Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3863459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3863953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3864439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3864931Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3865421Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3865939Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3866425Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3866911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3867399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3867920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3868404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3868879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3869366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3869852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3870365Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3870894Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3871384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3871869Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3872351Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3872833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3873354Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3873839Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3874324Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3874817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3875330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3875819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3876306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3876788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3877275Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3877760Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3878268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3878782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3879265Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3879736Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3880218Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3880755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3881234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3881726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3882215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3882734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3883346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3883837Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3884317Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3884806Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3885289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3885815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3886327Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3886816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3887289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3887780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3888299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3888783Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3889267Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3889756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3890270Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3890749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3891219Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3891706Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3892191Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3892675Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3893184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3893696Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3894186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3894673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3895159Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3895645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 95%] 2023-03-31T05:54:17.3896166Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3896646Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3897128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3897613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3898132Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3898617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3899100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3899587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3900069Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3900586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3901086Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3901571Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3902037Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3902521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3903074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3903608Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3904089Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3904572Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3905056Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3905575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3906054Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3906541Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3907024Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3907509Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3907990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3908510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3909011Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3909494Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3909978Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3910461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3910972Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3911458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3911944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3912434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3912928Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3913417Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3913899Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3914383Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3914863Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3915346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3915852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3916355Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3916841Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3917328Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3917815Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3918304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3918814Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3919298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3919780Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3920266Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3920775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3921260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3921739Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3922226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3922700Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3923393Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3923917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3924408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3924889Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3925375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3925855Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3926381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3926865Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3927348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3927827Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3928346Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3928826Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3929314Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3929796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3930284Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3930763Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3931268Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3931781Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3932264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3932744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3933222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3933732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3934198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3934673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3935157Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3935667Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3936153Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3936633Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3937114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3937588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3938065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3938565Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3939086Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3939564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3940038Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3940510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3940991Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3941497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3941984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3942462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3943022Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3943540Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3944016Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3944493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3944960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3945440Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3945943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3946439Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3946922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3947395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3947879Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3948353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3948858Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3949330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3949810Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3950283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3950799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3951273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3951755Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3952236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3952732Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3953216Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3953733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3954243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3954735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3955222Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3955697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3956183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3956693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3957176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3957662Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3958171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3958655Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3959138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3959626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3960115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3960596Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3961103Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3961617Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3962102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3962583Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3963186Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3963673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3964198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3964687Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3965171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3965657Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3966183Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3966651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3967131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3967618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3968100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3968620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3969135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3969618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3970096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3970585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3971064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3971575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3972058Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3972545Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3973025Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3973537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3974020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3974505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3974986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3975473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3975954Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3976477Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3976989Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3977462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3977944Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3978428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3978937Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3979426Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3979908Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3980395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3980909Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3981384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3981860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3982349Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3982916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3983406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3983920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3984432Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3984917Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3985404Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3985891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3986377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3986891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3987379Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3987862Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3988336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3988962Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3989455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3989936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3990422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3990907Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3991428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3991934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3992420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3992896Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3993373Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3993857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3994374Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3994853Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3995342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 96%] 2023-03-31T05:54:17.3995822Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3996337Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3996813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3997299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3997782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3998269Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3998730Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3999257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.3999751Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4000242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4000726Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4001212Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4001717Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4002205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4002681Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4003299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4003842Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4004319Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4004796Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4005273Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4005746Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4006230Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4006743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4007262Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4007748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4008229Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4008704Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4009189Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4009679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4010164Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4010641Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4011128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4011637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4012128Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4012615Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4013108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4013595Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4014107Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4014609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4015096Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4015580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4016068Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4016552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4017065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4017539Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4018026Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4018508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4019021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4019505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4019982Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4020444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4020923Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4021394Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4021902Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4022399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4022960Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4023435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4023918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4024401Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4024918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4025399Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4025876Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4026353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4026859Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4027342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4027824Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4028306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4028782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4029283Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4029800Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4030286Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4030758Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4031241Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4031722Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4032225Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4032703Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4033181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4033660Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4034167Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4034644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4035122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4035605Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4036082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4036557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4037059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4037558Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4038030Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4038502Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4038976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4039476Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4039951Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4040434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4040914Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_64_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4041420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4041906Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4042397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4042881Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4043515Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4044001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4044533Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4045048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4045537Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4046020Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4046510Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4046992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4047513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4048000Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4048493Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4048975Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4049495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4049973Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4050459Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4050943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4051427Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4051953Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4052458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4052943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4053428Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4053916Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4054402Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4054912Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4055395Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4055880Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4056364Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4056911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4057390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4057874Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4058363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4058844Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4059329Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4059840Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4060347Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4060831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4061320Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4061799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4062315Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4062797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4063348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4063833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4064348Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4064825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4065310Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4065795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4066288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4066768Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4067282Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4067788Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4068278Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4068756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4069243Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4069727Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4070234Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4070711Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4071205Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4071692Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4072215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4072701Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4073184Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4073671Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4074145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4074654Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4075162Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4075651Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4076135Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4076614Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4077105Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4077616Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4078103Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4078586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4079073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4079580Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4080063Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4080546Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4081029Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4081512Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4081990Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4082492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4083138Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4083622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4084108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4084588Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4085104Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4085585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4086070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4086555Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4087082Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4087561Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4088044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4088522Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4089007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4089487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4090111Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4090627Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4091117Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4091592Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4092078Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4092562Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4093081Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4093559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4094043Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4094526Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4095050Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 97%] 2023-03-31T05:54:17.4095534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4096008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4096492Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4096969Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4097475Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4097986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4098470Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4098947Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4099432Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4099911Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4100420Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_128_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4101074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4101575Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4102065Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4102585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4103178Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4103665Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4104151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4104637Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4105120Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4105634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4106145Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4106626Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4107102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4107589Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4108106Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4108587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4109074Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4109552Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4110077Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4110559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4111044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4111521Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4112008Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4112483Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4112997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4113504Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4113992Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4114473Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4114957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4115435Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4115976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4116457Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4116943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4117455Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4117927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4118406Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4118893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4119384Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4119871Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4120381Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4120893Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4121371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4121857Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4122334Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4122817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4123538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4124028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4124507Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4124997Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4125517Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4126001Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4126485Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4126968Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4127454Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4127976Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4128495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4128967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4129444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4129922Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4130398Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4130920Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4131405Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4131890Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4132372Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4132887Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4133375Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4133864Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4134342Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4134825Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4135306Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4135819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4136333Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4136819Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4137300Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4137784Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4138296Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4138782Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4139253Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4139734Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4140242Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4140720Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4141198Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4141672Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4142147Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4142635Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4143238Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4143753Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4144236Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4144716Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4145201Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4145684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4146196Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4146673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4147146Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4147620Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4148126Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4148611Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4149092Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4149574Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4150044Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4150557Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4151064Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4151551Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4152028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4152508Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4152983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4153488Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4153964Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4154453Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4154940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4155451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4155934Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4156414Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4156891Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4157371Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4157846Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4158350Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4158851Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4159336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4159817Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_256_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4160303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4160776Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4161288Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4161767Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4162249Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4162728Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4163389Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4163870Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4164353Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4164833Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4165311Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4165828Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4166338Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4166816Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4167289Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4167766Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4168244Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4168748Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4169226Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4169702Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4170180Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4170684Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4171151Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4171635Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4172115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4172597Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4173073Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4173585Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4174112Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4174586Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4175060Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4175530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4176028Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4176502Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4176986Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4177469Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4177980Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4178452Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4178937Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4179412Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4179895Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4180370Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4180873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4181366Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4181831Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4182303Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4182786Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4183345Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4183860Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4184336Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4184813Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4185287Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4185797Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4186274Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4186750Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4187224Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4187708Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4188181Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4188682Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4189187Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4189680Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4190161Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4190649Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4191156Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4191644Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4192118Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4192609Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4193115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4193594Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4194079Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4194564Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 98%] 2023-03-31T05:54:17.4195048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4195534Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4196049Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4196559Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4197042Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4197527Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4198009Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4198497Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4198999Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4199479Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4199957Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4200434Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4200936Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4201419Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4201904Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4202397Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4202867Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4203513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4204036Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4204529Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4205010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4205495Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4205971Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4206482Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4206958Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4207445Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4207930Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4208458Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4208938Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4209423Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4209905Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4210392Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4210873Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4211377Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4211877Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4212358Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4212830Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4213316Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4213818Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4214304Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4214795Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4215276Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4215799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4216279Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4216756Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4217244Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4217724Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4218199Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4218697Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_512_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4219207Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4219694Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4220175Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4220656Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4221141Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4221645Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4222131Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4222622Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4223182Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4223693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4224171Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4224638Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4225125Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4225613Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4226122Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4226630Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4227115Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4227591Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4228070Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4228547Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4229055Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4229532Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4230010Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4230487Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4230994Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4231482Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4231967Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4232451Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4232935Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4233408Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4233940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4234444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4234918Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4235390Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4235869Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4236344Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4236852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4237335Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4237820Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4238299Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4238799Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4239277Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4239749Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4240227Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4240705Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4241204Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4241710Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4242189Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4242673Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4243264Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4243744Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4244261Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4244743Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4245216Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4245679Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4246192Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4246664Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4247137Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4247618Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4248100Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_64_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4248587Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4249108Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4249629Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4250116Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4250593Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4251066Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4251573Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4252048Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4252530Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4253007Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4253512Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4253983Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4254462Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4254943Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4255422Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4255885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4256391Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4256885Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4257363Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4257832Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4258313Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4258787Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4259298Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4259775Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4260260Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4260747Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4261257Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4261735Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4262215Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4262689Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4263239Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4263714Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4264235Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4264733Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4265220Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4265693Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_32_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4266176Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4266669Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4267154Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4267634Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4268114Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4268623Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4269102Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4269582Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4270059Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4270538Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4271014Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4271513Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_64_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4272021Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4272505Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4272984Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4273461Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4273940Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4274444Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4274927Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4275403Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4275884Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4276387Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4276852Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_None SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4277330Z test_transformers.py::TestSDPA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1 SKIPPED (Does not support SDPA or pre-SM80 hardware) [ 99%] 2023-03-31T05:54:17.4277551Z test_transformers.py::TestSDPA::test_memory_efficeint_sm86_failure SKIPPED (Does not support fused SDPA or not SM86 hardware) [ 99%] 2023-03-31T05:54:17.4277865Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4278178Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel_SDPBackend_FLASH_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4278515Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel_SDPBackend_EFFICIENT_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4278843Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel_SDPBackend_FLASH_ATTENTION SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4279116Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4279384Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4279655Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4279926Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4280209Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_False_head_dims_match_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4280517Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_False_head_dims_match_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4280800Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_True_head_dims_match_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4281076Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_dense_is_contiguous_True_head_dims_match_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4281347Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_False_head_dims_match_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4281629Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_False_head_dims_match_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4281936Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_True_head_dims_match_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4282215Z test_transformers.py::TestSDPA::test_scaled_dot_product_attention_fused_kernels_type_nested_is_contiguous_True_head_dims_match_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4282442Z test_transformers.py::TestSDPA::test_sdp_choice_with_determinism_warn_only_False SKIPPED (Platform does not support fused SDPA) [ 99%] 2023-03-31T05:54:17.4282671Z test_transformers.py::TestSDPA::test_sdp_choice_with_determinism_warn_only_True SKIPPED (Platform does not support fused SDPA) [ 99%] 2023-03-31T05:54:17.4282950Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_bfloat16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4283442Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_float16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4283716Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_bfloat16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4283990Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_float16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4284307Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_bfloat16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4284611Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_float16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4284887Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_bfloat16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4285144Z test_transformers.py::TestSDPA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_float16 SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4285371Z test_transformers.py::TestSDPA::test_sdp_math_gradcheck_contiguous_inputs_False SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4285600Z test_transformers.py::TestSDPA::test_sdp_math_gradcheck_contiguous_inputs_True SKIPPED (Fused SDPA was not built for this system) [ 99%] 2023-03-31T05:54:17.4285866Z test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_False SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4286164Z test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_True SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4286431Z test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_False SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4286695Z test_transformers.py::TestSDPA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_True SKIPPED (Flash Attention was not built for this system) [ 99%] 2023-03-31T05:54:17.4287011Z test_transformers.py::TestSDPA::test_unaligned_tensors SKIPPED (Does not support fused SDPA or pre-SM80 hardware) [100%] 2023-03-31T05:54:17.4287025Z 2023-03-31T05:54:17.4287385Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_transformers/test_transformers-38a31af97a106b47.xml - 2023-03-31T05:54:17.4287510Z ====================== 70 passed, 20153 skipped in 28.61s ====================== 2023-03-31T05:54:17.4287752Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:54:17.4287762Z 2023-03-31T05:54:17.4288139Z ##[endgroup] 2023-03-31T05:54:17.4288464Z FINISHED PRINTING LOG FILE of test_transformers (/var/lib/jenkins/workspace/test/test-reports/test_transformers__3uqokqt.log) 2023-03-31T05:54:17.4288471Z 2023-03-31T05:54:17.7023988Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:54:17.7051228Z Ignoring disabled issues: [] 2023-03-31T05:54:17.7215430Z Running test_stateless ... [2023-03-31 05:54:17.721178] 2023-03-31T05:54:17.7217859Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_stateless.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:54:17.721545] 2023-03-31T05:54:26.2605288Z 2023-03-31T05:54:26.2605945Z Expand the folded group to see the log file of test_stateless 2023-03-31T05:54:26.2607084Z ##[group]PRINTING LOG FILE of test_stateless (/var/lib/jenkins/workspace/test/test-reports/test_stateless_ungh7mr3.log) 2023-03-31T05:54:26.2608248Z Test results will be stored in test-reports/python-pytest/test_stateless/test_stateless-bb1d952c6ff344c3.xml 2023-03-31T05:54:26.2608765Z ============================= test session starts ============================== 2023-03-31T05:54:26.2609163Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:54:26.2609442Z cachedir: .pytest_cache 2023-03-31T05:54:26.2610072Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:54:26.2610507Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:54:26.2610954Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:54:26.2611513Z collecting ... collected 48 items 2023-03-31T05:54:26.2618386Z Running 48 items in this shard: test/test_stateless.py::TestStatelessFunctionalAPI::test_circular_references_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_circular_references_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_batch_norm_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_batch_norm_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_member_reference_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_member_reference_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_multiple_dicts_error, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_tuple_dicts, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_data_parallel_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_data_parallel_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_gradient_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_gradient_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_jit_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_jit_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_kwargs_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_kwargs_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_in_place_operator_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_in_place_operator_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_module_fail_reset_to_original_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_module_fail_reset_to_original_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_some_weights_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_some_weights_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_special_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_special_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_strict_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_strict_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_some_weights_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_some_weights_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_strict_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_strict_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrized_module_change_parametrization_original_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_reparametrized_module_change_parametrization_original_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_setattr_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_setattr_strict_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_setattr_strict_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_setattr_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_errors_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_errors_torch_func, test/test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_no_error_without_flag, test/test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_warns_stateless, test/test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_warns_torch_func, test/test_stateless.py::TestStatelessDeprecation::test_private_stateless_warns, test/test_stateless.py::TestStatelessDeprecation::test_stateless_functional_call_warns, test/test_stateless.py::TestPythonOptimizeMode::test_runs_with_optimize_flag 2023-03-31T05:54:26.2624859Z 2023-03-31T05:54:26.2625054Z test_stateless.py::TestStatelessFunctionalAPI::test_circular_references_stateless PASSED [ 2%] 2023-03-31T05:54:26.2625545Z test_stateless.py::TestStatelessFunctionalAPI::test_circular_references_torch_func PASSED [ 4%] 2023-03-31T05:54:26.2625952Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_batch_norm_stateless PASSED [ 6%] 2023-03-31T05:54:26.2626436Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_batch_norm_torch_func PASSED [ 8%] 2023-03-31T05:54:26.2626850Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_member_reference_stateless PASSED [ 10%] 2023-03-31T05:54:26.2627417Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_member_reference_torch_func PASSED [ 12%] 2023-03-31T05:54:26.2627925Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_multiple_dicts_error PASSED [ 14%] 2023-03-31T05:54:26.2628335Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_stateless PASSED [ 16%] 2023-03-31T05:54:26.2628784Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_torch_func PASSED [ 18%] 2023-03-31T05:54:26.2629185Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_tuple_dicts PASSED [ 20%] 2023-03-31T05:54:26.2629827Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_data_parallel_stateless SKIPPED (multi-GPU not supported) [ 22%] 2023-03-31T05:54:26.2630518Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_data_parallel_torch_func SKIPPED (multi-GPU not supported) [ 25%] 2023-03-31T05:54:26.2631090Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_gradient_stateless PASSED [ 27%] 2023-03-31T05:54:26.2631519Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_gradient_torch_func PASSED [ 29%] 2023-03-31T05:54:26.2632003Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_jit_stateless PASSED [ 31%] 2023-03-31T05:54:26.2632401Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_jit_torch_func PASSED [ 33%] 2023-03-31T05:54:26.2632903Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_kwargs_stateless PASSED [ 35%] 2023-03-31T05:54:26.2633372Z test_stateless.py::TestStatelessFunctionalAPI::test_functional_call_with_kwargs_torch_func PASSED [ 37%] 2023-03-31T05:54:26.2633800Z test_stateless.py::TestStatelessFunctionalAPI::test_in_place_operator_stateless PASSED [ 39%] 2023-03-31T05:54:26.2634275Z test_stateless.py::TestStatelessFunctionalAPI::test_in_place_operator_torch_func PASSED [ 41%] 2023-03-31T05:54:26.2634711Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_module_fail_reset_to_original_stateless PASSED [ 43%] 2023-03-31T05:54:26.2635252Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_module_fail_reset_to_original_torch_func PASSED [ 45%] 2023-03-31T05:54:26.2635742Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_some_weights_stateless PASSED [ 47%] 2023-03-31T05:54:26.2636175Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_some_weights_torch_func PASSED [ 50%] 2023-03-31T05:54:26.2636717Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_special_stateless PASSED [ 52%] 2023-03-31T05:54:26.2637168Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_special_torch_func PASSED [ 54%] 2023-03-31T05:54:26.2637709Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_strict_stateless PASSED [ 56%] 2023-03-31T05:54:26.2638190Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_strict_torch_func PASSED [ 58%] 2023-03-31T05:54:26.2638616Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_some_weights_stateless PASSED [ 60%] 2023-03-31T05:54:26.2639121Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_some_weights_torch_func PASSED [ 62%] 2023-03-31T05:54:26.2639550Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_stateless PASSED [ 64%] 2023-03-31T05:54:26.2640055Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_strict_stateless PASSED [ 66%] 2023-03-31T05:54:26.2640573Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_strict_torch_func PASSED [ 68%] 2023-03-31T05:54:26.2641012Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrize_tie_weights_torch_func PASSED [ 70%] 2023-03-31T05:54:26.2641521Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrized_module_change_parametrization_original_stateless PASSED [ 72%] 2023-03-31T05:54:26.2642065Z test_stateless.py::TestStatelessFunctionalAPI::test_reparametrized_module_change_parametrization_original_torch_func PASSED [ 75%] 2023-03-31T05:54:26.2642591Z test_stateless.py::TestStatelessFunctionalAPI::test_setattr_stateless PASSED [ 77%] 2023-03-31T05:54:26.2642979Z test_stateless.py::TestStatelessFunctionalAPI::test_setattr_strict_stateless PASSED [ 79%] 2023-03-31T05:54:26.2643872Z test_stateless.py::TestStatelessFunctionalAPI::test_setattr_strict_torch_func PASSED [ 81%] 2023-03-31T05:54:26.2644266Z test_stateless.py::TestStatelessFunctionalAPI::test_setattr_torch_func PASSED [ 83%] 2023-03-31T05:54:26.2644658Z test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_errors_stateless PASSED [ 85%] 2023-03-31T05:54:26.2645056Z test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_errors_torch_func PASSED [ 87%] 2023-03-31T05:54:26.2645541Z test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_no_error_without_flag PASSED [ 89%] 2023-03-31T05:54:26.2646018Z test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_warns_stateless PASSED [ 91%] 2023-03-31T05:54:26.2646421Z test_stateless.py::TestStatelessFunctionalAPI::test_tied_weights_warns_torch_func PASSED [ 93%] 2023-03-31T05:54:26.2646794Z test_stateless.py::TestStatelessDeprecation::test_private_stateless_warns PASSED [ 95%] 2023-03-31T05:54:26.2647193Z test_stateless.py::TestStatelessDeprecation::test_stateless_functional_call_warns PASSED [ 97%] 2023-03-31T05:54:26.2647582Z test_stateless.py::TestPythonOptimizeMode::test_runs_with_optimize_flag PASSED [100%] 2023-03-31T05:54:26.2647791Z 2023-03-31T05:54:26.2648188Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_stateless/test_stateless-bb1d952c6ff344c3.xml - 2023-03-31T05:54:26.2648559Z ======================== 46 passed, 2 skipped in 6.34s ========================= 2023-03-31T05:54:26.2648889Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:54:26.2649094Z 2023-03-31T05:54:26.2649409Z ##[endgroup] 2023-03-31T05:54:26.2649890Z FINISHED PRINTING LOG FILE of test_stateless (/var/lib/jenkins/workspace/test/test-reports/test_stateless_ungh7mr3.log) 2023-03-31T05:54:26.2650124Z 2023-03-31T05:54:28.3409378Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:54:28.3444644Z Ignoring disabled issues: [] 2023-03-31T05:54:28.3609393Z Running test_ao_sparsity ... [2023-03-31 05:54:28.360654] 2023-03-31T05:54:28.3612818Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ao_sparsity.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:54:28.361007] 2023-03-31T05:54:36.1844949Z 2023-03-31T05:54:36.1845555Z Expand the folded group to see the log file of test_ao_sparsity 2023-03-31T05:54:36.1846661Z ##[group]PRINTING LOG FILE of test_ao_sparsity (/var/lib/jenkins/workspace/test/test-reports/test_ao_sparsity_t4j9ntwa.log) 2023-03-31T05:54:36.1847694Z Test results will be stored in test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-e05b502ec4c844d1.xml 2023-03-31T05:54:36.1848093Z ============================= test session starts ============================== 2023-03-31T05:54:36.1848555Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:54:36.1848895Z cachedir: .pytest_cache 2023-03-31T05:54:36.1849342Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:54:36.1849777Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:54:36.1850253Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:54:36.1850600Z collecting ... collected 85 items 2023-03-31T05:54:36.1859775Z Running 85 items in this shard: test/test_ao_sparsity.py::TestQuantizedSparseKernels::test_sparse_qlinear, test/test_ao_sparsity.py::TestQuantizedSparseLayers::test_sparse_qlinear, test/test_ao_sparsity.py::TestQuantizedSparseLayers::test_sparse_qlinear_serdes, test/test_ao_sparsity.py::TestFakeSparsity::test_jit_trace, test/test_ao_sparsity.py::TestFakeSparsity::test_masking_logic, test/test_ao_sparsity.py::TestFakeSparsity::test_state_dict_preserved, test/test_ao_sparsity.py::TestFakeSparsity::test_weights_parametrized, test/test_ao_sparsity.py::TestBaseSparsifier::test_constructor, test/test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash, test/test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params1, test/test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params2, test/test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params3, test/test_ao_sparsity.py::TestBaseSparsifier::test_prepare_config, test/test_ao_sparsity.py::TestBaseSparsifier::test_state_dict, test/test_ao_sparsity.py::TestBaseSparsifier::test_step, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_constructor, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_mask_squash, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_prepare, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_sparsity_levels, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_step, test/test_ao_sparsity.py::TestWeightNormSparsifier::test_step_2_of_4, test/test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_constructor, test/test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_mask_squash, test/test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_prepare, test/test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_sparsity_levels, test/test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_step, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_complex_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_constructor, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prepare_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prepare_linear, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_activation_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_bias_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_padding_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_pool_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_activation_linear, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_bias_linear, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_linear, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_layernorm_linear_multiple_layer, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_layernorm_linear_single_layer, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_linear_multiple_layer, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_linear_single_layer, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_step_conv2d, test/test_ao_sparsity.py::TestBaseStructuredSparsifier::test_step_linear, test/test_ao_sparsity.py::TestSaliencyPruner::test_lstm_saliency_pruner_update_mask, test/test_ao_sparsity.py::TestSaliencyPruner::test_saliency_pruner_update_mask, test/test_ao_sparsity.py::TestScheduler::test_constructor, test/test_ao_sparsity.py::TestScheduler::test_lambda_scheduler, test/test_ao_sparsity.py::TestScheduler::test_order_of_steps, test/test_ao_sparsity.py::TestScheduler::test_step, test/test_ao_sparsity.py::TestCubicScheduler::test_constructor, test/test_ao_sparsity.py::TestCubicScheduler::test_step, test/test_ao_sparsity.py::TestComposability::test_convert_without_squash_mask, test/test_ao_sparsity.py::TestComposability::test_fusion_before_s_prep, test/test_ao_sparsity.py::TestComposability::test_q_prep_before_s_prep, test/test_ao_sparsity.py::TestComposability::test_qat_prep_before_s_prep, test/test_ao_sparsity.py::TestComposability::test_s_prep_before_fusion, test/test_ao_sparsity.py::TestComposability::test_s_prep_before_q_prep, test/test_ao_sparsity.py::TestComposability::test_s_prep_before_qat_prep, test/test_ao_sparsity.py::TestFxComposability::test_q_prep_fx_before_s_prep, test/test_ao_sparsity.py::TestFxComposability::test_q_prep_fx_s_prep_ref_conv, test/test_ao_sparsity.py::TestFxComposability::test_s_prep_before_q_prep_fx, test/test_ao_sparsity.py::TestFxComposability::test_s_prep_before_qat_prep_fx, test/test_ao_sparsity.py::TestFxComposability::test_s_prep_q_prep_fx_ref, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module_fail, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module_for_tensors, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_get_arg_info_from_tensor_fqn, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_get_arg_info_from_tensor_fqn_fail, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn_fail, test/test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn_root, test/test_ao_sparsity.py::TestBaseDataSparsifier::test_nn_embeddings, test/test_ao_sparsity.py::TestBaseDataSparsifier::test_nn_parameters, test/test_ao_sparsity.py::TestBaseDataSparsifier::test_tensors, test/test_ao_sparsity.py::TestNormDataSparsifiers::test_nn_embeddings, test/test_ao_sparsity.py::TestNormDataSparsifiers::test_nn_parameters, test/test_ao_sparsity.py::TestNormDataSparsifiers::test_tensors, test/test_ao_sparsity.py::TestQuantizationUtils::test_ptq_quantize_first, test/test_ao_sparsity.py::TestQuantizationUtils::test_ptq_sparsify_first, test/test_ao_sparsity.py::TestBaseDataScheduler::test_constructor, test/test_ao_sparsity.py::TestBaseDataScheduler::test_order_of_steps, test/test_ao_sparsity.py::TestBaseDataScheduler::test_state_dict, test/test_ao_sparsity.py::TestBaseDataScheduler::test_step, test/test_ao_sparsity.py::TestActivationSparsifier::test_activation_sparsifier 2023-03-31T05:54:36.1868271Z 2023-03-31T05:54:36.1868916Z test_ao_sparsity.py::TestQuantizedSparseKernels::test_sparse_qlinear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py 2023-03-31 05:54:30,425 - root - INFO - static sparse qlinear is only available in fbgemm 2023-03-31T05:54:36.1869578Z 2023-03-31 05:54:30,476 - root - INFO - static sparse qlinear is only available in fbgemm 2023-03-31T05:54:36.1870045Z 2023-03-31 05:54:30,481 - root - INFO - dynamic sparse qlinear is only available in qnnpack 2023-03-31T05:54:36.1870572Z 2023-03-31 05:54:30,482 - root - INFO - dynamic sparse qlinear is only available in qnnpack 2023-03-31T05:54:36.1870820Z PASSED [ 1%] 2023-03-31T05:54:36.1871680Z test_ao_sparsity.py::TestQuantizedSparseLayers::test_sparse_qlinear <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py [W qlinear_dynamic.cpp:247] Warning: Currently, qnnpack incorrectly ignores reduce_range when it is set to true; this may change in a future release. (function operator()) 2023-03-31T05:54:36.1872249Z PASSED [ 2%] 2023-03-31T05:54:36.1872866Z test_ao_sparsity.py::TestQuantizedSparseLayers::test_sparse_qlinear_serdes <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_quantized.py PASSED [ 3%] 2023-03-31T05:54:36.1873530Z test_ao_sparsity.py::TestFakeSparsity::test_jit_trace <- test/ao/sparsity/test_parametrization.py PASSED [ 4%] 2023-03-31T05:54:36.1874104Z test_ao_sparsity.py::TestFakeSparsity::test_masking_logic <- test/ao/sparsity/test_parametrization.py PASSED [ 5%] 2023-03-31T05:54:36.1874646Z test_ao_sparsity.py::TestFakeSparsity::test_state_dict_preserved <- test/ao/sparsity/test_parametrization.py PASSED [ 7%] 2023-03-31T05:54:36.1875268Z test_ao_sparsity.py::TestFakeSparsity::test_weights_parametrized <- test/ao/sparsity/test_parametrization.py PASSED [ 8%] 2023-03-31T05:54:36.1875898Z test_ao_sparsity.py::TestBaseSparsifier::test_constructor <- test/ao/sparsity/test_sparsifier.py PASSED [ 9%] 2023-03-31T05:54:36.1876461Z test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash <- test/ao/sparsity/test_sparsifier.py PASSED [ 10%] 2023-03-31T05:54:36.1876986Z test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params1 <- test/ao/sparsity/test_sparsifier.py PASSED [ 11%] 2023-03-31T05:54:36.1877609Z test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params2 <- test/ao/sparsity/test_sparsifier.py PASSED [ 12%] 2023-03-31T05:54:36.1878213Z test_ao_sparsity.py::TestBaseSparsifier::test_mask_squash_with_params3 <- test/ao/sparsity/test_sparsifier.py PASSED [ 14%] 2023-03-31T05:54:36.1878783Z test_ao_sparsity.py::TestBaseSparsifier::test_prepare_config <- test/ao/sparsity/test_sparsifier.py PASSED [ 15%] 2023-03-31T05:54:36.1879325Z test_ao_sparsity.py::TestBaseSparsifier::test_state_dict <- test/ao/sparsity/test_sparsifier.py PASSED [ 16%] 2023-03-31T05:54:36.1879897Z test_ao_sparsity.py::TestBaseSparsifier::test_step <- test/ao/sparsity/test_sparsifier.py PASSED [ 17%] 2023-03-31T05:54:36.1892531Z test_ao_sparsity.py::TestWeightNormSparsifier::test_constructor <- test/ao/sparsity/test_sparsifier.py PASSED [ 18%] 2023-03-31T05:54:36.1893280Z test_ao_sparsity.py::TestWeightNormSparsifier::test_mask_squash <- test/ao/sparsity/test_sparsifier.py PASSED [ 20%] 2023-03-31T05:54:36.1893816Z test_ao_sparsity.py::TestWeightNormSparsifier::test_prepare <- test/ao/sparsity/test_sparsifier.py PASSED [ 21%] 2023-03-31T05:54:36.1894382Z test_ao_sparsity.py::TestWeightNormSparsifier::test_sparsity_levels <- test/ao/sparsity/test_sparsifier.py PASSED [ 22%] 2023-03-31T05:54:36.1894915Z test_ao_sparsity.py::TestWeightNormSparsifier::test_step <- test/ao/sparsity/test_sparsifier.py PASSED [ 23%] 2023-03-31T05:54:36.1895474Z test_ao_sparsity.py::TestWeightNormSparsifier::test_step_2_of_4 <- test/ao/sparsity/test_sparsifier.py PASSED [ 24%] 2023-03-31T05:54:36.1896011Z test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_constructor <- test/ao/sparsity/test_sparsifier.py PASSED [ 25%] 2023-03-31T05:54:36.1896571Z test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_mask_squash <- test/ao/sparsity/test_sparsifier.py PASSED [ 27%] 2023-03-31T05:54:36.1897109Z test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_prepare <- test/ao/sparsity/test_sparsifier.py PASSED [ 28%] 2023-03-31T05:54:36.1897810Z test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_sparsity_levels <- test/ao/sparsity/test_sparsifier.py PASSED [ 29%] 2023-03-31T05:54:36.1898390Z test_ao_sparsity.py::TestNearlyDiagonalSparsifier::test_step <- test/ao/sparsity/test_sparsifier.py PASSED [ 30%] 2023-03-31T05:54:36.1898966Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_complex_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 31%] 2023-03-31T05:54:36.1899562Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_constructor <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 32%] 2023-03-31T05:54:36.1900162Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prepare_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 34%] 2023-03-31T05:54:36.1900743Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prepare_linear <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 35%] 2023-03-31T05:54:36.1901363Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_activation_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 36%] 2023-03-31T05:54:36.1901989Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_bias_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 37%] 2023-03-31T05:54:36.1902609Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 38%] 2023-03-31T05:54:36.1903337Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_padding_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 40%] 2023-03-31T05:54:36.1904018Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_conv2d_pool_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 41%] 2023-03-31T05:54:36.1904649Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_activation_linear <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 42%] 2023-03-31T05:54:36.1905271Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_bias_linear <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 43%] 2023-03-31T05:54:36.1905872Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_linear_linear <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 44%] 2023-03-31T05:54:36.1906522Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_layernorm_linear_multiple_layer <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 45%] 2023-03-31T05:54:36.1907270Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_layernorm_linear_single_layer <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 47%] 2023-03-31T05:54:36.1907917Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_linear_multiple_layer <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 48%] 2023-03-31T05:54:36.1908561Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_prune_lstm_linear_single_layer <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 49%] 2023-03-31T05:54:36.1909150Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_step_conv2d <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 50%] 2023-03-31T05:54:36.1909731Z test_ao_sparsity.py::TestBaseStructuredSparsifier::test_step_linear <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 51%] 2023-03-31T05:54:36.1910331Z test_ao_sparsity.py::TestSaliencyPruner::test_lstm_saliency_pruner_update_mask <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 52%] 2023-03-31T05:54:36.1910924Z test_ao_sparsity.py::TestSaliencyPruner::test_saliency_pruner_update_mask <- test/ao/sparsity/test_structured_sparsifier.py PASSED [ 54%] 2023-03-31T05:54:36.1911436Z test_ao_sparsity.py::TestScheduler::test_constructor <- test/ao/sparsity/test_scheduler.py PASSED [ 55%] 2023-03-31T05:54:36.1911925Z test_ao_sparsity.py::TestScheduler::test_lambda_scheduler <- test/ao/sparsity/test_scheduler.py PASSED [ 56%] 2023-03-31T05:54:36.1912419Z test_ao_sparsity.py::TestScheduler::test_order_of_steps <- test/ao/sparsity/test_scheduler.py PASSED [ 57%] 2023-03-31T05:54:36.1912920Z test_ao_sparsity.py::TestScheduler::test_step <- test/ao/sparsity/test_scheduler.py PASSED [ 58%] 2023-03-31T05:54:36.1913419Z test_ao_sparsity.py::TestCubicScheduler::test_constructor <- test/ao/sparsity/test_scheduler.py PASSED [ 60%] 2023-03-31T05:54:36.1913912Z test_ao_sparsity.py::TestCubicScheduler::test_step <- test/ao/sparsity/test_scheduler.py PASSED [ 61%] 2023-03-31T05:54:36.1914439Z test_ao_sparsity.py::TestComposability::test_convert_without_squash_mask <- test/ao/sparsity/test_composability.py PASSED [ 62%] 2023-03-31T05:54:36.1914969Z test_ao_sparsity.py::TestComposability::test_fusion_before_s_prep <- test/ao/sparsity/test_composability.py PASSED [ 63%] 2023-03-31T05:54:36.1915496Z test_ao_sparsity.py::TestComposability::test_q_prep_before_s_prep <- test/ao/sparsity/test_composability.py PASSED [ 64%] 2023-03-31T05:54:36.1916029Z test_ao_sparsity.py::TestComposability::test_qat_prep_before_s_prep <- test/ao/sparsity/test_composability.py PASSED [ 65%] 2023-03-31T05:54:36.1916564Z test_ao_sparsity.py::TestComposability::test_s_prep_before_fusion <- test/ao/sparsity/test_composability.py PASSED [ 67%] 2023-03-31T05:54:36.1917075Z test_ao_sparsity.py::TestComposability::test_s_prep_before_q_prep <- test/ao/sparsity/test_composability.py PASSED [ 68%] 2023-03-31T05:54:36.1917606Z test_ao_sparsity.py::TestComposability::test_s_prep_before_qat_prep <- test/ao/sparsity/test_composability.py PASSED [ 69%] 2023-03-31T05:54:36.1918180Z test_ao_sparsity.py::TestFxComposability::test_q_prep_fx_before_s_prep <- test/ao/sparsity/test_composability.py PASSED [ 70%] 2023-03-31T05:54:36.1918734Z test_ao_sparsity.py::TestFxComposability::test_q_prep_fx_s_prep_ref_conv <- test/ao/sparsity/test_composability.py PASSED [ 71%] 2023-03-31T05:54:36.1919272Z test_ao_sparsity.py::TestFxComposability::test_s_prep_before_q_prep_fx <- test/ao/sparsity/test_composability.py PASSED [ 72%] 2023-03-31T05:54:36.1919823Z test_ao_sparsity.py::TestFxComposability::test_s_prep_before_qat_prep_fx <- test/ao/sparsity/test_composability.py PASSED [ 74%] 2023-03-31T05:54:36.1920363Z test_ao_sparsity.py::TestFxComposability::test_s_prep_q_prep_fx_ref <- test/ao/sparsity/test_composability.py PASSED [ 75%] 2023-03-31T05:54:36.1920910Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 76%] 2023-03-31T05:54:36.1921485Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module_fail <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 77%] 2023-03-31T05:54:36.1922061Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_fqn_to_module_for_tensors <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 78%] 2023-03-31T05:54:36.1922649Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_get_arg_info_from_tensor_fqn <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 80%] 2023-03-31T05:54:36.1923434Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_get_arg_info_from_tensor_fqn_fail <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 81%] 2023-03-31T05:54:36.1924009Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 82%] 2023-03-31T05:54:36.1924559Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn_fail <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 83%] 2023-03-31T05:54:36.1925112Z test_ao_sparsity.py::TestSparsityUtilFunctions::test_module_to_fqn_root <- test/ao/sparsity/test_sparsity_utils.py PASSED [ 84%] 2023-03-31T05:54:36.1925659Z test_ao_sparsity.py::TestBaseDataSparsifier::test_nn_embeddings <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 85%] 2023-03-31T05:54:36.1926190Z test_ao_sparsity.py::TestBaseDataSparsifier::test_nn_parameters <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 87%] 2023-03-31T05:54:36.1926718Z test_ao_sparsity.py::TestBaseDataSparsifier::test_tensors <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 88%] 2023-03-31T05:54:36.1927253Z test_ao_sparsity.py::TestNormDataSparsifiers::test_nn_embeddings <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 89%] 2023-03-31T05:54:36.1927867Z test_ao_sparsity.py::TestNormDataSparsifiers::test_nn_parameters <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 90%] 2023-03-31T05:54:36.1928421Z test_ao_sparsity.py::TestNormDataSparsifiers::test_tensors <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 91%] 2023-03-31T05:54:36.1928951Z test_ao_sparsity.py::TestQuantizationUtils::test_ptq_quantize_first <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 92%] 2023-03-31T05:54:36.1929499Z test_ao_sparsity.py::TestQuantizationUtils::test_ptq_sparsify_first <- test/ao/sparsity/test_data_sparsifier.py PASSED [ 94%] 2023-03-31T05:54:36.1930031Z test_ao_sparsity.py::TestBaseDataScheduler::test_constructor <- test/ao/sparsity/test_data_scheduler.py PASSED [ 95%] 2023-03-31T05:54:36.1930548Z test_ao_sparsity.py::TestBaseDataScheduler::test_order_of_steps <- test/ao/sparsity/test_data_scheduler.py PASSED [ 96%] 2023-03-31T05:54:36.1931075Z test_ao_sparsity.py::TestBaseDataScheduler::test_state_dict <- test/ao/sparsity/test_data_scheduler.py PASSED [ 97%] 2023-03-31T05:54:36.1931583Z test_ao_sparsity.py::TestBaseDataScheduler::test_step <- test/ao/sparsity/test_data_scheduler.py PASSED [ 98%] 2023-03-31T05:54:36.1932144Z test_ao_sparsity.py::TestActivationSparsifier::test_activation_sparsifier <- test/ao/sparsity/test_activation_sparsifier.py PASSED [100%] 2023-03-31T05:54:36.1932405Z 2023-03-31T05:54:36.1932737Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-e05b502ec4c844d1.xml - 2023-03-31T05:54:36.1933167Z ============================== 85 passed in 5.59s ============================== 2023-03-31T05:54:36.1933486Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:54:36.1933690Z 2023-03-31T05:54:36.1933992Z ##[endgroup] 2023-03-31T05:54:36.1934409Z FINISHED PRINTING LOG FILE of test_ao_sparsity (/var/lib/jenkins/workspace/test/test-reports/test_ao_sparsity_t4j9ntwa.log) 2023-03-31T05:54:36.1934640Z 2023-03-31T05:54:38.2717709Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:54:38.2744919Z Ignoring disabled issues: [] 2023-03-31T05:54:38.2909649Z Running nn/test_embedding ... [2023-03-31 05:54:38.290637] 2023-03-31T05:54:38.2911888Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_embedding.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:54:38.290957] 2023-03-31T05:54:46.1389924Z 2023-03-31T05:54:46.1391805Z Expand the folded group to see the log file of nn/test_embedding 2023-03-31T05:54:46.1392866Z ##[group]PRINTING LOG FILE of nn/test_embedding (/var/lib/jenkins/workspace/test/test-reports/nn-test_embedding_e49oayjw.log) 2023-03-31T05:54:46.1393911Z Test results will be stored in test-reports/python-pytest/nn.test_embedding/nn.test_embedding-970b421025a9958a.xml 2023-03-31T05:54:46.1394492Z ============================= test session starts ============================== 2023-03-31T05:54:46.1395162Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:54:46.1395455Z cachedir: .pytest_cache 2023-03-31T05:54:46.1395880Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:54:46.1396309Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:54:46.1396853Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:54:46.1397212Z collecting ... collected 160 items 2023-03-31T05:54:46.1422142Z Running 160 items in this shard: test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_from_pretrained, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_from_pretrained_padding_idx, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_functional, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_padding_idx_error, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_float32, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_float64, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int16, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int32, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int64, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int8, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_options, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_padding_idx, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_uint8, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_functional, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_max_norm, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_max_norm_unsorted_repeating_indices, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_sparse_basic, test/nn/test_embedding.py::TestEmbeddingNN::test_embedding_sparse_empty_tensor, test/nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_from_pretrained, test/nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_from_pretrained_options, test/nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_include_last_offset, test/nn/test_embedding.py::TestEmbeddingNN::test_move_sparse_half_embedding, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_backward_cpu_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_1D_padding_idx_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_1D_padding_idx_cpu_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_2D_padding_idx_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_2D_padding_idx_cpu_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_dimension_errors_cpu, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_bfloat16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float32_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float32_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float64_int32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float64_int64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_dense_grad_cpu, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_backward_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float16, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float64, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_fwd_AD_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_padding_idx_cpu_float32, test/nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_scalar_weight_error_cpu 2023-03-31T05:54:46.1445750Z 2023-03-31T05:54:46.1445935Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_from_pretrained PASSED [ 0%] 2023-03-31T05:54:46.1446332Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_from_pretrained_padding_idx PASSED [ 1%] 2023-03-31T05:54:46.1446793Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_functional PASSED [ 1%] 2023-03-31T05:54:46.1447159Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_bag_padding_idx_error PASSED [ 2%] 2023-03-31T05:54:46.1447518Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_float32 PASSED [ 3%] 2023-03-31T05:54:46.1447985Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_float64 PASSED [ 3%] 2023-03-31T05:54:46.1448389Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int16 PASSED [ 4%] 2023-03-31T05:54:46.1448754Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int32 PASSED [ 5%] 2023-03-31T05:54:46.1449107Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int64 PASSED [ 5%] 2023-03-31T05:54:46.1449465Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_int8 PASSED [ 6%] 2023-03-31T05:54:46.1449836Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_options PASSED [ 6%] 2023-03-31T05:54:46.1450200Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_padding_idx PASSED [ 7%] 2023-03-31T05:54:46.1450574Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_from_pretrained_uint8 PASSED [ 8%] 2023-03-31T05:54:46.1450924Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_functional PASSED [ 8%] 2023-03-31T05:54:46.1451265Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_max_norm PASSED [ 9%] 2023-03-31T05:54:46.1451704Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_max_norm_unsorted_repeating_indices SKIPPED (CUDA unavailable) [ 10%] 2023-03-31T05:54:46.1452133Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_sparse_basic PASSED [ 10%] 2023-03-31T05:54:46.1452485Z nn/test_embedding.py::TestEmbeddingNN::test_embedding_sparse_empty_tensor PASSED [ 11%] 2023-03-31T05:54:46.1452837Z nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_from_pretrained PASSED [ 11%] 2023-03-31T05:54:46.1453208Z nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_from_pretrained_options PASSED [ 12%] 2023-03-31T05:54:46.1453579Z nn/test_embedding.py::TestEmbeddingNN::test_embeddingbag_include_last_offset PASSED [ 13%] 2023-03-31T05:54:46.1453937Z nn/test_embedding.py::TestEmbeddingNN::test_move_sparse_half_embedding PASSED [ 13%] 2023-03-31T05:54:46.1454363Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_bfloat16 PASSED [ 14%] 2023-03-31T05:54:46.1454880Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float16 PASSED [ 15%] 2023-03-31T05:54:46.1455389Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float32 PASSED [ 15%] 2023-03-31T05:54:46.1455885Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int32_float64 PASSED [ 16%] 2023-03-31T05:54:46.1456427Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_bfloat16 PASSED [ 16%] 2023-03-31T05:54:46.1457029Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float16 PASSED [ 17%] 2023-03-31T05:54:46.1457574Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float32 PASSED [ 18%] 2023-03-31T05:54:46.1458076Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int32_int64_float64 PASSED [ 18%] 2023-03-31T05:54:46.1458571Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_bfloat16 PASSED [ 19%] 2023-03-31T05:54:46.1459066Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float16 PASSED [ 20%] 2023-03-31T05:54:46.1459602Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float32 PASSED [ 20%] 2023-03-31T05:54:46.1460186Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int32_float64 PASSED [ 21%] 2023-03-31T05:54:46.1460722Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_bfloat16 PASSED [ 21%] 2023-03-31T05:54:46.1461221Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float16 PASSED [ 22%] 2023-03-31T05:54:46.1461709Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float32 PASSED [ 23%] 2023-03-31T05:54:46.1462200Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_empty_per_sample_weights_and_offsets_cpu_int64_int64_float64 PASSED [ 23%] 2023-03-31T05:54:46.1462700Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_bfloat16 PASSED [ 24%] 2023-03-31T05:54:46.1463274Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float16 PASSED [ 25%] 2023-03-31T05:54:46.1463768Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float32 PASSED [ 25%] 2023-03-31T05:54:46.1464305Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int32_float64 PASSED [ 26%] 2023-03-31T05:54:46.1464832Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_bfloat16 PASSED [ 26%] 2023-03-31T05:54:46.1465313Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float16 PASSED [ 27%] 2023-03-31T05:54:46.1465807Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float32 PASSED [ 28%] 2023-03-31T05:54:46.1466298Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int32_int64_float64 PASSED [ 28%] 2023-03-31T05:54:46.1466793Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_bfloat16 PASSED [ 29%] 2023-03-31T05:54:46.1467280Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float16 PASSED [ 30%] 2023-03-31T05:54:46.1467770Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float32 PASSED [ 30%] 2023-03-31T05:54:46.1468316Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int32_float64 PASSED [ 31%] 2023-03-31T05:54:46.1468911Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_bfloat16 PASSED [ 31%] 2023-03-31T05:54:46.1469396Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float16 PASSED [ 32%] 2023-03-31T05:54:46.1469879Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float32 PASSED [ 33%] 2023-03-31T05:54:46.1470365Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_new_offsets_cpu_int64_int64_float64 PASSED [ 33%] 2023-03-31T05:54:46.1470851Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int32_float32 PASSED [ 34%] 2023-03-31T05:54:46.1471371Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int32_float64 PASSED [ 35%] 2023-03-31T05:54:46.1471851Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int64_float32 PASSED [ 35%] 2023-03-31T05:54:46.1472331Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_no_offsets_cpu_int64_float64 PASSED [ 36%] 2023-03-31T05:54:46.1472816Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_bfloat16 PASSED [ 36%] 2023-03-31T05:54:46.1473293Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float16 PASSED [ 37%] 2023-03-31T05:54:46.1473812Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float32 PASSED [ 38%] 2023-03-31T05:54:46.1474387Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int32_float64 PASSED [ 38%] 2023-03-31T05:54:46.1474891Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_bfloat16 PASSED [ 39%] 2023-03-31T05:54:46.1475364Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float16 PASSED [ 40%] 2023-03-31T05:54:46.1475843Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float32 PASSED [ 40%] 2023-03-31T05:54:46.1476375Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int32_int64_float64 PASSED [ 41%] 2023-03-31T05:54:46.1476966Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_bfloat16 PASSED [ 41%] 2023-03-31T05:54:46.1477496Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float16 PASSED [ 42%] 2023-03-31T05:54:46.1477968Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float32 PASSED [ 43%] 2023-03-31T05:54:46.1478454Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int32_float64 PASSED [ 43%] 2023-03-31T05:54:46.1478939Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_bfloat16 PASSED [ 44%] 2023-03-31T05:54:46.1479424Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float16 PASSED [ 45%] 2023-03-31T05:54:46.1479899Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float32 PASSED [ 45%] 2023-03-31T05:54:46.1480388Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_and_offsets_cpu_int64_int64_float64 PASSED [ 46%] 2023-03-31T05:54:46.1480909Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int32_int32 PASSED [ 46%] 2023-03-31T05:54:46.1481382Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int32_int64 PASSED [ 47%] 2023-03-31T05:54:46.1481839Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int64_int32 PASSED [ 48%] 2023-03-31T05:54:46.1482300Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_EmbeddingBag_per_sample_weights_failures_cpu_int64_int64 PASSED [ 48%] 2023-03-31T05:54:46.1482746Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_backward_cpu_float64 PASSED [ 49%] 2023-03-31T05:54:46.1483320Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_1D_padding_idx_cpu_float32 PASSED [ 50%] 2023-03-31T05:54:46.1483749Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_1D_padding_idx_cpu_float64 PASSED [ 50%] 2023-03-31T05:54:46.1484256Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_2D_padding_idx_cpu_float32 PASSED [ 51%] 2023-03-31T05:54:46.1484689Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_2D_padding_idx_cpu_float64 PASSED [ 51%] 2023-03-31T05:54:46.1485109Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int32_int32 PASSED [ 52%] 2023-03-31T05:54:46.1485538Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int32_int64 PASSED [ 53%] 2023-03-31T05:54:46.1485969Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int64_int32 PASSED [ 53%] 2023-03-31T05:54:46.1486398Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_bfloat16_cpu_int64_int64 PASSED [ 54%] 2023-03-31T05:54:46.1486909Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_bfloat16 PASSED [ 55%] 2023-03-31T05:54:46.1487416Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float16 PASSED [ 55%] 2023-03-31T05:54:46.1487864Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float32 PASSED [ 56%] 2023-03-31T05:54:46.1488302Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int32_float64 PASSED [ 56%] 2023-03-31T05:54:46.1488726Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_bfloat16 PASSED [ 57%] 2023-03-31T05:54:46.1489220Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float16 PASSED [ 58%] 2023-03-31T05:54:46.1489698Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float32 PASSED [ 58%] 2023-03-31T05:54:46.1490136Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int32_int64_float64 PASSED [ 59%] 2023-03-31T05:54:46.1490566Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_bfloat16 PASSED [ 60%] 2023-03-31T05:54:46.1490999Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float16 PASSED [ 60%] 2023-03-31T05:54:46.1491435Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float32 PASSED [ 61%] 2023-03-31T05:54:46.1491868Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int32_float64 PASSED [ 61%] 2023-03-31T05:54:46.1492291Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_bfloat16 PASSED [ 62%] 2023-03-31T05:54:46.1492731Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float16 PASSED [ 63%] 2023-03-31T05:54:46.1493164Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float32 PASSED [ 63%] 2023-03-31T05:54:46.1493597Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_device_cpu_int64_int64_float64 PASSED [ 64%] 2023-03-31T05:54:46.1494057Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_dimension_errors_cpu PASSED [ 65%] 2023-03-31T05:54:46.1494489Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int32_int32 PASSED [ 65%] 2023-03-31T05:54:46.1494920Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int32_int64 PASSED [ 66%] 2023-03-31T05:54:46.1495340Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int64_int32 PASSED [ 66%] 2023-03-31T05:54:46.1495773Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_empty_input_cpu_int64_int64 PASSED [ 67%] 2023-03-31T05:54:46.1496195Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int32_int32 PASSED [ 68%] 2023-03-31T05:54:46.1496642Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int32_int64 PASSED [ 68%] 2023-03-31T05:54:46.1497045Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int64_int32 PASSED [ 69%] 2023-03-31T05:54:46.1497455Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_half_cpu_int64_int64 PASSED [ 70%] 2023-03-31T05:54:46.1497909Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_bfloat16 PASSED [ 70%] 2023-03-31T05:54:46.1498486Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float16 PASSED [ 71%] 2023-03-31T05:54:46.1498956Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float32 PASSED [ 71%] 2023-03-31T05:54:46.1499474Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int32_float64 PASSED [ 72%] 2023-03-31T05:54:46.1499948Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_bfloat16 PASSED [ 73%] 2023-03-31T05:54:46.1500415Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float16 PASSED [ 73%] 2023-03-31T05:54:46.1500871Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float32 PASSED [ 74%] 2023-03-31T05:54:46.1501342Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int32_int64_float64 PASSED [ 75%] 2023-03-31T05:54:46.1501849Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_bfloat16 PASSED [ 75%] 2023-03-31T05:54:46.1502353Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float16 PASSED [ 76%] 2023-03-31T05:54:46.1502804Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float32 PASSED [ 76%] 2023-03-31T05:54:46.1503366Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int32_float64 PASSED [ 77%] 2023-03-31T05:54:46.1503838Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_bfloat16 PASSED [ 78%] 2023-03-31T05:54:46.1504303Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float16 PASSED [ 78%] 2023-03-31T05:54:46.1504755Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float32 PASSED [ 79%] 2023-03-31T05:54:46.1505222Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_non_contiguous_weight_cpu_int64_int64_float64 PASSED [ 80%] 2023-03-31T05:54:46.1505704Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float32_int32 PASSED [ 80%] 2023-03-31T05:54:46.1506238Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float32_int64 PASSED [ 81%] 2023-03-31T05:54:46.1506717Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float64_int32 PASSED [ 81%] 2023-03-31T05:54:46.1507202Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_max_cpu_float64_int64 PASSED [ 82%] 2023-03-31T05:54:46.1507692Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float32_int32 PASSED [ 83%] 2023-03-31T05:54:46.1508185Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float32_int64 PASSED [ 83%] 2023-03-31T05:54:46.1508662Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float64_int32 PASSED [ 84%] 2023-03-31T05:54:46.1509256Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_mean_cpu_float64_int64 PASSED [ 85%] 2023-03-31T05:54:46.1509776Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float32_int32 PASSED [ 85%] 2023-03-31T05:54:46.1510317Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float32_int64 PASSED [ 86%] 2023-03-31T05:54:46.1510794Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float64_int32 PASSED [ 86%] 2023-03-31T05:54:46.1511279Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_0_mode_sum_cpu_float64_int64 PASSED [ 87%] 2023-03-31T05:54:46.1511772Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float32_int32 PASSED [ 88%] 2023-03-31T05:54:46.1512279Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float32_int64 PASSED [ 88%] 2023-03-31T05:54:46.1512778Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float64_int32 PASSED [ 89%] 2023-03-31T05:54:46.1513263Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_max_cpu_float64_int64 PASSED [ 90%] 2023-03-31T05:54:46.1513804Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float32_int32 PASSED [ 90%] 2023-03-31T05:54:46.1514343Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float32_int64 PASSED [ 91%] 2023-03-31T05:54:46.1514847Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float64_int32 PASSED [ 91%] 2023-03-31T05:54:46.1515331Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_mean_cpu_float64_int64 PASSED [ 92%] 2023-03-31T05:54:46.1515830Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float32_int32 PASSED [ 93%] 2023-03-31T05:54:46.1516406Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float32_int64 PASSED [ 93%] 2023-03-31T05:54:46.1516956Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float64_int32 PASSED [ 94%] 2023-03-31T05:54:46.1517440Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_bag_out_of_bounds_idx_padding_idx_None_mode_sum_cpu_float64_int64 PASSED [ 95%] 2023-03-31T05:54:46.1517886Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_dense_grad_cpu PASSED [ 95%] 2023-03-31T05:54:46.1518342Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_backward_cpu_float32 PASSED [ 96%] 2023-03-31T05:54:46.1518795Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float16 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T05:54:46.1519251Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float32 SKIPPED (Only runs on cuda) [ 97%] 2023-03-31T05:54:46.1519717Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_device_cpu_float64 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T05:54:46.1520166Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_max_norm_fwd_AD_cpu_float32 PASSED [ 98%] 2023-03-31T05:54:46.1520598Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_padding_idx_cpu_float32 PASSED [ 99%] 2023-03-31T05:54:46.1521043Z nn/test_embedding.py::TestEmbeddingNNDeviceTypeCPU::test_embedding_scalar_weight_error_cpu PASSED [100%] 2023-03-31T05:54:46.1521273Z 2023-03-31T05:54:46.1521729Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_embedding/nn.test_embedding-970b421025a9958a.xml - 2023-03-31T05:54:46.1522111Z ======================== 156 passed, 4 skipped in 5.66s ======================== 2023-03-31T05:54:46.1522433Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:54:46.1522624Z 2023-03-31T05:54:46.1522966Z ##[endgroup] 2023-03-31T05:54:46.1523548Z FINISHED PRINTING LOG FILE of nn/test_embedding (/var/lib/jenkins/workspace/test/test-reports/nn-test_embedding_e49oayjw.log) 2023-03-31T05:54:46.1523866Z 2023-03-31T05:54:48.3106419Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:54:48.3134852Z Ignoring disabled issues: [] 2023-03-31T05:54:48.3324140Z Running inductor/test_standalone_compile ... [2023-03-31 05:54:48.329732] 2023-03-31T05:54:48.3324964Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'inductor/test_standalone_compile.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:54:48.330078] 2023-03-31T05:55:00.7729168Z 2023-03-31T05:55:00.7730814Z Expand the folded group to see the log file of inductor/test_standalone_compile 2023-03-31T05:55:00.7731587Z ##[group]PRINTING LOG FILE of inductor/test_standalone_compile (/var/lib/jenkins/workspace/test/test-reports/inductor-test_standalone_compile_r4l3kldt.log) 2023-03-31T05:55:00.7732375Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:00.7732979Z Test results will be stored in test-reports/python-pytest/inductor.test_standalone_compile/inductor.test_standalone_compile-12f270e795e46dc6.xml 2023-03-31T05:55:00.7733363Z ============================= test session starts ============================== 2023-03-31T05:55:00.7733771Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:00.7734037Z cachedir: .pytest_cache 2023-03-31T05:55:00.7734477Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:00.7734851Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:00.7735287Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:00.7735596Z collecting ... collected 8 items 2023-03-31T05:55:00.7736764Z Running 8 items in this shard: test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_bare_module, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_export1, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_export2, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx_dict_input, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx_tensor_return, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_make_fx, test/inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_op_with_multiple_outputs 2023-03-31T05:55:00.7737742Z 2023-03-31T05:55:00.7737930Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_bare_module PASSED [ 12%] 2023-03-31T05:55:00.7738345Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_export1 PASSED [ 25%] 2023-03-31T05:55:00.7738754Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_export2 PASSED [ 37%] 2023-03-31T05:55:00.7739153Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx PASSED [ 50%] 2023-03-31T05:55:00.7739589Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx_dict_input PASSED [ 62%] 2023-03-31T05:55:00.7740010Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_fx_tensor_return PASSED [ 75%] 2023-03-31T05:55:00.7740420Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_make_fx PASSED [ 87%] 2023-03-31T05:55:00.7740839Z inductor/test_standalone_compile.py::TestStandaloneInductor::test_inductor_via_op_with_multiple_outputs PASSED [100%] 2023-03-31T05:55:00.7741060Z 2023-03-31T05:55:00.7741470Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_standalone_compile/inductor.test_standalone_compile-12f270e795e46dc6.xml - 2023-03-31T05:55:00.7741877Z ============================== 8 passed in 5.59s =============================== 2023-03-31T05:55:00.7742279Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:00.7742486Z 2023-03-31T05:55:00.7742721Z ##[endgroup] 2023-03-31T05:55:00.7743201Z FINISHED PRINTING LOG FILE of inductor/test_standalone_compile (/var/lib/jenkins/workspace/test/test-reports/inductor-test_standalone_compile_r4l3kldt.log) 2023-03-31T05:55:00.7743469Z 2023-03-31T05:55:02.8631920Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:02.8660797Z Ignoring disabled issues: [] 2023-03-31T05:55:02.8827069Z Running test_datapipe ... [2023-03-31 05:55:02.882370] 2023-03-31T05:55:02.8829314Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_datapipe.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:02.882679] 2023-03-31T05:55:10.3103514Z 2023-03-31T05:55:10.3104035Z Expand the folded group to see the log file of test_datapipe 2023-03-31T05:55:10.3105107Z ##[group]PRINTING LOG FILE of test_datapipe (/var/lib/jenkins/workspace/test/test-reports/test_datapipe_e7t239wc.log) 2023-03-31T05:55:10.3106118Z Test results will be stored in test-reports/python-pytest/test_datapipe/test_datapipe-a20c618c60f64e38.xml 2023-03-31T05:55:10.3106477Z ============================= test session starts ============================== 2023-03-31T05:55:10.3106886Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:10.3107153Z cachedir: .pytest_cache 2023-03-31T05:55:10.3107591Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:10.3107967Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:10.3108413Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:10.3108732Z collecting ... collected 90 items 2023-03-31T05:55:10.3117967Z Running 90 items in this shard: test/test_datapipe.py::TestDataChunk::test_as_string, test/test_datapipe.py::TestDataChunk::test_getitem, test/test_datapipe.py::TestDataChunk::test_iter, test/test_datapipe.py::TestDataChunk::test_len, test/test_datapipe.py::TestDataChunk::test_random_shuffle, test/test_datapipe.py::TestDataChunk::test_reverse, test/test_datapipe.py::TestDataChunk::test_sort, test/test_datapipe.py::TestStreamWrapper::test_api, test/test_datapipe.py::TestStreamWrapper::test_dir, test/test_datapipe.py::TestStreamWrapper::test_pickle, test/test_datapipe.py::TestStreamWrapper::test_repr, test/test_datapipe.py::TestIterableDataPipeBasic::test_demux_mux_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_groupby_iterable_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_listdirfiles_iterable_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_listdirfilesdeterministic_iterable_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_map_with_col_file_handle_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_openfilesfromdisk_iterable_datapipe, test/test_datapipe.py::TestIterableDataPipeBasic::test_routeddecoder_iterable_datapipe, test/test_datapipe.py::TestCaptureDataFrame::test_basic_capture, test/test_datapipe.py::TestDataFramesPipes::test_batch, test/test_datapipe.py::TestDataFramesPipes::test_capture, test/test_datapipe.py::TestDataFramesPipes::test_collate, test/test_datapipe.py::TestDataFramesPipes::test_filter, test/test_datapipe.py::TestDataFramesPipes::test_shuffle, test/test_datapipe.py::TestDataFramesPipes::test_unbatch, test/test_datapipe.py::TestFunctionalIterDataPipe::test_batch_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_collate_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_concat_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_demux_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_filter_datapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_fork_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_iterable_wrapper_datapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_map_dict_with_col_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_map_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_map_tuple_list_with_col_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_mux_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_sampler_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_serializable, test/test_datapipe.py::TestFunctionalIterDataPipe::test_serializable_with_dill, test/test_datapipe.py::TestFunctionalIterDataPipe::test_shuffler_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_stream_reader_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_unbatch_iterdatapipe, test/test_datapipe.py::TestFunctionalIterDataPipe::test_zip_iterdatapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_batch_mapdatapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_concat_mapdatapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_map_mapdatapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_sequence_wrapper_datapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_serializable, test/test_datapipe.py::TestFunctionalMapDataPipe::test_serializable_with_dill, test/test_datapipe.py::TestFunctionalMapDataPipe::test_shuffler_mapdatapipe, test/test_datapipe.py::TestFunctionalMapDataPipe::test_zip_mapdatapipe, test/test_datapipe.py::TestTyping::test_compile_time, test/test_datapipe.py::TestTyping::test_construct_time, test/test_datapipe.py::TestTyping::test_isinstance, test/test_datapipe.py::TestTyping::test_issubinstance, test/test_datapipe.py::TestTyping::test_protocol, test/test_datapipe.py::TestTyping::test_reinforce, test/test_datapipe.py::TestTyping::test_runtime, test/test_datapipe.py::TestTyping::test_subtype, test/test_datapipe.py::TestGraph::test_simple_traverse, test/test_datapipe.py::TestGraph::test_traverse_circular_datapipe, test/test_datapipe.py::TestGraph::test_traverse_forked, test/test_datapipe.py::TestGraph::test_traverse_mapdatapipe, test/test_datapipe.py::TestGraph::test_traverse_mixdatapipe, test/test_datapipe.py::TestGraph::test_traverse_unhashable_datapipe, test/test_datapipe.py::TestSerialization::test_spawn_lambdas_iter, test/test_datapipe.py::TestSerialization::test_spawn_lambdas_map, test/test_datapipe.py::TestCircularSerialization::test_circular_serialization_with_dill, test/test_datapipe.py::TestCircularSerialization::test_circular_serialization_with_pickle, test/test_datapipe.py::TestSharding::test_multi_sharding, test/test_datapipe.py::TestSharding::test_old_dataloader, test/test_datapipe.py::TestSharding::test_sharding_groups, test/test_datapipe.py::TestSharding::test_sharding_groups_in_legacy_grouping_package, test/test_datapipe.py::TestSharding::test_sharding_length, test/test_datapipe.py::TestSharding::test_simple_sharding, test/test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_buggy, test/test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_constraint_multiple_outputs, test/test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_generator, test/test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_new_object, test/test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_self_next, test/test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_generator_function, test/test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_generator_function_exception, test/test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_next, test/test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_next_exception, test/test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_return_self, test/test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_custom_non_generator, test/test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_custom_self_next, test/test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph, test/test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph_repeated, test/test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph_with_serialization 2023-03-31T05:55:10.3126914Z 2023-03-31T05:55:10.3127063Z test_datapipe.py::TestDataChunk::test_as_string PASSED [ 1%] 2023-03-31T05:55:10.3127391Z test_datapipe.py::TestDataChunk::test_getitem PASSED [ 2%] 2023-03-31T05:55:10.3127752Z test_datapipe.py::TestDataChunk::test_iter PASSED [ 3%] 2023-03-31T05:55:10.3128091Z test_datapipe.py::TestDataChunk::test_len PASSED [ 4%] 2023-03-31T05:55:10.3128389Z test_datapipe.py::TestDataChunk::test_random_shuffle PASSED [ 5%] 2023-03-31T05:55:10.3128693Z test_datapipe.py::TestDataChunk::test_reverse PASSED [ 6%] 2023-03-31T05:55:10.3129003Z test_datapipe.py::TestDataChunk::test_sort PASSED [ 7%] 2023-03-31T05:55:10.3129543Z test_datapipe.py::TestStreamWrapper::test_api <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py PASSED [ 8%] 2023-03-31T05:55:10.3129957Z test_datapipe.py::TestStreamWrapper::test_dir PASSED [ 10%] 2023-03-31T05:55:10.3130272Z test_datapipe.py::TestStreamWrapper::test_pickle PASSED [ 11%] 2023-03-31T05:55:10.3130590Z test_datapipe.py::TestStreamWrapper::test_repr PASSED [ 12%] 2023-03-31T05:55:10.3130931Z test_datapipe.py::TestIterableDataPipeBasic::test_demux_mux_datapipe PASSED [ 13%] 2023-03-31T05:55:10.3131317Z test_datapipe.py::TestIterableDataPipeBasic::test_groupby_iterable_datapipe PASSED [ 14%] 2023-03-31T05:55:10.3131713Z test_datapipe.py::TestIterableDataPipeBasic::test_listdirfiles_iterable_datapipe PASSED [ 15%] 2023-03-31T05:55:10.3132149Z test_datapipe.py::TestIterableDataPipeBasic::test_listdirfilesdeterministic_iterable_datapipe PASSED [ 16%] 2023-03-31T05:55:10.3132607Z test_datapipe.py::TestIterableDataPipeBasic::test_map_with_col_file_handle_datapipe PASSED [ 17%] 2023-03-31T05:55:10.3133028Z test_datapipe.py::TestIterableDataPipeBasic::test_openfilesfromdisk_iterable_datapipe PASSED [ 18%] 2023-03-31T05:55:10.3133441Z test_datapipe.py::TestIterableDataPipeBasic::test_routeddecoder_iterable_datapipe PASSED [ 20%] 2023-03-31T05:55:10.3133831Z test_datapipe.py::TestCaptureDataFrame::test_basic_capture SKIPPED (no dataframes (pandas)) [ 21%] 2023-03-31T05:55:10.3134221Z test_datapipe.py::TestDataFramesPipes::test_batch SKIPPED (no dataframes (pandas)) [ 22%] 2023-03-31T05:55:10.3134597Z test_datapipe.py::TestDataFramesPipes::test_capture SKIPPED (no dataframes (pandas)) [ 23%] 2023-03-31T05:55:10.3134971Z test_datapipe.py::TestDataFramesPipes::test_collate SKIPPED (no dataframes (pandas)) [ 24%] 2023-03-31T05:55:10.3135365Z test_datapipe.py::TestDataFramesPipes::test_filter SKIPPED (no dataframes (pandas)) [ 25%] 2023-03-31T05:55:10.3135741Z test_datapipe.py::TestDataFramesPipes::test_shuffle SKIPPED (no dataframes (pandas)) [ 26%] 2023-03-31T05:55:10.3136112Z test_datapipe.py::TestDataFramesPipes::test_unbatch SKIPPED (no dataframes (pandas)) [ 27%] 2023-03-31T05:55:10.3136487Z test_datapipe.py::TestFunctionalIterDataPipe::test_batch_iterdatapipe PASSED [ 28%] 2023-03-31T05:55:10.3136865Z test_datapipe.py::TestFunctionalIterDataPipe::test_collate_iterdatapipe PASSED [ 30%] 2023-03-31T05:55:10.3137249Z test_datapipe.py::TestFunctionalIterDataPipe::test_concat_iterdatapipe PASSED [ 31%] 2023-03-31T05:55:10.3137635Z test_datapipe.py::TestFunctionalIterDataPipe::test_demux_iterdatapipe PASSED [ 32%] 2023-03-31T05:55:10.3138005Z test_datapipe.py::TestFunctionalIterDataPipe::test_filter_datapipe PASSED [ 33%] 2023-03-31T05:55:10.3138382Z test_datapipe.py::TestFunctionalIterDataPipe::test_fork_iterdatapipe PASSED [ 34%] 2023-03-31T05:55:10.3138768Z test_datapipe.py::TestFunctionalIterDataPipe::test_iterable_wrapper_datapipe PASSED [ 35%] 2023-03-31T05:55:10.3139172Z test_datapipe.py::TestFunctionalIterDataPipe::test_map_dict_with_col_iterdatapipe PASSED [ 36%] 2023-03-31T05:55:10.3139550Z test_datapipe.py::TestFunctionalIterDataPipe::test_map_iterdatapipe PASSED [ 37%] 2023-03-31T05:55:10.3139947Z test_datapipe.py::TestFunctionalIterDataPipe::test_map_tuple_list_with_col_iterdatapipe PASSED [ 38%] 2023-03-31T05:55:10.3140345Z test_datapipe.py::TestFunctionalIterDataPipe::test_mux_iterdatapipe PASSED [ 40%] 2023-03-31T05:55:10.3140745Z test_datapipe.py::TestFunctionalIterDataPipe::test_sampler_iterdatapipe PASSED [ 41%] 2023-03-31T05:55:10.3141148Z test_datapipe.py::TestFunctionalIterDataPipe::test_serializable PASSED [ 42%] 2023-03-31T05:55:10.3141525Z test_datapipe.py::TestFunctionalIterDataPipe::test_serializable_with_dill PASSED [ 43%] 2023-03-31T05:55:10.3141908Z test_datapipe.py::TestFunctionalIterDataPipe::test_shuffler_iterdatapipe PASSED [ 44%] 2023-03-31T05:55:10.3142399Z test_datapipe.py::TestFunctionalIterDataPipe::test_stream_reader_iterdatapipe PASSED [ 45%] 2023-03-31T05:55:10.3142788Z test_datapipe.py::TestFunctionalIterDataPipe::test_unbatch_iterdatapipe PASSED [ 46%] 2023-03-31T05:55:10.3143168Z test_datapipe.py::TestFunctionalIterDataPipe::test_zip_iterdatapipe PASSED [ 47%] 2023-03-31T05:55:10.3143527Z test_datapipe.py::TestFunctionalMapDataPipe::test_batch_mapdatapipe PASSED [ 48%] 2023-03-31T05:55:10.3143902Z test_datapipe.py::TestFunctionalMapDataPipe::test_concat_mapdatapipe PASSED [ 50%] 2023-03-31T05:55:10.3144276Z test_datapipe.py::TestFunctionalMapDataPipe::test_map_mapdatapipe PASSED [ 51%] 2023-03-31T05:55:10.3144660Z test_datapipe.py::TestFunctionalMapDataPipe::test_sequence_wrapper_datapipe PASSED [ 52%] 2023-03-31T05:55:10.3145023Z test_datapipe.py::TestFunctionalMapDataPipe::test_serializable PASSED [ 53%] 2023-03-31T05:55:10.3145393Z test_datapipe.py::TestFunctionalMapDataPipe::test_serializable_with_dill PASSED [ 54%] 2023-03-31T05:55:10.3145776Z test_datapipe.py::TestFunctionalMapDataPipe::test_shuffler_mapdatapipe PASSED [ 55%] 2023-03-31T05:55:10.3146178Z test_datapipe.py::TestFunctionalMapDataPipe::test_zip_mapdatapipe PASSED [ 56%] 2023-03-31T05:55:10.3146544Z test_datapipe.py::TestTyping::test_compile_time SKIPPED (TODO: Fix typing bug) [ 57%] 2023-03-31T05:55:10.3146896Z test_datapipe.py::TestTyping::test_construct_time SKIPPED (TODO: Fix typing bug) [ 58%] 2023-03-31T05:55:10.3147225Z test_datapipe.py::TestTyping::test_isinstance PASSED [ 60%] 2023-03-31T05:55:10.3147549Z test_datapipe.py::TestTyping::test_issubinstance SKIPPED (TODO: Fix typing bug) [ 61%] 2023-03-31T05:55:10.3147882Z test_datapipe.py::TestTyping::test_protocol PASSED [ 62%] 2023-03-31T05:55:10.3148212Z test_datapipe.py::TestTyping::test_reinforce SKIPPED (TODO: Fix typing bug) [ 63%] 2023-03-31T05:55:10.3148536Z test_datapipe.py::TestTyping::test_runtime SKIPPED (TODO: Fix typing bug) [ 64%] 2023-03-31T05:55:10.3148912Z test_datapipe.py::TestTyping::test_subtype SKIPPED (TODO: Fix typing bug) [ 65%] 2023-03-31T05:55:10.3149238Z test_datapipe.py::TestGraph::test_simple_traverse PASSED [ 66%] 2023-03-31T05:55:10.3149557Z test_datapipe.py::TestGraph::test_traverse_circular_datapipe PASSED [ 67%] 2023-03-31T05:55:10.3149867Z test_datapipe.py::TestGraph::test_traverse_forked PASSED [ 68%] 2023-03-31T05:55:10.3150180Z test_datapipe.py::TestGraph::test_traverse_mapdatapipe PASSED [ 70%] 2023-03-31T05:55:10.3150498Z test_datapipe.py::TestGraph::test_traverse_mixdatapipe PASSED [ 71%] 2023-03-31T05:55:10.3150818Z test_datapipe.py::TestGraph::test_traverse_unhashable_datapipe PASSED [ 72%] 2023-03-31T05:55:10.3151164Z test_datapipe.py::TestSerialization::test_spawn_lambdas_iter SKIPPED (no dill) [ 73%] 2023-03-31T05:55:10.3151519Z test_datapipe.py::TestSerialization::test_spawn_lambdas_map SKIPPED (no dill) [ 74%] 2023-03-31T05:55:10.3151915Z test_datapipe.py::TestCircularSerialization::test_circular_serialization_with_dill SKIPPED (no dill) [ 75%] 2023-03-31T05:55:10.3152324Z test_datapipe.py::TestCircularSerialization::test_circular_serialization_with_pickle PASSED [ 76%] 2023-03-31T05:55:10.3152688Z test_datapipe.py::TestSharding::test_multi_sharding PASSED [ 77%] 2023-03-31T05:55:10.3153007Z test_datapipe.py::TestSharding::test_old_dataloader PASSED [ 78%] 2023-03-31T05:55:10.3153313Z test_datapipe.py::TestSharding::test_sharding_groups PASSED [ 80%] 2023-03-31T05:55:10.3153662Z test_datapipe.py::TestSharding::test_sharding_groups_in_legacy_grouping_package PASSED [ 81%] 2023-03-31T05:55:10.3154057Z test_datapipe.py::TestSharding::test_sharding_length PASSED [ 82%] 2023-03-31T05:55:10.3154405Z test_datapipe.py::TestSharding::test_simple_sharding PASSED [ 83%] 2023-03-31T05:55:10.3154782Z test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_buggy PASSED [ 84%] 2023-03-31T05:55:10.3155265Z test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_constraint_multiple_outputs PASSED [ 85%] 2023-03-31T05:55:10.3155749Z test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_generator PASSED [ 86%] 2023-03-31T05:55:10.3156208Z test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_new_object PASSED [ 87%] 2023-03-31T05:55:10.3156652Z test_datapipe.py::TestIterDataPipeSingletonConstraint::test_iterdatapipe_singleton_self_next PASSED [ 88%] 2023-03-31T05:55:10.3157127Z test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_generator_function PASSED [ 90%] 2023-03-31T05:55:10.3157636Z test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_generator_function_exception PASSED [ 91%] 2023-03-31T05:55:10.3158126Z test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_next PASSED [ 92%] 2023-03-31T05:55:10.3158588Z test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_next_exception PASSED [ 93%] 2023-03-31T05:55:10.3159098Z test_datapipe.py::TestIterDataPipeCountSampleYielded::test_iterdatapipe_sample_yielded_return_self PASSED [ 94%] 2023-03-31T05:55:10.3159562Z test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_custom_non_generator PASSED [ 95%] 2023-03-31T05:55:10.3159992Z test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_custom_self_next PASSED [ 96%] 2023-03-31T05:55:10.3160416Z test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph PASSED [ 97%] 2023-03-31T05:55:10.3160845Z test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph_repeated PASSED [ 98%] 2023-03-31T05:55:10.3161299Z test_datapipe.py::TestIterDataPipeGraphFastForward::test_simple_snapshot_graph_with_serialization PASSED [100%] 2023-03-31T05:55:10.3161545Z 2023-03-31T05:55:10.3161955Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_datapipe/test_datapipe-a20c618c60f64e38.xml - 2023-03-31T05:55:10.3162330Z ======================== 74 passed, 16 skipped in 5.44s ======================== 2023-03-31T05:55:10.3162654Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:10.3162857Z 2023-03-31T05:55:10.3163298Z ##[endgroup] 2023-03-31T05:55:10.3163710Z FINISHED PRINTING LOG FILE of test_datapipe (/var/lib/jenkins/workspace/test/test-reports/test_datapipe_e7t239wc.log) 2023-03-31T05:55:10.3163943Z 2023-03-31T05:55:12.3930835Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:12.3959579Z Ignoring disabled issues: [] 2023-03-31T05:55:12.4127703Z Running test_functionalization ... [2023-03-31 05:55:12.412330] 2023-03-31T05:55:12.4129042Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_functionalization.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:12.412638] 2023-03-31T05:55:18.9405877Z 2023-03-31T05:55:18.9406613Z Expand the folded group to see the log file of test_functionalization 2023-03-31T05:55:18.9407761Z ##[group]PRINTING LOG FILE of test_functionalization (/var/lib/jenkins/workspace/test/test-reports/test_functionalization_ksr14jeb.log) 2023-03-31T05:55:18.9408915Z Test results will be stored in test-reports/python-pytest/test_functionalization/test_functionalization-c2940ed962564e0b.xml 2023-03-31T05:55:18.9409531Z ============================= test session starts ============================== 2023-03-31T05:55:18.9410393Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:18.9410924Z cachedir: .pytest_cache 2023-03-31T05:55:18.9411704Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:18.9412304Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:18.9444588Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:18.9445177Z collecting ... collected 88 items 2023-03-31T05:55:18.9457682Z Running 88 items in this shard: test/test_functionalization.py::TestFunctionalization::test_advanced_indexing, test/test_functionalization.py::TestFunctionalization::test_advanced_indexing_correct_strides, test/test_functionalization.py::TestFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views, test/test_functionalization.py::TestFunctionalization::test_as_strided, test/test_functionalization.py::TestFunctionalization::test_batch_norm, test/test_functionalization.py::TestFunctionalization::test_cat, test/test_functionalization.py::TestFunctionalization::test_channels_last_contiguous, test/test_functionalization.py::TestFunctionalization::test_copy_, test/test_functionalization.py::TestFunctionalization::test_copy_stride_mismatch, test/test_functionalization.py::TestFunctionalization::test_diagonal, test/test_functionalization.py::TestFunctionalization::test_diagonal_mutated_input, test/test_functionalization.py::TestFunctionalization::test_everything, test/test_functionalization.py::TestFunctionalization::test_expand_symint, test/test_functionalization.py::TestFunctionalization::test_fill_, test/test_functionalization.py::TestFunctionalization::test_freeze, test/test_functionalization.py::TestFunctionalization::test_index_mutation_on_non_input, test/test_functionalization.py::TestFunctionalization::test_inplace_on_non_view, test/test_functionalization.py::TestFunctionalization::test_instance_norm, test/test_functionalization.py::TestFunctionalization::test_metadata_change, test/test_functionalization.py::TestFunctionalization::test_metadata_change_out_op, test/test_functionalization.py::TestFunctionalization::test_mixed_wrappers_invalid, test/test_functionalization.py::TestFunctionalization::test_mixed_wrappers_valid, test/test_functionalization.py::TestFunctionalization::test_multi_out, test/test_functionalization.py::TestFunctionalization::test_multiple_views_of_same_base, test/test_functionalization.py::TestFunctionalization::test_mutable_op_not_inplace_or_other, test/test_functionalization.py::TestFunctionalization::test_nested_functions_propagate_updates, test/test_functionalization.py::TestFunctionalization::test_only_one_view, test/test_functionalization.py::TestFunctionalization::test_optional_tensor_list, test/test_functionalization.py::TestFunctionalization::test_reapply_views_simple, test/test_functionalization.py::TestFunctionalization::test_resize_larger_invalid, test/test_functionalization.py::TestFunctionalization::test_resize_larger_valid, test/test_functionalization.py::TestFunctionalization::test_resize_same_size_diff_rank, test/test_functionalization.py::TestFunctionalization::test_resize_smaller, test/test_functionalization.py::TestFunctionalization::test_save_for_backwards_segfault, test/test_functionalization.py::TestFunctionalization::test_scalars, test/test_functionalization.py::TestFunctionalization::test_set_, test/test_functionalization.py::TestFunctionalization::test_simple, test/test_functionalization.py::TestFunctionalization::test_simple_out, test/test_functionalization.py::TestFunctionalization::test_split, test/test_functionalization.py::TestFunctionalization::test_tensor_ctr, test/test_functionalization.py::TestFunctionalization::test_tensor_list_composite, test/test_functionalization.py::TestFunctionalization::test_tensor_list_mixed_functional_nonfunctional, test/test_functionalization.py::TestFunctionalization::test_view_clone_view_inplace, test/test_functionalization.py::TestFunctionalization::test_view_inplace, test/test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing, test/test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing_correct_strides, test/test_functionalization.py::TestCrossRefFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views, test/test_functionalization.py::TestCrossRefFunctionalization::test_as_strided, test/test_functionalization.py::TestCrossRefFunctionalization::test_batch_norm, test/test_functionalization.py::TestCrossRefFunctionalization::test_cat, test/test_functionalization.py::TestCrossRefFunctionalization::test_channels_last_contiguous, test/test_functionalization.py::TestCrossRefFunctionalization::test_copy_, test/test_functionalization.py::TestCrossRefFunctionalization::test_copy_stride_mismatch, test/test_functionalization.py::TestCrossRefFunctionalization::test_diagonal, test/test_functionalization.py::TestCrossRefFunctionalization::test_diagonal_mutated_input, test/test_functionalization.py::TestCrossRefFunctionalization::test_everything, test/test_functionalization.py::TestCrossRefFunctionalization::test_expand_symint, test/test_functionalization.py::TestCrossRefFunctionalization::test_fill_, test/test_functionalization.py::TestCrossRefFunctionalization::test_freeze, test/test_functionalization.py::TestCrossRefFunctionalization::test_index_mutation_on_non_input, test/test_functionalization.py::TestCrossRefFunctionalization::test_inplace_on_non_view, test/test_functionalization.py::TestCrossRefFunctionalization::test_instance_norm, test/test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change, test/test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change_out_op, test/test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_invalid, test/test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_valid, test/test_functionalization.py::TestCrossRefFunctionalization::test_multi_out, test/test_functionalization.py::TestCrossRefFunctionalization::test_multiple_views_of_same_base, test/test_functionalization.py::TestCrossRefFunctionalization::test_mutable_op_not_inplace_or_other, test/test_functionalization.py::TestCrossRefFunctionalization::test_nested_functions_propagate_updates, test/test_functionalization.py::TestCrossRefFunctionalization::test_only_one_view, test/test_functionalization.py::TestCrossRefFunctionalization::test_optional_tensor_list, test/test_functionalization.py::TestCrossRefFunctionalization::test_reapply_views_simple, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_invalid, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_valid, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_same_size_diff_rank, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_smaller, test/test_functionalization.py::TestCrossRefFunctionalization::test_save_for_backwards_segfault, test/test_functionalization.py::TestCrossRefFunctionalization::test_scalars, test/test_functionalization.py::TestCrossRefFunctionalization::test_set_, test/test_functionalization.py::TestCrossRefFunctionalization::test_simple, test/test_functionalization.py::TestCrossRefFunctionalization::test_simple_out, test/test_functionalization.py::TestCrossRefFunctionalization::test_split, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_ctr, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_composite, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_mixed_functional_nonfunctional, test/test_functionalization.py::TestCrossRefFunctionalization::test_view_clone_view_inplace, test/test_functionalization.py::TestCrossRefFunctionalization::test_view_inplace 2023-03-31T05:55:18.9467127Z 2023-03-31T05:55:18.9467333Z test_functionalization.py::TestFunctionalization::test_advanced_indexing PASSED [ 1%] 2023-03-31T05:55:18.9467777Z test_functionalization.py::TestFunctionalization::test_advanced_indexing_correct_strides PASSED [ 2%] 2023-03-31T05:55:18.9468255Z test_functionalization.py::TestFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views PASSED [ 3%] 2023-03-31T05:55:18.9468666Z test_functionalization.py::TestFunctionalization::test_as_strided PASSED [ 4%] 2023-03-31T05:55:18.9469022Z test_functionalization.py::TestFunctionalization::test_batch_norm PASSED [ 5%] 2023-03-31T05:55:18.9469385Z test_functionalization.py::TestFunctionalization::test_cat PASSED [ 6%] 2023-03-31T05:55:18.9469763Z test_functionalization.py::TestFunctionalization::test_channels_last_contiguous PASSED [ 7%] 2023-03-31T05:55:18.9470139Z test_functionalization.py::TestFunctionalization::test_copy_ PASSED [ 9%] 2023-03-31T05:55:18.9470496Z test_functionalization.py::TestFunctionalization::test_copy_stride_mismatch PASSED [ 10%] 2023-03-31T05:55:18.9470868Z test_functionalization.py::TestFunctionalization::test_diagonal PASSED [ 11%] 2023-03-31T05:55:18.9471244Z test_functionalization.py::TestFunctionalization::test_diagonal_mutated_input PASSED [ 12%] 2023-03-31T05:55:18.9471634Z test_functionalization.py::TestFunctionalization::test_everything PASSED [ 13%] 2023-03-31T05:55:18.9472036Z test_functionalization.py::TestFunctionalization::test_expand_symint PASSED [ 14%] 2023-03-31T05:55:18.9472429Z test_functionalization.py::TestFunctionalization::test_fill_ PASSED [ 15%] 2023-03-31T05:55:18.9472817Z test_functionalization.py::TestFunctionalization::test_freeze PASSED [ 17%] 2023-03-31T05:55:18.9473186Z test_functionalization.py::TestFunctionalization::test_index_mutation_on_non_input PASSED [ 18%] 2023-03-31T05:55:18.9473573Z test_functionalization.py::TestFunctionalization::test_inplace_on_non_view PASSED [ 19%] 2023-03-31T05:55:18.9473946Z test_functionalization.py::TestFunctionalization::test_instance_norm PASSED [ 20%] 2023-03-31T05:55:18.9474323Z test_functionalization.py::TestFunctionalization::test_metadata_change PASSED [ 21%] 2023-03-31T05:55:18.9474693Z test_functionalization.py::TestFunctionalization::test_metadata_change_out_op PASSED [ 22%] 2023-03-31T05:55:18.9475079Z test_functionalization.py::TestFunctionalization::test_mixed_wrappers_invalid PASSED [ 23%] 2023-03-31T05:55:18.9475459Z test_functionalization.py::TestFunctionalization::test_mixed_wrappers_valid PASSED [ 25%] 2023-03-31T05:55:18.9475862Z test_functionalization.py::TestFunctionalization::test_multi_out PASSED [ 26%] 2023-03-31T05:55:18.9476231Z test_functionalization.py::TestFunctionalization::test_multiple_views_of_same_base PASSED [ 27%] 2023-03-31T05:55:18.9476631Z test_functionalization.py::TestFunctionalization::test_mutable_op_not_inplace_or_other PASSED [ 28%] 2023-03-31T05:55:18.9477044Z test_functionalization.py::TestFunctionalization::test_nested_functions_propagate_updates PASSED [ 29%] 2023-03-31T05:55:18.9477421Z test_functionalization.py::TestFunctionalization::test_only_one_view PASSED [ 30%] 2023-03-31T05:55:18.9477795Z test_functionalization.py::TestFunctionalization::test_optional_tensor_list PASSED [ 31%] 2023-03-31T05:55:18.9478179Z test_functionalization.py::TestFunctionalization::test_reapply_views_simple PASSED [ 32%] 2023-03-31T05:55:18.9478565Z test_functionalization.py::TestFunctionalization::test_resize_larger_invalid PASSED [ 34%] 2023-03-31T05:55:18.9478934Z test_functionalization.py::TestFunctionalization::test_resize_larger_valid PASSED [ 35%] 2023-03-31T05:55:18.9479320Z test_functionalization.py::TestFunctionalization::test_resize_same_size_diff_rank PASSED [ 36%] 2023-03-31T05:55:18.9479701Z test_functionalization.py::TestFunctionalization::test_resize_smaller PASSED [ 37%] 2023-03-31T05:55:18.9480079Z test_functionalization.py::TestFunctionalization::test_save_for_backwards_segfault PASSED [ 38%] 2023-03-31T05:55:18.9480460Z test_functionalization.py::TestFunctionalization::test_scalars PASSED [ 39%] 2023-03-31T05:55:18.9480813Z test_functionalization.py::TestFunctionalization::test_set_ PASSED [ 40%] 2023-03-31T05:55:18.9481190Z test_functionalization.py::TestFunctionalization::test_simple PASSED [ 42%] 2023-03-31T05:55:18.9481569Z test_functionalization.py::TestFunctionalization::test_simple_out PASSED [ 43%] 2023-03-31T05:55:18.9481915Z test_functionalization.py::TestFunctionalization::test_split PASSED [ 44%] 2023-03-31T05:55:18.9482269Z test_functionalization.py::TestFunctionalization::test_tensor_ctr PASSED [ 45%] 2023-03-31T05:55:18.9482630Z test_functionalization.py::TestFunctionalization::test_tensor_list_composite PASSED [ 46%] 2023-03-31T05:55:18.9483252Z test_functionalization.py::TestFunctionalization::test_tensor_list_mixed_functional_nonfunctional PASSED [ 47%] 2023-03-31T05:55:18.9483786Z test_functionalization.py::TestFunctionalization::test_view_clone_view_inplace PASSED [ 48%] 2023-03-31T05:55:18.9484406Z test_functionalization.py::TestFunctionalization::test_view_inplace PASSED [ 50%] 2023-03-31T05:55:18.9484988Z test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing PASSED [ 51%] 2023-03-31T05:55:18.9485433Z test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing_correct_strides PASSED [ 52%] 2023-03-31T05:55:18.9485903Z test_functionalization.py::TestCrossRefFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views PASSED [ 53%] 2023-03-31T05:55:18.9486345Z test_functionalization.py::TestCrossRefFunctionalization::test_as_strided XFAIL [ 54%] 2023-03-31T05:55:18.9486807Z test_functionalization.py::TestCrossRefFunctionalization::test_batch_norm PASSED [ 55%] 2023-03-31T05:55:18.9487197Z test_functionalization.py::TestCrossRefFunctionalization::test_cat PASSED [ 56%] 2023-03-31T05:55:18.9487604Z test_functionalization.py::TestCrossRefFunctionalization::test_channels_last_contiguous PASSED [ 57%] 2023-03-31T05:55:18.9487997Z test_functionalization.py::TestCrossRefFunctionalization::test_copy_ XFAIL [ 59%] 2023-03-31T05:55:18.9488399Z test_functionalization.py::TestCrossRefFunctionalization::test_copy_stride_mismatch PASSED [ 60%] 2023-03-31T05:55:18.9488802Z test_functionalization.py::TestCrossRefFunctionalization::test_diagonal XFAIL [ 61%] 2023-03-31T05:55:18.9489206Z test_functionalization.py::TestCrossRefFunctionalization::test_diagonal_mutated_input XFAIL [ 62%] 2023-03-31T05:55:18.9489600Z test_functionalization.py::TestCrossRefFunctionalization::test_everything XFAIL [ 63%] 2023-03-31T05:55:18.9490041Z test_functionalization.py::TestCrossRefFunctionalization::test_expand_symint PASSED [ 64%] 2023-03-31T05:55:18.9490430Z test_functionalization.py::TestCrossRefFunctionalization::test_fill_ XFAIL [ 65%] 2023-03-31T05:55:18.9490801Z test_functionalization.py::TestCrossRefFunctionalization::test_freeze PASSED [ 67%] 2023-03-31T05:55:18.9491212Z test_functionalization.py::TestCrossRefFunctionalization::test_index_mutation_on_non_input PASSED [ 68%] 2023-03-31T05:55:18.9491632Z test_functionalization.py::TestCrossRefFunctionalization::test_inplace_on_non_view PASSED [ 69%] 2023-03-31T05:55:18.9492041Z test_functionalization.py::TestCrossRefFunctionalization::test_instance_norm PASSED [ 70%] 2023-03-31T05:55:18.9492437Z test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change PASSED [ 71%] 2023-03-31T05:55:18.9492848Z test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change_out_op PASSED [ 72%] 2023-03-31T05:55:18.9493274Z test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_invalid PASSED [ 73%] 2023-03-31T05:55:18.9493699Z test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_valid PASSED [ 75%] 2023-03-31T05:55:18.9494090Z test_functionalization.py::TestCrossRefFunctionalization::test_multi_out PASSED [ 76%] 2023-03-31T05:55:18.9494517Z test_functionalization.py::TestCrossRefFunctionalization::test_multiple_views_of_same_base PASSED [ 77%] 2023-03-31T05:55:18.9494954Z test_functionalization.py::TestCrossRefFunctionalization::test_mutable_op_not_inplace_or_other PASSED [ 78%] 2023-03-31T05:55:18.9495447Z test_functionalization.py::TestCrossRefFunctionalization::test_nested_functions_propagate_updates PASSED [ 79%] 2023-03-31T05:55:18.9495896Z test_functionalization.py::TestCrossRefFunctionalization::test_only_one_view PASSED [ 80%] 2023-03-31T05:55:18.9496302Z test_functionalization.py::TestCrossRefFunctionalization::test_optional_tensor_list PASSED [ 81%] 2023-03-31T05:55:18.9496717Z test_functionalization.py::TestCrossRefFunctionalization::test_reapply_views_simple PASSED [ 82%] 2023-03-31T05:55:18.9497130Z test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_invalid PASSED [ 84%] 2023-03-31T05:55:18.9497546Z test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_valid PASSED [ 85%] 2023-03-31T05:55:18.9498096Z test_functionalization.py::TestCrossRefFunctionalization::test_resize_same_size_diff_rank PASSED [ 86%] 2023-03-31T05:55:18.9498649Z test_functionalization.py::TestCrossRefFunctionalization::test_resize_smaller PASSED [ 87%] 2023-03-31T05:55:18.9499355Z test_functionalization.py::TestCrossRefFunctionalization::test_save_for_backwards_segfault PASSED [ 88%] 2023-03-31T05:55:18.9499925Z test_functionalization.py::TestCrossRefFunctionalization::test_scalars PASSED [ 89%] 2023-03-31T05:55:18.9500795Z test_functionalization.py::TestCrossRefFunctionalization::test_set_ PASSED [ 90%] 2023-03-31T05:55:18.9501474Z test_functionalization.py::TestCrossRefFunctionalization::test_simple PASSED [ 92%] 2023-03-31T05:55:18.9502160Z test_functionalization.py::TestCrossRefFunctionalization::test_simple_out PASSED [ 93%] 2023-03-31T05:55:18.9502873Z test_functionalization.py::TestCrossRefFunctionalization::test_split XFAIL [ 94%] 2023-03-31T05:55:18.9503381Z test_functionalization.py::TestCrossRefFunctionalization::test_tensor_ctr PASSED [ 95%] 2023-03-31T05:55:18.9503790Z test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_composite PASSED [ 96%] 2023-03-31T05:55:18.9504233Z test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_mixed_functional_nonfunctional PASSED [ 97%] 2023-03-31T05:55:18.9504680Z test_functionalization.py::TestCrossRefFunctionalization::test_view_clone_view_inplace XFAIL [ 98%] 2023-03-31T05:55:18.9505086Z test_functionalization.py::TestCrossRefFunctionalization::test_view_inplace XFAIL [100%] 2023-03-31T05:55:18.9505303Z 2023-03-31T05:55:18.9505799Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functionalization/test_functionalization-c2940ed962564e0b.xml - 2023-03-31T05:55:18.9506204Z ======================== 79 passed, 9 xfailed in 4.55s ========================= 2023-03-31T05:55:18.9506532Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:18.9506735Z 2023-03-31T05:55:18.9507071Z ##[endgroup] 2023-03-31T05:55:18.9507510Z FINISHED PRINTING LOG FILE of test_functionalization (/var/lib/jenkins/workspace/test/test-reports/test_functionalization_ksr14jeb.log) 2023-03-31T05:55:18.9507763Z 2023-03-31T05:55:20.9987772Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:21.0015374Z Ignoring disabled issues: [] 2023-03-31T05:55:21.0181218Z Running profiler/test_memory_profiler ... [2023-03-31 05:55:21.017735] 2023-03-31T05:55:21.0183123Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'profiler/test_memory_profiler.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:21.018054] 2023-03-31T05:55:27.0721799Z 2023-03-31T05:55:27.0722868Z Expand the folded group to see the log file of profiler/test_memory_profiler 2023-03-31T05:55:27.0724172Z ##[group]PRINTING LOG FILE of profiler/test_memory_profiler (/var/lib/jenkins/workspace/test/test-reports/profiler-test_memory_profiler_fshex0e3.log) 2023-03-31T05:55:27.0725148Z Test results will be stored in test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-f170e6cb75f486bf.xml 2023-03-31T05:55:27.0725766Z ============================= test session starts ============================== 2023-03-31T05:55:27.0726272Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:27.0726607Z cachedir: .pytest_cache 2023-03-31T05:55:27.0727240Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:27.0727802Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:27.0728507Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:27.0728883Z collecting ... collected 32 items 2023-03-31T05:55:27.0732826Z Running 32 items in this shard: test/profiler/test_memory_profiler.py::TestMemoryProfiler::test_config_check, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_module, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_module_and_optimizer, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_low_level, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_complicated, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_non_op_allocations, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple_backward, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple_inplace, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_stacked, test/profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_with_annotations, test/profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas, test/profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas_backward, test/profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas_tensorlist, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_sequential_fwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_sequential_fwd_bwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd_bwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd_bwd_step, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd_bwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd_bwd_step, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd_bwd, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd_lazy, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_lazily_initialized, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_manual_optimizer_step, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_memory_timeline, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_parameters_and_gradients, test/profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_parameters_and_gradients_set_to_none 2023-03-31T05:55:27.0736328Z 2023-03-31T05:55:27.0736748Z profiler/test_memory_profiler.py::TestMemoryProfiler::test_config_check STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0737298Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0737838Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0738321Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0738780Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0739254Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0739708Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0740147Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0740614Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0740891Z PASSED [ 3%] 2023-03-31T05:55:27.0741437Z profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_module STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0741984Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0742447Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0742942Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0743483Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0743946Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0744404Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0744860Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0745309Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0745763Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0746214Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0746719Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0746979Z PASSED [ 6%] 2023-03-31T05:55:27.0747540Z profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_module_and_optimizer STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0748112Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0748581Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0748846Z PASSED [ 9%] 2023-03-31T05:55:27.0749389Z profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0749948Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0750403Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0750855Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0751304Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0751768Z STAGE:2023-03-31 05:55:22 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0752250Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0752737Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0753200Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0753655Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0754093Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0754561Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0754836Z PASSED [ 12%] 2023-03-31T05:55:27.0755383Z profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0755957Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0756423Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0756873Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0757312Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0757816Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0758268Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0758718Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0759168Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0759622Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0760072Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0760520Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0760835Z PASSED [ 15%] 2023-03-31T05:55:27.0761369Z profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_low_level STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0761921Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0762369Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0762825Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0763461Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0764115Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0764573Z PASSED [ 18%] 2023-03-31T05:55:27.0766240Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_complicated STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0767158Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0767982Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0768716Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0769489Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0770437Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0770997Z PASSED [ 21%] 2023-03-31T05:55:27.0771939Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_non_op_allocations STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0772872Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0773631Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0774093Z PASSED [ 25%] 2023-03-31T05:55:27.0775010Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0775938Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0776765Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0777570Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0778387Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0779321Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0779769Z PASSED [ 28%] 2023-03-31T05:55:27.0780706Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple_backward STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0781676Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0782158Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0782421Z PASSED [ 31%] 2023-03-31T05:55:27.0782931Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_simple_inplace STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0783635Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0784100Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0784540Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0784989Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0785449Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0785721Z PASSED [ 34%] 2023-03-31T05:55:27.0786209Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_stacked STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0786737Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0787196Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0787650Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0788084Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0788545Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0788994Z STAGE:2023-03-31 05:55:23 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0789473Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0789964Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0790239Z PASSED [ 37%] 2023-03-31T05:55:27.0790753Z profiler/test_memory_profiler.py::TestDataFlow::test_data_flow_graph_with_annotations STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0791279Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0791736Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0792007Z PASSED [ 40%] 2023-03-31T05:55:27.0792476Z profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0793000Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0793460Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0793729Z PASSED [ 43%] 2023-03-31T05:55:27.0794213Z profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas_backward STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0794873Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0795338Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0795611Z PASSED [ 46%] 2023-03-31T05:55:27.0796101Z profiler/test_memory_profiler.py::TestDataFlow::test_match_schemas_tensorlist STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0796627Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0797097Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0797358Z PASSED [ 50%] 2023-03-31T05:55:27.0797889Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_sequential_fwd STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0798479Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0798938Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0799199Z PASSED [ 53%] 2023-03-31T05:55:27.0799732Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_sequential_fwd_bwd STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0800286Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0800807Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0801081Z PASSED [ 56%] 2023-03-31T05:55:27.0801592Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0802142Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0802604Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0802873Z PASSED [ 59%] 2023-03-31T05:55:27.0803567Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd_bwd STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0804192Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0804696Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0804961Z PASSED [ 62%] 2023-03-31T05:55:27.0805494Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_fwd_bwd_step STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0806055Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0806522Z STAGE:2023-03-31 05:55:24 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0806781Z PASSED [ 65%] 2023-03-31T05:55:27.0807317Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0807875Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0808344Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0808604Z PASSED [ 68%] 2023-03-31T05:55:27.0809148Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd_bwd STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0809751Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0810203Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0810478Z PASSED [ 71%] 2023-03-31T05:55:27.0811030Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_categories_e2e_simple_module_fwd_bwd_step STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0811591Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0812045Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0812366Z PASSED [ 75%] 2023-03-31T05:55:27.0812867Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0813399Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0813847Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0814120Z PASSED [ 78%] 2023-03-31T05:55:27.0814630Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd_bwd STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0815154Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0815613Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0815887Z PASSED [ 81%] 2023-03-31T05:55:27.0816401Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_inputs_fwd_lazy STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0816926Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0817389Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0817663Z PASSED [ 84%] 2023-03-31T05:55:27.0818180Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_lazily_initialized STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0818754Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0819254Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0819529Z PASSED [ 87%] 2023-03-31T05:55:27.0820035Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_manual_optimizer_step STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0820575Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0821034Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0821305Z PASSED [ 90%] 2023-03-31T05:55:27.0821801Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_memory_timeline STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0822340Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0822800Z STAGE:2023-03-31 05:55:25 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0823159Z PASSED [ 93%] 2023-03-31T05:55:27.0823725Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_parameters_and_gradients STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0824272Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0824733Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0825175Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0825630Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0826097Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0826551Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0827034Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0827500Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0827951Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0828398Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0828849Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0829122Z PASSED [ 96%] 2023-03-31T05:55:27.0829672Z profiler/test_memory_profiler.py::TestMemoryProfilerE2E::test_parameters_and_gradients_set_to_none STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0830224Z [W CPUAllocator.cpp:235] Memory block of unknown size was allocated before the profiling started, profiler results will not include the deallocation event 2023-03-31T05:55:27.0830731Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0831193Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0831643Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T05:55:27.0832077Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T05:55:27.0832539Z STAGE:2023-03-31 05:55:26 3058:3058 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T05:55:27.0832860Z PASSED [100%] 2023-03-31T05:55:27.0833002Z 2023-03-31T05:55:27.0833401Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-f170e6cb75f486bf.xml - 2023-03-31T05:55:27.0833791Z ============================== 32 passed in 4.08s ============================== 2023-03-31T05:55:27.0834110Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:27.0834318Z 2023-03-31T05:55:27.0834687Z ##[endgroup] 2023-03-31T05:55:27.0835137Z FINISHED PRINTING LOG FILE of profiler/test_memory_profiler (/var/lib/jenkins/workspace/test/test-reports/profiler-test_memory_profiler_fshex0e3.log) 2023-03-31T05:55:27.0835403Z 2023-03-31T05:55:29.1814976Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:29.1845937Z Ignoring disabled issues: [] 2023-03-31T05:55:29.2019578Z Running test_weak ... [2023-03-31 05:55:29.201542] 2023-03-31T05:55:29.2022186Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_weak.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:29.201912] 2023-03-31T05:55:35.0826705Z 2023-03-31T05:55:35.0827252Z Expand the folded group to see the log file of test_weak 2023-03-31T05:55:35.0828320Z ##[group]PRINTING LOG FILE of test_weak (/var/lib/jenkins/workspace/test/test-reports/test_weak_c7fpnl6w.log) 2023-03-31T05:55:35.0829534Z Test results will be stored in test-reports/python-pytest/test_weak/test_weak-1e7688e5b758402f.xml 2023-03-31T05:55:35.0830061Z ============================= test session starts ============================== 2023-03-31T05:55:35.0830470Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:35.0830754Z cachedir: .pytest_cache 2023-03-31T05:55:35.0831204Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:35.0831584Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:35.0832038Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:35.0832456Z collecting ... collected 25 items 2023-03-31T05:55:35.0834426Z Running 25 items in this shard: test/test_weak.py::WeakTest::test_make_weak_keyed_dict_from_dict, test/test_weak.py::WeakTest::test_make_weak_keyed_dict_from_weak_keyed_dict, test/test_weak.py::WeakTest::test_make_weak_keyed_dict_repr, test/test_weak.py::WeakTest::test_threaded_weak_key_dict_copy, test/test_weak.py::WeakTest::test_threaded_weak_key_dict_deepcopy, test/test_weak.py::WeakTest::test_weak_keyed_bad_delitem, test/test_weak.py::WeakTest::test_weak_keyed_delitem, test/test_weak.py::WeakTest::test_weak_keyed_dict_popitem, test/test_weak.py::WeakTest::test_weak_keyed_dict_setdefault, test/test_weak.py::WeakTest::test_weak_keyed_dict_update, test/test_weak.py::WeakTest::test_weak_keyed_union_operators, test/test_weak.py::WeakKeyDictionaryTestCase::test_bool, test/test_weak.py::WeakKeyDictionaryTestCase::test_constructor, test/test_weak.py::WeakKeyDictionaryTestCase::test_get, test/test_weak.py::WeakKeyDictionaryTestCase::test_getitem, test/test_weak.py::WeakKeyDictionaryTestCase::test_items, test/test_weak.py::WeakKeyDictionaryTestCase::test_keys, test/test_weak.py::WeakKeyDictionaryTestCase::test_len, test/test_weak.py::WeakKeyDictionaryTestCase::test_pop, test/test_weak.py::WeakKeyDictionaryTestCase::test_popitem, test/test_weak.py::WeakKeyDictionaryTestCase::test_read, test/test_weak.py::WeakKeyDictionaryTestCase::test_setdefault, test/test_weak.py::WeakKeyDictionaryTestCase::test_update, test/test_weak.py::WeakKeyDictionaryTestCase::test_values, test/test_weak.py::WeakKeyDictionaryTestCase::test_write 2023-03-31T05:55:35.0836331Z 2023-03-31T05:55:35.0836487Z test_weak.py::WeakTest::test_make_weak_keyed_dict_from_dict PASSED [ 4%] 2023-03-31T05:55:35.0836898Z test_weak.py::WeakTest::test_make_weak_keyed_dict_from_weak_keyed_dict PASSED [ 8%] 2023-03-31T05:55:35.0837275Z test_weak.py::WeakTest::test_make_weak_keyed_dict_repr PASSED [ 12%] 2023-03-31T05:55:35.0837594Z test_weak.py::WeakTest::test_threaded_weak_key_dict_copy PASSED [ 16%] 2023-03-31T05:55:35.0837918Z test_weak.py::WeakTest::test_threaded_weak_key_dict_deepcopy PASSED [ 20%] 2023-03-31T05:55:35.0838221Z test_weak.py::WeakTest::test_weak_keyed_bad_delitem PASSED [ 24%] 2023-03-31T05:55:35.0838528Z test_weak.py::WeakTest::test_weak_keyed_delitem PASSED [ 28%] 2023-03-31T05:55:35.0838829Z test_weak.py::WeakTest::test_weak_keyed_dict_popitem PASSED [ 32%] 2023-03-31T05:55:35.0839142Z test_weak.py::WeakTest::test_weak_keyed_dict_setdefault PASSED [ 36%] 2023-03-31T05:55:35.0839434Z test_weak.py::WeakTest::test_weak_keyed_dict_update PASSED [ 40%] 2023-03-31T05:55:35.0839799Z test_weak.py::WeakTest::test_weak_keyed_union_operators SKIPPED (dict union not supported in this Python) [ 44%] 2023-03-31T05:55:35.0840167Z test_weak.py::WeakKeyDictionaryTestCase::test_bool PASSED [ 48%] 2023-03-31T05:55:35.0840490Z test_weak.py::WeakKeyDictionaryTestCase::test_constructor PASSED [ 52%] 2023-03-31T05:55:35.0840829Z test_weak.py::WeakKeyDictionaryTestCase::test_get PASSED [ 56%] 2023-03-31T05:55:35.0841198Z test_weak.py::WeakKeyDictionaryTestCase::test_getitem PASSED [ 60%] 2023-03-31T05:55:35.0841527Z test_weak.py::WeakKeyDictionaryTestCase::test_items PASSED [ 64%] 2023-03-31T05:55:35.0841842Z test_weak.py::WeakKeyDictionaryTestCase::test_keys PASSED [ 68%] 2023-03-31T05:55:35.0842163Z test_weak.py::WeakKeyDictionaryTestCase::test_len PASSED [ 72%] 2023-03-31T05:55:35.0842488Z test_weak.py::WeakKeyDictionaryTestCase::test_pop PASSED [ 76%] 2023-03-31T05:55:35.0842811Z test_weak.py::WeakKeyDictionaryTestCase::test_popitem PASSED [ 80%] 2023-03-31T05:55:35.0843447Z test_weak.py::WeakKeyDictionaryTestCase::test_read PASSED [ 84%] 2023-03-31T05:55:35.0843780Z test_weak.py::WeakKeyDictionaryTestCase::test_setdefault PASSED [ 88%] 2023-03-31T05:55:35.0844198Z test_weak.py::WeakKeyDictionaryTestCase::test_update PASSED [ 92%] 2023-03-31T05:55:35.0844516Z test_weak.py::WeakKeyDictionaryTestCase::test_values PASSED [ 96%] 2023-03-31T05:55:35.0844843Z test_weak.py::WeakKeyDictionaryTestCase::test_write PASSED [100%] 2023-03-31T05:55:35.0845023Z 2023-03-31T05:55:35.0866267Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_weak/test_weak-1e7688e5b758402f.xml - 2023-03-31T05:55:35.0866740Z ======================== 24 passed, 1 skipped in 3.87s ========================= 2023-03-31T05:55:35.0867083Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:35.0867299Z 2023-03-31T05:55:35.0867618Z ##[endgroup] 2023-03-31T05:55:35.0868089Z FINISHED PRINTING LOG FILE of test_weak (/var/lib/jenkins/workspace/test/test-reports/test_weak_c7fpnl6w.log) 2023-03-31T05:55:35.0868310Z 2023-03-31T05:55:37.1747889Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:37.1775756Z Ignoring disabled issues: [] 2023-03-31T05:55:37.1943335Z Running test_mkl_verbose ... [2023-03-31 05:55:37.193976] 2023-03-31T05:55:37.1946047Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_mkl_verbose.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:37.194331] 2023-03-31T05:55:42.7926992Z 2023-03-31T05:55:42.7927576Z Expand the folded group to see the log file of test_mkl_verbose 2023-03-31T05:55:42.7928669Z ##[group]PRINTING LOG FILE of test_mkl_verbose (/var/lib/jenkins/workspace/test/test-reports/test_mkl_verbose_udnnjy9e.log) 2023-03-31T05:55:42.7930137Z Test results will be stored in test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-4148f53ee5bf8b62.xml 2023-03-31T05:55:42.7930831Z ============================= test session starts ============================== 2023-03-31T05:55:42.7931564Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:42.7931873Z cachedir: .pytest_cache 2023-03-31T05:55:42.7932313Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:42.7932692Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:42.7933131Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:42.7933442Z collecting ... collected 2 items 2023-03-31T05:55:42.7933816Z Running 2 items in this shard: test/test_mkl_verbose.py::TestMKLVerbose::test_verbose_off, test/test_mkl_verbose.py::TestMKLVerbose::test_verbose_on 2023-03-31T05:55:42.7934073Z 2023-03-31T05:55:42.7934214Z test_mkl_verbose.py::TestMKLVerbose::test_verbose_off PASSED [ 50%] 2023-03-31T05:55:42.7934537Z test_mkl_verbose.py::TestMKLVerbose::test_verbose_on PASSED [100%] 2023-03-31T05:55:42.7934713Z 2023-03-31T05:55:42.7935056Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-4148f53ee5bf8b62.xml - 2023-03-31T05:55:42.7935508Z ============================== 2 passed in 3.65s =============================== 2023-03-31T05:55:42.7935827Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:42.7936032Z 2023-03-31T05:55:42.7936277Z ##[endgroup] 2023-03-31T05:55:42.7936698Z FINISHED PRINTING LOG FILE of test_mkl_verbose (/var/lib/jenkins/workspace/test/test-reports/test_mkl_verbose_udnnjy9e.log) 2023-03-31T05:55:42.7936929Z 2023-03-31T05:55:44.8852515Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:44.8880396Z Ignoring disabled issues: [] 2023-03-31T05:55:44.9050246Z Running dynamo/test_export ... [2023-03-31 05:55:44.904727] 2023-03-31T05:55:44.9052796Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_export.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:44.905040] 2023-03-31T05:55:50.6689009Z 2023-03-31T05:55:50.6689415Z Expand the folded group to see the log file of dynamo/test_export 2023-03-31T05:55:50.6690195Z ##[group]PRINTING LOG FILE of dynamo/test_export (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_export_c2oltxw7.log) 2023-03-31T05:55:50.6691005Z Test results will be stored in test-reports/python-pytest/dynamo.test_export/dynamo.test_export-c5d005f8365c4b22.xml 2023-03-31T05:55:50.6691445Z ============================= test session starts ============================== 2023-03-31T05:55:50.6691895Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:50.6692194Z cachedir: .pytest_cache 2023-03-31T05:55:50.6692662Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:50.6693073Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:50.6693593Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:50.6693913Z collecting ... collected 100 items 2023-03-31T05:55:50.6703968Z Running 100 items in this shard: test/dynamo/test_export.py::ExportTests::test_dict_return, test/dynamo/test_export.py::ExportTests::test_dict_return_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes, test/dynamo/test_export.py::ExportTests::test_dupes_2, test/dynamo/test_export.py::ExportTests::test_dupes_2_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_arg, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_arg_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_output, test/dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_output_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dupes_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_dynamic_slicing, test/dynamo/test_export.py::ExportTests::test_dynamic_slicing_invalid, test/dynamo/test_export.py::ExportTests::test_dynamic_slicing_simple, test/dynamo/test_export.py::ExportTests::test_export, test/dynamo/test_export.py::ExportTests::test_export_compare_optimize_with_make_fx, test/dynamo/test_export.py::ExportTests::test_export_cond_in_aten_symbolic, test/dynamo/test_export.py::ExportTests::test_export_control_flow_with_getattr, test/dynamo/test_export.py::ExportTests::test_export_decomp, test/dynamo/test_export.py::ExportTests::test_export_decomp_asserts_bad_args, test/dynamo/test_export.py::ExportTests::test_export_decomp_asserts_bad_args_mode, test/dynamo/test_export.py::ExportTests::test_export_dynamic_dim_cleanup, test/dynamo/test_export.py::ExportTests::test_export_dynamic_dim_not_1, test/dynamo/test_export.py::ExportTests::test_export_graph_bypass, test/dynamo/test_export.py::ExportTests::test_export_graph_bypass_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_graph_with_complex_reorder, test/dynamo/test_export.py::ExportTests::test_export_graph_with_complex_reorder_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_graph_with_list, test/dynamo/test_export.py::ExportTests::test_export_graph_with_list_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_identity, test/dynamo/test_export.py::ExportTests::test_export_mark_dynamic_conflict_dynamic_dim, test/dynamo/test_export.py::ExportTests::test_export_meta, test/dynamo/test_export.py::ExportTests::test_export_meta_val, test/dynamo/test_export.py::ExportTests::test_export_mismatched_out, test/dynamo/test_export.py::ExportTests::test_export_mismatched_out_2, test/dynamo/test_export.py::ExportTests::test_export_mismatched_out_2_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_mismatched_out_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_multi_dynamic_dim_constraint, test/dynamo/test_export.py::ExportTests::test_export_multi_dynamic_dim_safe_relationship, test/dynamo/test_export.py::ExportTests::test_export_no_raise, test/dynamo/test_export.py::ExportTests::test_export_no_raise_on_relationship, test/dynamo/test_export.py::ExportTests::test_export_pass_arg_by_name, test/dynamo/test_export.py::ExportTests::test_export_pass_arg_by_name_star_args, test/dynamo/test_export.py::ExportTests::test_export_preserves_nn_module_stack_for_get_attr, test/dynamo/test_export.py::ExportTests::test_export_raise_guard_full_constraint, test/dynamo/test_export.py::ExportTests::test_export_raise_guard_partial_constraint, test/dynamo/test_export.py::ExportTests::test_export_shape_control_flow_1, test/dynamo/test_export.py::ExportTests::test_export_specialized_int, test/dynamo/test_export.py::ExportTests::test_export_symbolic_shape, test/dynamo/test_export.py::ExportTests::test_export_with_args_and_empty_kwargs, test/dynamo/test_export.py::ExportTests::test_export_with_args_with_default_None, test/dynamo/test_export.py::ExportTests::test_export_with_args_with_default_float, test/dynamo/test_export.py::ExportTests::test_export_with_args_with_default_tensor, test/dynamo/test_export.py::ExportTests::test_export_with_args_with_default_tuple, test/dynamo/test_export.py::ExportTests::test_export_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_export_with_cond_dynamic_shape_pred, test/dynamo/test_export.py::ExportTests::test_export_with_constant_dict_values, test/dynamo/test_export.py::ExportTests::test_export_with_constant_free_function, test/dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method, test/dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method_multiarg, test/dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method_multiarg_diff, test/dynamo/test_export.py::ExportTests::test_export_with_constant_list_nonzero, test/dynamo/test_export.py::ExportTests::test_export_with_constant_list_nonzero_free_function, test/dynamo/test_export.py::ExportTests::test_export_with_constant_method_on_module, test/dynamo/test_export.py::ExportTests::test_export_with_constant_method_on_module_invoke_twice, test/dynamo/test_export.py::ExportTests::test_export_with_constant_none_control_flow, test/dynamo/test_export.py::ExportTests::test_export_with_constant_none_control_flow_free_func, test/dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow, test/dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow_free_func, test/dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow_pos, test/dynamo/test_export.py::ExportTests::test_export_with_constant_not_return_const, test/dynamo/test_export.py::ExportTests::test_export_with_constant_tuple_nonzero, test/dynamo/test_export.py::ExportTests::test_export_with_functools_wrapped_fn, test/dynamo/test_export.py::ExportTests::test_export_with_functools_wrapped_method, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs_and_empty_args, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_None, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_float, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_tensor, test/dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_tuple, test/dynamo/test_export.py::ExportTests::test_export_with_map_cond, test/dynamo/test_export.py::ExportTests::test_export_with_map_zero_sized_tensor, test/dynamo/test_export.py::ExportTests::test_export_with_module_layer, test/dynamo/test_export.py::ExportTests::test_export_with_stack_trace, test/dynamo/test_export.py::ExportTests::test_export_with_wrapped_fn, test/dynamo/test_export.py::ExportTests::test_func_return, test/dynamo/test_export.py::ExportTests::test_func_return_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_input_container_type, test/dynamo/test_export.py::ExportTests::test_list_contains, test/dynamo/test_export.py::ExportTests::test_list_not_contains, test/dynamo/test_export.py::ExportTests::test_list_unpack, test/dynamo/test_export.py::ExportTests::test_list_unpack_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_zeroes_in_and_out_different_shape_on_test, test/dynamo/test_export.py::ExportTests::test_zeroes_in_and_out_different_shape_on_test_with_aten_graph, test/dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out, test/dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out_permute, test/dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out_permute_dupe_and_bypass 2023-03-31T05:55:50.6713245Z 2023-03-31T05:55:50.6713465Z dynamo/test_export.py::ExportTests::test_dict_return PASSED [ 1%] 2023-03-31T05:55:50.6713809Z dynamo/test_export.py::ExportTests::test_dict_return_with_aten_graph PASSED [ 2%] 2023-03-31T05:55:50.6714209Z dynamo/test_export.py::ExportTests::test_dupes PASSED [ 3%] 2023-03-31T05:55:50.6714525Z dynamo/test_export.py::ExportTests::test_dupes_2 PASSED [ 4%] 2023-03-31T05:55:50.6714848Z dynamo/test_export.py::ExportTests::test_dupes_2_with_aten_graph PASSED [ 5%] 2023-03-31T05:55:50.6715258Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass PASSED [ 6%] 2023-03-31T05:55:50.6715679Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg PASSED [ 7%] 2023-03-31T05:55:50.6716099Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg_with_aten_graph PASSED [ 8%] 2023-03-31T05:55:50.6716539Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_aten_graph PASSED [ 9%] 2023-03-31T05:55:50.6716900Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_arg PASSED [ 10%] 2023-03-31T05:55:50.6717352Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_arg_with_aten_graph PASSED [ 11%] 2023-03-31T05:55:50.6717721Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_output PASSED [ 12%] 2023-03-31T05:55:50.6718171Z dynamo/test_export.py::ExportTests::test_dupes_and_bypass_with_non_tensor_output_with_aten_graph PASSED [ 13%] 2023-03-31T05:55:50.6718539Z dynamo/test_export.py::ExportTests::test_dupes_with_aten_graph PASSED [ 14%] 2023-03-31T05:55:50.6718926Z dynamo/test_export.py::ExportTests::test_dynamic_slicing PASSED [ 15%] 2023-03-31T05:55:50.6719241Z dynamo/test_export.py::ExportTests::test_dynamic_slicing_invalid PASSED [ 16%] 2023-03-31T05:55:50.6719636Z dynamo/test_export.py::ExportTests::test_dynamic_slicing_simple PASSED [ 17%] 2023-03-31T05:55:50.6719994Z dynamo/test_export.py::ExportTests::test_export PASSED [ 18%] 2023-03-31T05:55:50.6720387Z dynamo/test_export.py::ExportTests::test_export_compare_optimize_with_make_fx PASSED [ 19%] 2023-03-31T05:55:50.6720744Z dynamo/test_export.py::ExportTests::test_export_cond_in_aten_symbolic PASSED [ 20%] 2023-03-31T05:55:50.6721154Z dynamo/test_export.py::ExportTests::test_export_control_flow_with_getattr PASSED [ 21%] 2023-03-31T05:55:50.6721492Z dynamo/test_export.py::ExportTests::test_export_decomp PASSED [ 22%] 2023-03-31T05:55:50.6721820Z dynamo/test_export.py::ExportTests::test_export_decomp_asserts_bad_args PASSED [ 23%] 2023-03-31T05:55:50.6722232Z dynamo/test_export.py::ExportTests::test_export_decomp_asserts_bad_args_mode PASSED [ 24%] 2023-03-31T05:55:50.6722581Z dynamo/test_export.py::ExportTests::test_export_dynamic_dim_cleanup PASSED [ 25%] 2023-03-31T05:55:50.6722958Z dynamo/test_export.py::ExportTests::test_export_dynamic_dim_not_1 PASSED [ 26%] 2023-03-31T05:55:50.6723564Z dynamo/test_export.py::ExportTests::test_export_graph_bypass PASSED [ 27%] 2023-03-31T05:55:50.6723909Z dynamo/test_export.py::ExportTests::test_export_graph_bypass_with_aten_graph PASSED [ 28%] 2023-03-31T05:55:50.6724266Z dynamo/test_export.py::ExportTests::test_export_graph_with_complex_reorder PASSED [ 29%] 2023-03-31T05:55:50.6724633Z dynamo/test_export.py::ExportTests::test_export_graph_with_complex_reorder_with_aten_graph PASSED [ 30%] 2023-03-31T05:55:50.6724993Z dynamo/test_export.py::ExportTests::test_export_graph_with_list PASSED [ 31%] 2023-03-31T05:55:50.6725368Z dynamo/test_export.py::ExportTests::test_export_graph_with_list_with_aten_graph PASSED [ 32%] 2023-03-31T05:55:50.6725746Z dynamo/test_export.py::ExportTests::test_export_identity PASSED [ 33%] 2023-03-31T05:55:50.6726096Z dynamo/test_export.py::ExportTests::test_export_mark_dynamic_conflict_dynamic_dim PASSED [ 34%] 2023-03-31T05:55:50.6726439Z dynamo/test_export.py::ExportTests::test_export_meta PASSED [ 35%] 2023-03-31T05:55:50.6726756Z dynamo/test_export.py::ExportTests::test_export_meta_val PASSED [ 36%] 2023-03-31T05:55:50.6727067Z dynamo/test_export.py::ExportTests::test_export_mismatched_out PASSED [ 37%] 2023-03-31T05:55:50.6727387Z dynamo/test_export.py::ExportTests::test_export_mismatched_out_2 PASSED [ 38%] 2023-03-31T05:55:50.6727735Z dynamo/test_export.py::ExportTests::test_export_mismatched_out_2_with_aten_graph PASSED [ 39%] 2023-03-31T05:55:50.6728098Z dynamo/test_export.py::ExportTests::test_export_mismatched_out_with_aten_graph PASSED [ 40%] 2023-03-31T05:55:50.6728519Z dynamo/test_export.py::ExportTests::test_export_multi_dynamic_dim_constraint PASSED [ 41%] 2023-03-31T05:55:50.6728934Z dynamo/test_export.py::ExportTests::test_export_multi_dynamic_dim_safe_relationship PASSED [ 42%] 2023-03-31T05:55:50.6729348Z dynamo/test_export.py::ExportTests::test_export_no_raise PASSED [ 43%] 2023-03-31T05:55:50.6729678Z dynamo/test_export.py::ExportTests::test_export_no_raise_on_relationship PASSED [ 44%] 2023-03-31T05:55:50.6730020Z dynamo/test_export.py::ExportTests::test_export_pass_arg_by_name PASSED [ 45%] 2023-03-31T05:55:50.6730362Z dynamo/test_export.py::ExportTests::test_export_pass_arg_by_name_star_args PASSED [ 46%] 2023-03-31T05:55:50.6730727Z dynamo/test_export.py::ExportTests::test_export_preserves_nn_module_stack_for_get_attr PASSED [ 47%] 2023-03-31T05:55:50.6731080Z dynamo/test_export.py::ExportTests::test_export_raise_guard_full_constraint PASSED [ 48%] 2023-03-31T05:55:50.6731442Z dynamo/test_export.py::ExportTests::test_export_raise_guard_partial_constraint PASSED [ 49%] 2023-03-31T05:55:50.6731795Z dynamo/test_export.py::ExportTests::test_export_shape_control_flow_1 PASSED [ 50%] 2023-03-31T05:55:50.6732121Z dynamo/test_export.py::ExportTests::test_export_specialized_int PASSED [ 51%] 2023-03-31T05:55:50.6732451Z dynamo/test_export.py::ExportTests::test_export_symbolic_shape PASSED [ 52%] 2023-03-31T05:55:50.6732793Z dynamo/test_export.py::ExportTests::test_export_with_args_and_empty_kwargs PASSED [ 53%] 2023-03-31T05:55:50.6733189Z dynamo/test_export.py::ExportTests::test_export_with_args_with_default_None PASSED [ 54%] 2023-03-31T05:55:50.6733529Z dynamo/test_export.py::ExportTests::test_export_with_args_with_default_float PASSED [ 55%] 2023-03-31T05:55:50.6733936Z dynamo/test_export.py::ExportTests::test_export_with_args_with_default_tensor XFAIL [ 56%] 2023-03-31T05:55:50.6734302Z dynamo/test_export.py::ExportTests::test_export_with_args_with_default_tuple XFAIL [ 57%] 2023-03-31T05:55:50.6734632Z dynamo/test_export.py::ExportTests::test_export_with_aten_graph PASSED [ 58%] 2023-03-31T05:55:50.6734978Z dynamo/test_export.py::ExportTests::test_export_with_cond_dynamic_shape_pred PASSED [ 59%] 2023-03-31T05:55:50.6735332Z dynamo/test_export.py::ExportTests::test_export_with_constant_dict_values PASSED [ 60%] 2023-03-31T05:55:50.6735737Z dynamo/test_export.py::ExportTests::test_export_with_constant_free_function PASSED [ 61%] 2023-03-31T05:55:50.6736108Z dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method PASSED [ 62%] 2023-03-31T05:55:50.6736582Z dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method_multiarg PASSED [ 63%] 2023-03-31T05:55:50.6736997Z dynamo/test_export.py::ExportTests::test_export_with_constant_free_function_and_class_method_multiarg_diff PASSED [ 64%] 2023-03-31T05:55:50.6737380Z dynamo/test_export.py::ExportTests::test_export_with_constant_list_nonzero PASSED [ 65%] 2023-03-31T05:55:50.6737743Z dynamo/test_export.py::ExportTests::test_export_with_constant_list_nonzero_free_function PASSED [ 66%] 2023-03-31T05:55:50.6738115Z dynamo/test_export.py::ExportTests::test_export_with_constant_method_on_module PASSED [ 67%] 2023-03-31T05:55:50.6738489Z dynamo/test_export.py::ExportTests::test_export_with_constant_method_on_module_invoke_twice PASSED [ 68%] 2023-03-31T05:55:50.6738855Z dynamo/test_export.py::ExportTests::test_export_with_constant_none_control_flow PASSED [ 69%] 2023-03-31T05:55:50.6739232Z dynamo/test_export.py::ExportTests::test_export_with_constant_none_control_flow_free_func PASSED [ 70%] 2023-03-31T05:55:50.6739606Z dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow PASSED [ 71%] 2023-03-31T05:55:50.6739990Z dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow_free_func PASSED [ 72%] 2023-03-31T05:55:50.6740363Z dynamo/test_export.py::ExportTests::test_export_with_constant_not_none_control_flow_pos PASSED [ 73%] 2023-03-31T05:55:50.6740777Z dynamo/test_export.py::ExportTests::test_export_with_constant_not_return_const PASSED [ 74%] 2023-03-31T05:55:50.6741169Z dynamo/test_export.py::ExportTests::test_export_with_constant_tuple_nonzero PASSED [ 75%] 2023-03-31T05:55:50.6741524Z dynamo/test_export.py::ExportTests::test_export_with_functools_wrapped_fn PASSED [ 76%] 2023-03-31T05:55:50.6741871Z dynamo/test_export.py::ExportTests::test_export_with_functools_wrapped_method PASSED [ 77%] 2023-03-31T05:55:50.6742217Z dynamo/test_export.py::ExportTests::test_export_with_kwargs PASSED [ 78%] 2023-03-31T05:55:50.6742557Z dynamo/test_export.py::ExportTests::test_export_with_kwargs_and_empty_args PASSED [ 79%] 2023-03-31T05:55:50.6742900Z dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_None PASSED [ 80%] 2023-03-31T05:55:50.6743358Z dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_float PASSED [ 81%] 2023-03-31T05:55:50.6743717Z dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_tensor XFAIL [ 82%] 2023-03-31T05:55:50.6744072Z dynamo/test_export.py::ExportTests::test_export_with_kwargs_with_default_tuple XFAIL [ 83%] 2023-03-31T05:55:50.6744402Z dynamo/test_export.py::ExportTests::test_export_with_map_cond PASSED [ 84%] 2023-03-31T05:55:50.6744749Z dynamo/test_export.py::ExportTests::test_export_with_map_zero_sized_tensor PASSED [ 85%] 2023-03-31T05:55:50.6745092Z dynamo/test_export.py::ExportTests::test_export_with_module_layer PASSED [ 86%] 2023-03-31T05:55:50.6745450Z dynamo/test_export.py::ExportTests::test_export_with_stack_trace PASSED [ 87%] 2023-03-31T05:55:50.6745780Z dynamo/test_export.py::ExportTests::test_export_with_wrapped_fn PASSED [ 88%] 2023-03-31T05:55:50.6746095Z dynamo/test_export.py::ExportTests::test_func_return PASSED [ 89%] 2023-03-31T05:55:50.6746423Z dynamo/test_export.py::ExportTests::test_func_return_with_aten_graph PASSED [ 90%] 2023-03-31T05:55:50.6746744Z dynamo/test_export.py::ExportTests::test_input_container_type PASSED [ 91%] 2023-03-31T05:55:50.6747070Z dynamo/test_export.py::ExportTests::test_list_contains PASSED [ 92%] 2023-03-31T05:55:50.6747387Z dynamo/test_export.py::ExportTests::test_list_not_contains PASSED [ 93%] 2023-03-31T05:55:50.6747688Z dynamo/test_export.py::ExportTests::test_list_unpack PASSED [ 94%] 2023-03-31T05:55:50.6748052Z dynamo/test_export.py::ExportTests::test_list_unpack_with_aten_graph PASSED [ 95%] 2023-03-31T05:55:50.6748412Z dynamo/test_export.py::ExportTests::test_zeroes_in_and_out_different_shape_on_test PASSED [ 96%] 2023-03-31T05:55:50.6748800Z dynamo/test_export.py::ExportTests::test_zeroes_in_and_out_different_shape_on_test_with_aten_graph PASSED [ 97%] 2023-03-31T05:55:50.6749160Z dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out PASSED [ 98%] 2023-03-31T05:55:50.6749529Z dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out_permute PASSED [ 99%] 2023-03-31T05:55:50.6749969Z dynamo/test_export.py::ExportTests::test_zeroes_in_new_shape_scalar_out_permute_dupe_and_bypass PASSED [100%] 2023-03-31T05:55:50.6750189Z 2023-03-31T05:55:50.6750616Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_export/dynamo.test_export-c5d005f8365c4b22.xml - 2023-03-31T05:55:50.6750988Z ======================== 96 passed, 4 xfailed in 3.57s ========================= 2023-03-31T05:55:50.6751312Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:50.6751516Z 2023-03-31T05:55:50.6751839Z ##[endgroup] 2023-03-31T05:55:50.6752253Z FINISHED PRINTING LOG FILE of dynamo/test_export (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_export_c2oltxw7.log) 2023-03-31T05:55:50.6752491Z 2023-03-31T05:55:52.8482272Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:55:52.8509538Z Ignoring disabled issues: [] 2023-03-31T05:55:52.8681372Z Running dynamo/test_aot_autograd ... [2023-03-31 05:55:52.867734] 2023-03-31T05:55:52.8684530Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_aot_autograd.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:55:52.868115] 2023-03-31T05:55:58.3246563Z 2023-03-31T05:55:58.3247170Z Expand the folded group to see the log file of dynamo/test_aot_autograd 2023-03-31T05:55:58.3248250Z ##[group]PRINTING LOG FILE of dynamo/test_aot_autograd (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_aot_autograd_uwtoio9u.log) 2023-03-31T05:55:58.3249301Z Test results will be stored in test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-3e275992b40418f7.xml 2023-03-31T05:55:58.3249841Z ============================= test session starts ============================== 2023-03-31T05:55:58.3250251Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:55:58.3250518Z cachedir: .pytest_cache 2023-03-31T05:55:58.3250962Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:55:58.3251348Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:55:58.3251789Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:55:58.3252104Z collecting ... collected 18 items 2023-03-31T05:55:58.3254618Z Running 18 items in this shard: test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_LSTM, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg_list, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_with_global, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_metadata_mutation_on_input_causes_recompile, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_safe, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe_control_flow, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_double_backward_errors, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_multiple_aot_autograd_calls_dupe_args, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation1, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_requires_grad_fake_via_dynamo_recompiles 2023-03-31T05:55:58.3256914Z 2023-03-31T05:55:58.3257094Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_LSTM PASSED [ 5%] 2023-03-31T05:55:58.3257489Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles PASSED [ 11%] 2023-03-31T05:55:58.3257928Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args PASSED [ 16%] 2023-03-31T05:55:58.3258376Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param PASSED [ 22%] 2023-03-31T05:55:58.3258841Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg PASSED [ 27%] 2023-03-31T05:55:58.3259362Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg_list PASSED [ 33%] 2023-03-31T05:55:58.3259876Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_with_global PASSED [ 38%] 2023-03-31T05:55:58.3260329Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_metadata_mutation_on_input_causes_recompile PASSED [ 44%] 2023-03-31T05:55:58.3260768Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_safe PASSED [ 50%] 2023-03-31T05:55:58.3261182Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe PASSED [ 55%] 2023-03-31T05:55:58.3261626Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe_control_flow PASSED [ 61%] 2023-03-31T05:55:58.3262050Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_double_backward_errors PASSED [ 66%] 2023-03-31T05:55:58.3262473Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_multiple_aot_autograd_calls_dupe_args PASSED [ 72%] 2023-03-31T05:55:58.3262858Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation PASSED [ 77%] 2023-03-31T05:55:58.3263329Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation1 PASSED [ 83%] 2023-03-31T05:55:58.3263716Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing PASSED [ 88%] 2023-03-31T05:55:58.3264155Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing_mutation PASSED [ 94%] 2023-03-31T05:55:58.3264564Z dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_requires_grad_fake_via_dynamo_recompiles PASSED [100%] 2023-03-31T05:55:58.3264799Z 2023-03-31T05:55:58.3265187Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-3e275992b40418f7.xml - 2023-03-31T05:55:58.3265581Z ============================== 18 passed in 3.34s ============================== 2023-03-31T05:55:58.3265889Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:55:58.3266093Z 2023-03-31T05:55:58.3266352Z ##[endgroup] 2023-03-31T05:55:58.3266804Z FINISHED PRINTING LOG FILE of dynamo/test_aot_autograd (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_aot_autograd_uwtoio9u.log) 2023-03-31T05:55:58.3267107Z 2023-03-31T05:56:00.4188664Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:00.4216465Z Ignoring disabled issues: [] 2023-03-31T05:56:00.4382562Z Running dynamo/test_logging ... [2023-03-31 05:56:00.437889] 2023-03-31T05:56:00.4384326Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_logging.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:00.438202] 2023-03-31T05:56:10.5878251Z 2023-03-31T05:56:10.5878787Z Expand the folded group to see the log file of dynamo/test_logging 2023-03-31T05:56:10.5879934Z ##[group]PRINTING LOG FILE of dynamo/test_logging (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_logging_tjdflc9z.log) 2023-03-31T05:56:10.5880855Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:10.5881727Z Test results will be stored in test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-0ff7e4912c9e147c.xml 2023-03-31T05:56:10.5882350Z ============================= test session starts ============================== 2023-03-31T05:56:10.5883276Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:10.5883759Z cachedir: .pytest_cache 2023-03-31T05:56:10.5884546Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:10.5885191Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:10.5886231Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:10.5886841Z collecting ... collected 17 items 2023-03-31T05:56:10.5889200Z Running 17 items in this shard: test/dynamo/test_logging.py::LoggingTests::test_aot, test/dynamo/test_logging.py::LoggingTests::test_aot_graphs, test/dynamo/test_logging.py::LoggingTests::test_aot_joint_graph, test/dynamo/test_logging.py::LoggingTests::test_bytecode, test/dynamo/test_logging.py::LoggingTests::test_dynamo_debug, test/dynamo/test_logging.py::LoggingTests::test_dynamo_error, test/dynamo/test_logging.py::LoggingTests::test_dynamo_info, test/dynamo/test_logging.py::LoggingTests::test_graph, test/dynamo/test_logging.py::LoggingTests::test_graph_code, test/dynamo/test_logging.py::LoggingTests::test_guards, test/dynamo/test_logging.py::LoggingTests::test_inductor_debug, test/dynamo/test_logging.py::LoggingTests::test_inductor_error, test/dynamo/test_logging.py::LoggingTests::test_inductor_info, test/dynamo/test_logging.py::LoggingTests::test_open_registration, test/dynamo/test_logging.py::LoggingTests::test_open_registration_with_registered_parent, test/dynamo/test_logging.py::LoggingTests::test_output_code, test/dynamo/test_logging.py::LoggingTests::test_schedule 2023-03-31T05:56:10.5891549Z 2023-03-31T05:56:10.5892286Z dynamo/test_logging.py::LoggingTests::test_aot <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 5%] 2023-03-31T05:56:10.5893537Z dynamo/test_logging.py::LoggingTests::test_aot_graphs <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 11%] 2023-03-31T05:56:10.5894718Z dynamo/test_logging.py::LoggingTests::test_aot_joint_graph <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 17%] 2023-03-31T05:56:10.5895887Z dynamo/test_logging.py::LoggingTests::test_bytecode <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 23%] 2023-03-31T05:56:10.5897043Z dynamo/test_logging.py::LoggingTests::test_dynamo_debug <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 29%] 2023-03-31T05:56:10.5898330Z dynamo/test_logging.py::LoggingTests::test_dynamo_error <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 35%] 2023-03-31T05:56:10.5899498Z dynamo/test_logging.py::LoggingTests::test_dynamo_info <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 41%] 2023-03-31T05:56:10.5900646Z dynamo/test_logging.py::LoggingTests::test_graph <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 47%] 2023-03-31T05:56:10.5901797Z dynamo/test_logging.py::LoggingTests::test_graph_code <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 52%] 2023-03-31T05:56:10.5903005Z dynamo/test_logging.py::LoggingTests::test_guards <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 58%] 2023-03-31T05:56:10.5904318Z dynamo/test_logging.py::LoggingTests::test_inductor_debug <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 64%] 2023-03-31T05:56:10.5905564Z dynamo/test_logging.py::LoggingTests::test_inductor_error <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 70%] 2023-03-31T05:56:10.5906799Z dynamo/test_logging.py::LoggingTests::test_inductor_info <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 76%] 2023-03-31T05:56:10.5908147Z dynamo/test_logging.py::LoggingTests::test_open_registration <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 82%] 2023-03-31T05:56:10.5909495Z dynamo/test_logging.py::LoggingTests::test_open_registration_with_registered_parent <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 88%] 2023-03-31T05:56:10.5910753Z dynamo/test_logging.py::LoggingTests::test_output_code <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py PASSED [ 94%] 2023-03-31T05:56:10.5911969Z dynamo/test_logging.py::LoggingTests::test_schedule <- ../../../../opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/logging_utils.py SKIPPED (requires cuda) [100%] 2023-03-31T05:56:10.5912487Z 2023-03-31T05:56:10.5913120Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-0ff7e4912c9e147c.xml - 2023-03-31T05:56:10.5913800Z ======================== 16 passed, 1 skipped in 3.26s ========================= 2023-03-31T05:56:10.5914372Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:10.5914746Z 2023-03-31T05:56:10.5915194Z ##[endgroup] 2023-03-31T05:56:10.5915944Z FINISHED PRINTING LOG FILE of dynamo/test_logging (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_logging_tjdflc9z.log) 2023-03-31T05:56:10.5916390Z 2023-03-31T05:56:12.7644599Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:12.7677590Z Ignoring disabled issues: [] 2023-03-31T05:56:12.7844743Z Running test_import_stats ... [2023-03-31 05:56:12.784066] 2023-03-31T05:56:12.7846656Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_import_stats.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:12.784425] 2023-03-31T05:56:17.8097503Z 2023-03-31T05:56:17.8098344Z Expand the folded group to see the log file of test_import_stats 2023-03-31T05:56:17.8099454Z ##[group]PRINTING LOG FILE of test_import_stats (/var/lib/jenkins/workspace/test/test-reports/test_import_stats_69phv1hk.log) 2023-03-31T05:56:17.8100510Z Test results will be stored in test-reports/python-pytest/test_import_stats/test_import_stats-9f6b211aaa7d9826.xml 2023-03-31T05:56:17.8101353Z ============================= test session starts ============================== 2023-03-31T05:56:17.8102067Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:17.8102598Z cachedir: .pytest_cache 2023-03-31T05:56:17.8103423Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:17.8104398Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:17.8105249Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:17.8105846Z collecting ... collected 2 items 2023-03-31T05:56:17.8106613Z Running 2 items in this shard: test/test_import_stats.py::TestImportTime::test_time_cuda_device_count, test/test_import_stats.py::TestImportTime::test_time_import_torch 2023-03-31T05:56:17.8107156Z 2023-03-31T05:56:17.8107462Z test_import_stats.py::TestImportTime::test_time_cuda_device_count PASSED [ 50%] 2023-03-31T05:56:17.8108095Z test_import_stats.py::TestImportTime::test_time_import_torch PASSED [100%] 2023-03-31T05:56:17.8108446Z 2023-03-31T05:56:17.8109137Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_import_stats/test_import_stats-9f6b211aaa7d9826.xml - 2023-03-31T05:56:17.8109852Z ============================== 2 passed in 3.10s =============================== 2023-03-31T05:56:17.8110436Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:17.8110830Z 2023-03-31T05:56:17.8111315Z ##[endgroup] 2023-03-31T05:56:17.8112287Z FINISHED PRINTING LOG FILE of test_import_stats (/var/lib/jenkins/workspace/test/test-reports/test_import_stats_69phv1hk.log) 2023-03-31T05:56:17.8112810Z 2023-03-31T05:56:19.8471277Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:19.8498592Z Ignoring disabled issues: [] 2023-03-31T05:56:19.8663601Z Running test_mkldnn_verbose ... [2023-03-31 05:56:19.865991] 2023-03-31T05:56:19.8665187Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_mkldnn_verbose.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:19.866279] 2023-03-31T05:56:24.8935815Z 2023-03-31T05:56:24.8936344Z Expand the folded group to see the log file of test_mkldnn_verbose 2023-03-31T05:56:24.8937350Z ##[group]PRINTING LOG FILE of test_mkldnn_verbose (/var/lib/jenkins/workspace/test/test-reports/test_mkldnn_verbose_jlawt3ge.log) 2023-03-31T05:56:24.8938458Z Test results will be stored in test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-7df56168c4f02b69.xml 2023-03-31T05:56:24.8939110Z ============================= test session starts ============================== 2023-03-31T05:56:24.8939862Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:24.8940361Z cachedir: .pytest_cache 2023-03-31T05:56:24.8941440Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:24.8942028Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:24.8942568Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:24.8946933Z collecting ... collected 2 items 2023-03-31T05:56:24.8947786Z Running 2 items in this shard: test/test_mkldnn_verbose.py::TestMKLDNNVerbose::test_verbose_off, test/test_mkldnn_verbose.py::TestMKLDNNVerbose::test_verbose_on 2023-03-31T05:56:24.8948090Z 2023-03-31T05:56:24.8948258Z test_mkldnn_verbose.py::TestMKLDNNVerbose::test_verbose_off PASSED [ 50%] 2023-03-31T05:56:24.8948595Z test_mkldnn_verbose.py::TestMKLDNNVerbose::test_verbose_on PASSED [100%] 2023-03-31T05:56:24.8948781Z 2023-03-31T05:56:24.8949220Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-7df56168c4f02b69.xml - 2023-03-31T05:56:24.8965404Z ============================== 2 passed in 3.10s =============================== 2023-03-31T05:56:24.8966062Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:24.8966465Z 2023-03-31T05:56:24.8967040Z ##[endgroup] 2023-03-31T05:56:24.8967934Z FINISHED PRINTING LOG FILE of test_mkldnn_verbose (/var/lib/jenkins/workspace/test/test-reports/test_mkldnn_verbose_jlawt3ge.log) 2023-03-31T05:56:24.8968411Z 2023-03-31T05:56:26.9625303Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:26.9653607Z Ignoring disabled issues: [] 2023-03-31T05:56:26.9820224Z Running dynamo/test_modules ... [2023-03-31 05:56:26.981704] 2023-03-31T05:56:26.9822485Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_modules.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:26.982023] 2023-03-31T05:56:32.2104476Z 2023-03-31T05:56:32.2105117Z Expand the folded group to see the log file of dynamo/test_modules 2023-03-31T05:56:32.2106412Z ##[group]PRINTING LOG FILE of dynamo/test_modules (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_modules_6eq1t1zk.log) 2023-03-31T05:56:32.2107465Z Test results will be stored in test-reports/python-pytest/dynamo.test_modules/dynamo.test_modules-e361d6bdc2cfdd69.xml 2023-03-31T05:56:32.2108102Z ============================= test session starts ============================== 2023-03-31T05:56:32.2108987Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:32.2109540Z cachedir: .pytest_cache 2023-03-31T05:56:32.2110347Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:32.2110975Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:32.2111434Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:32.2111742Z collecting ... collected 64 items 2023-03-31T05:56:32.2117403Z Running 64 items in this shard: test/dynamo/test_modules.py::NNModuleTests::test_access_by_keys, test/dynamo/test_modules.py::NNModuleTests::test_basicmodule1, test/dynamo/test_modules.py::NNModuleTests::test_basicmodule2, test/dynamo/test_modules.py::NNModuleTests::test_call_fn_with_non_const_inputs_safe, test/dynamo/test_modules.py::NNModuleTests::test_cfgmod, test/dynamo/test_modules.py::NNModuleTests::test_children, test/dynamo/test_modules.py::NNModuleTests::test_constloop, test/dynamo/test_modules.py::NNModuleTests::test_densenet, test/dynamo/test_modules.py::NNModuleTests::test_enumvalues, test/dynamo/test_modules.py::NNModuleTests::test_fnmember, test/dynamo/test_modules.py::NNModuleTests::test_fnmembercmp1, test/dynamo/test_modules.py::NNModuleTests::test_fnmembercmp2, test/dynamo/test_modules.py::NNModuleTests::test_forward_directly, test/dynamo/test_modules.py::NNModuleTests::test_generation_tag, test/dynamo/test_modules.py::NNModuleTests::test_hasattr, test/dynamo/test_modules.py::NNModuleTests::test_intarg, test/dynamo/test_modules.py::NNModuleTests::test_iseval1, test/dynamo/test_modules.py::NNModuleTests::test_iseval2, test/dynamo/test_modules.py::NNModuleTests::test_isnonelayer, test/dynamo/test_modules.py::NNModuleTests::test_istraining1, test/dynamo/test_modules.py::NNModuleTests::test_istraining2, test/dynamo/test_modules.py::NNModuleTests::test_layerlist, test/dynamo/test_modules.py::NNModuleTests::test_lazy_module, test/dynamo/test_modules.py::NNModuleTests::test_module_attribute_precedence, test/dynamo/test_modules.py::NNModuleTests::test_module_call_module_with_static_forward, test/dynamo/test_modules.py::NNModuleTests::test_module_class_method, test/dynamo/test_modules.py::NNModuleTests::test_module_forward_has_graph_break, test/dynamo/test_modules.py::NNModuleTests::test_module_guard_name_is_valid, test/dynamo/test_modules.py::NNModuleTests::test_module_name_string, test/dynamo/test_modules.py::NNModuleTests::test_module_property, test/dynamo/test_modules.py::NNModuleTests::test_module_static_method, test/dynamo/test_modules.py::NNModuleTests::test_moduledict, test/dynamo/test_modules.py::NNModuleTests::test_modulelist, test/dynamo/test_modules.py::NNModuleTests::test_modulemethod1, test/dynamo/test_modules.py::NNModuleTests::test_modulemethod2, test/dynamo/test_modules.py::NNModuleTests::test_named_children, test/dynamo/test_modules.py::NNModuleTests::test_nn_moduledict_contains, test/dynamo/test_modules.py::NNModuleTests::test_parameters1, test/dynamo/test_modules.py::NNModuleTests::test_parameters2, test/dynamo/test_modules.py::NNModuleTests::test_parameters3, test/dynamo/test_modules.py::NNModuleTests::test_self_mutating1, test/dynamo/test_modules.py::NNModuleTests::test_seq, test/dynamo/test_modules.py::NNModuleTests::test_simple_torch_function, test/dynamo/test_modules.py::NNModuleTests::test_stringmember, test/dynamo/test_modules.py::NNModuleTests::test_submodules1, test/dynamo/test_modules.py::NNModuleTests::test_submodules2, test/dynamo/test_modules.py::NNModuleTests::test_super1, test/dynamo/test_modules.py::NNModuleTests::test_super2, test/dynamo/test_modules.py::NNModuleTests::test_super_class_method, test/dynamo/test_modules.py::NNModuleTests::test_tensorlist, test/dynamo/test_modules.py::NNModuleTests::test_torch_function_with_closure, test/dynamo/test_modules.py::NNModuleTests::test_unsupportedmethod, test/dynamo/test_modules.py::NNModuleTests::test_unsupportedmodule, test/dynamo/test_modules.py::NNModuleTests::test_viamodulecall, test/dynamo/test_modules.py::OptimizedModuleTest::test_attr, test/dynamo/test_modules.py::OptimizedModuleTest::test_composition, test/dynamo/test_modules.py::OptimizedModuleTest::test_composition_with_opt_mod, test/dynamo/test_modules.py::OptimizedModuleTest::test_hooks_inner, test/dynamo/test_modules.py::OptimizedModuleTest::test_hooks_outer, test/dynamo/test_modules.py::OptimizedModuleTest::test_hooks_skip_guards, test/dynamo/test_modules.py::OptimizedModuleTest::test_module_patch, test/dynamo/test_modules.py::OptimizedModuleTest::test_nn_module, test/dynamo/test_modules.py::OptimizedModuleTest::test_recursion, test/dynamo/test_modules.py::OptimizedModuleTest::test_to 2023-03-31T05:56:32.2122134Z 2023-03-31T05:56:32.2122288Z dynamo/test_modules.py::NNModuleTests::test_access_by_keys PASSED [ 1%] 2023-03-31T05:56:32.2122631Z dynamo/test_modules.py::NNModuleTests::test_basicmodule1 PASSED [ 3%] 2023-03-31T05:56:32.2122964Z dynamo/test_modules.py::NNModuleTests::test_basicmodule2 PASSED [ 4%] 2023-03-31T05:56:32.2123692Z dynamo/test_modules.py::NNModuleTests::test_call_fn_with_non_const_inputs_safe PASSED [ 6%] 2023-03-31T05:56:32.2124164Z dynamo/test_modules.py::NNModuleTests::test_cfgmod PASSED [ 7%] 2023-03-31T05:56:32.2125002Z dynamo/test_modules.py::NNModuleTests::test_children PASSED [ 9%] 2023-03-31T05:56:32.2125499Z dynamo/test_modules.py::NNModuleTests::test_constloop PASSED [ 10%] 2023-03-31T05:56:32.2126190Z dynamo/test_modules.py::NNModuleTests::test_densenet PASSED [ 12%] 2023-03-31T05:56:32.2126827Z dynamo/test_modules.py::NNModuleTests::test_enumvalues PASSED [ 14%] 2023-03-31T05:56:32.2127322Z dynamo/test_modules.py::NNModuleTests::test_fnmember PASSED [ 15%] 2023-03-31T05:56:32.2127643Z dynamo/test_modules.py::NNModuleTests::test_fnmembercmp1 PASSED [ 17%] 2023-03-31T05:56:32.2127964Z dynamo/test_modules.py::NNModuleTests::test_fnmembercmp2 PASSED [ 18%] 2023-03-31T05:56:32.2128297Z dynamo/test_modules.py::NNModuleTests::test_forward_directly PASSED [ 20%] 2023-03-31T05:56:32.2128610Z dynamo/test_modules.py::NNModuleTests::test_generation_tag PASSED [ 21%] 2023-03-31T05:56:32.2128926Z dynamo/test_modules.py::NNModuleTests::test_hasattr PASSED [ 23%] 2023-03-31T05:56:32.2129308Z dynamo/test_modules.py::NNModuleTests::test_intarg PASSED [ 25%] 2023-03-31T05:56:32.2129607Z dynamo/test_modules.py::NNModuleTests::test_iseval1 PASSED [ 26%] 2023-03-31T05:56:32.2129917Z dynamo/test_modules.py::NNModuleTests::test_iseval2 PASSED [ 28%] 2023-03-31T05:56:32.2130225Z dynamo/test_modules.py::NNModuleTests::test_isnonelayer PASSED [ 29%] 2023-03-31T05:56:32.2130543Z dynamo/test_modules.py::NNModuleTests::test_istraining1 PASSED [ 31%] 2023-03-31T05:56:32.2130844Z dynamo/test_modules.py::NNModuleTests::test_istraining2 PASSED [ 32%] 2023-03-31T05:56:32.2131168Z dynamo/test_modules.py::NNModuleTests::test_layerlist PASSED [ 34%] 2023-03-31T05:56:32.2131799Z dynamo/test_modules.py::NNModuleTests::test_lazy_module [2023-03-31 05:56:29,451] torch._dynamo.output_graph: [WARNING] nn.Module hooks are not fully supported, they may be ignored 2023-03-31T05:56:32.2132149Z PASSED [ 35%] 2023-03-31T05:56:32.2132417Z dynamo/test_modules.py::NNModuleTests::test_module_attribute_precedence PASSED [ 37%] 2023-03-31T05:56:32.2132792Z dynamo/test_modules.py::NNModuleTests::test_module_call_module_with_static_forward PASSED [ 39%] 2023-03-31T05:56:32.2133151Z dynamo/test_modules.py::NNModuleTests::test_module_class_method PASSED [ 40%] 2023-03-31T05:56:32.2133486Z dynamo/test_modules.py::NNModuleTests::test_module_forward_has_graph_break PASSED [ 42%] 2023-03-31T05:56:32.2133844Z dynamo/test_modules.py::NNModuleTests::test_module_guard_name_is_valid PASSED [ 43%] 2023-03-31T05:56:32.2166183Z dynamo/test_modules.py::NNModuleTests::test_module_name_string PASSED [ 45%] 2023-03-31T05:56:32.2166963Z dynamo/test_modules.py::NNModuleTests::test_module_property PASSED [ 46%] 2023-03-31T05:56:32.2167654Z dynamo/test_modules.py::NNModuleTests::test_module_static_method PASSED [ 48%] 2023-03-31T05:56:32.2168307Z dynamo/test_modules.py::NNModuleTests::test_moduledict PASSED [ 50%] 2023-03-31T05:56:32.2168931Z dynamo/test_modules.py::NNModuleTests::test_modulelist PASSED [ 51%] 2023-03-31T05:56:32.2169728Z dynamo/test_modules.py::NNModuleTests::test_modulemethod1 PASSED [ 53%] 2023-03-31T05:56:32.2170359Z dynamo/test_modules.py::NNModuleTests::test_modulemethod2 PASSED [ 54%] 2023-03-31T05:56:32.2170990Z dynamo/test_modules.py::NNModuleTests::test_named_children PASSED [ 56%] 2023-03-31T05:56:32.2171622Z dynamo/test_modules.py::NNModuleTests::test_nn_moduledict_contains PASSED [ 57%] 2023-03-31T05:56:32.2172269Z dynamo/test_modules.py::NNModuleTests::test_parameters1 PASSED [ 59%] 2023-03-31T05:56:32.2172905Z dynamo/test_modules.py::NNModuleTests::test_parameters2 PASSED [ 60%] 2023-03-31T05:56:32.2173527Z dynamo/test_modules.py::NNModuleTests::test_parameters3 PASSED [ 62%] 2023-03-31T05:56:32.2174170Z dynamo/test_modules.py::NNModuleTests::test_self_mutating1 PASSED [ 64%] 2023-03-31T05:56:32.2174781Z dynamo/test_modules.py::NNModuleTests::test_seq PASSED [ 65%] 2023-03-31T05:56:32.2175550Z dynamo/test_modules.py::NNModuleTests::test_simple_torch_function PASSED [ 67%] 2023-03-31T05:56:32.2176184Z dynamo/test_modules.py::NNModuleTests::test_stringmember PASSED [ 68%] 2023-03-31T05:56:32.2176810Z dynamo/test_modules.py::NNModuleTests::test_submodules1 PASSED [ 70%] 2023-03-31T05:56:32.2177431Z dynamo/test_modules.py::NNModuleTests::test_submodules2 PASSED [ 71%] 2023-03-31T05:56:32.2178026Z dynamo/test_modules.py::NNModuleTests::test_super1 PASSED [ 73%] 2023-03-31T05:56:32.2178629Z dynamo/test_modules.py::NNModuleTests::test_super2 PASSED [ 75%] 2023-03-31T05:56:32.2179260Z dynamo/test_modules.py::NNModuleTests::test_super_class_method PASSED [ 76%] 2023-03-31T05:56:32.2179889Z dynamo/test_modules.py::NNModuleTests::test_tensorlist PASSED [ 78%] 2023-03-31T05:56:32.2180548Z dynamo/test_modules.py::NNModuleTests::test_torch_function_with_closure PASSED [ 79%] 2023-03-31T05:56:32.2181320Z dynamo/test_modules.py::NNModuleTests::test_unsupportedmethod PASSED [ 81%] 2023-03-31T05:56:32.2181995Z dynamo/test_modules.py::NNModuleTests::test_unsupportedmodule PASSED [ 82%] 2023-03-31T05:56:32.2182634Z dynamo/test_modules.py::NNModuleTests::test_viamodulecall PASSED [ 84%] 2023-03-31T05:56:32.2183275Z dynamo/test_modules.py::OptimizedModuleTest::test_attr PASSED [ 85%] 2023-03-31T05:56:32.2183939Z dynamo/test_modules.py::OptimizedModuleTest::test_composition PASSED [ 87%] 2023-03-31T05:56:32.2184784Z dynamo/test_modules.py::OptimizedModuleTest::test_composition_with_opt_mod PASSED [ 89%] 2023-03-31T05:56:32.2185521Z dynamo/test_modules.py::OptimizedModuleTest::test_hooks_inner PASSED [ 90%] 2023-03-31T05:56:32.2186199Z dynamo/test_modules.py::OptimizedModuleTest::test_hooks_outer PASSED [ 92%] 2023-03-31T05:56:32.2186882Z dynamo/test_modules.py::OptimizedModuleTest::test_hooks_skip_guards PASSED [ 93%] 2023-03-31T05:56:32.2187572Z dynamo/test_modules.py::OptimizedModuleTest::test_module_patch PASSED [ 95%] 2023-03-31T05:56:32.2188215Z dynamo/test_modules.py::OptimizedModuleTest::test_nn_module PASSED [ 96%] 2023-03-31T05:56:32.2188884Z dynamo/test_modules.py::OptimizedModuleTest::test_recursion PASSED [ 98%] 2023-03-31T05:56:32.2189558Z dynamo/test_modules.py::OptimizedModuleTest::test_to PASSED [100%] 2023-03-31T05:56:32.2189891Z 2023-03-31T05:56:32.2190658Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_modules/dynamo.test_modules-e361d6bdc2cfdd69.xml - 2023-03-31T05:56:32.2191496Z ============================== 64 passed in 3.09s ============================== 2023-03-31T05:56:32.2192172Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:32.2192570Z 2023-03-31T05:56:32.2193217Z ##[endgroup] 2023-03-31T05:56:32.2194051Z FINISHED PRINTING LOG FILE of dynamo/test_modules (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_modules_6eq1t1zk.log) 2023-03-31T05:56:32.2194526Z 2023-03-31T05:56:34.3385239Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:34.3412563Z Ignoring disabled issues: [] 2023-03-31T05:56:34.3578201Z Running functorch/test_control_flow ... [2023-03-31 05:56:34.357445] 2023-03-31T05:56:34.3580210Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'functorch/test_control_flow.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:34.357773] 2023-03-31T05:56:39.3778218Z 2023-03-31T05:56:39.3778824Z Expand the folded group to see the log file of functorch/test_control_flow 2023-03-31T05:56:39.3779953Z ##[group]PRINTING LOG FILE of functorch/test_control_flow (/var/lib/jenkins/workspace/test/test-reports/functorch-test_control_flow_rnmpf4z1.log) 2023-03-31T05:56:39.3781081Z Test results will be stored in test-reports/python-pytest/functorch.test_control_flow/functorch.test_control_flow-d5e8f048f0857b56.xml 2023-03-31T05:56:39.3782003Z ============================= test session starts ============================== 2023-03-31T05:56:39.3782725Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:39.3783206Z cachedir: .pytest_cache 2023-03-31T05:56:39.3783977Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:39.3784564Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:39.3785391Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:39.3785975Z collecting ... collected 33 items 2023-03-31T05:56:39.3794448Z Running 33 items in this shard: test/functorch/test_control_flow.py::TestControlFlow::test_cond_gpu, test/functorch/test_control_flow.py::TestControlFlow::test_cond_no_trace, test/functorch/test_control_flow.py::TestControlFlow::test_map_gpu, test/functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_tensor_size, test/functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_tensor_size_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_type_size, test/functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_type_size_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_data_dependent_pred, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_false_branch, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_true_branch, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested_input_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_output_alias_input, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_retrace_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_sym_pred, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_arg_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_alias, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_real, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_symbolic, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_cond_map_cond_symbolic, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_real, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_symbolic 2023-03-31T05:56:39.3802586Z 2023-03-31T05:56:39.3802997Z functorch/test_control_flow.py::TestControlFlow::test_cond_gpu SKIPPED (Test requires CUDA.) [ 3%] 2023-03-31T05:56:39.3804080Z functorch/test_control_flow.py::TestControlFlow::test_cond_no_trace PASSED [ 6%] 2023-03-31T05:56:39.3804898Z functorch/test_control_flow.py::TestControlFlow::test_map_gpu SKIPPED (Test requires CUDA.) [ 9%] 2023-03-31T05:56:39.3805747Z functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_tensor_size PASSED [ 12%] 2023-03-31T05:56:39.3806666Z functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_tensor_size_fake_tensor PASSED [ 15%] 2023-03-31T05:56:39.3807632Z functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_type_size PASSED [ 18%] 2023-03-31T05:56:39.3808544Z functorch/test_control_flow.py::TestControlFlowTraced::test_assert_on_mismatch_type_size_fake_tensor PASSED [ 21%] 2023-03-31T05:56:39.3809430Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized PASSED [ 24%] 2023-03-31T05:56:39.3810398Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_data_dependent_pred PASSED [ 27%] 2023-03-31T05:56:39.3811390Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_false_branch PASSED [ 30%] 2023-03-31T05:56:39.3812384Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_true_branch PASSED [ 33%] 2023-03-31T05:56:39.3813294Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested PASSED [ 36%] 2023-03-31T05:56:39.3814213Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested_input_mutation PASSED [ 39%] 2023-03-31T05:56:39.3815173Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_output_alias_input PASSED [ 42%] 2023-03-31T05:56:39.3816066Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced PASSED [ 45%] 2023-03-31T05:56:39.3816906Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_fake_tensor PASSED [ 48%] 2023-03-31T05:56:39.3817787Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi PASSED [ 51%] 2023-03-31T05:56:39.3818674Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi_fake_tensor PASSED [ 54%] 2023-03-31T05:56:39.3819568Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs PASSED [ 57%] 2023-03-31T05:56:39.3820470Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs_fake_tensor PASSED [ 60%] 2023-03-31T05:56:39.3821466Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_retrace_functionalized PASSED [ 63%] 2023-03-31T05:56:39.3822401Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested PASSED [ 66%] 2023-03-31T05:56:39.3823280Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested_fake_tensor PASSED [ 69%] 2023-03-31T05:56:39.3824194Z functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_sym_pred PASSED [ 72%] 2023-03-31T05:56:39.3825040Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized PASSED [ 75%] 2023-03-31T05:56:39.3825921Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_arg_mutation PASSED [ 78%] 2023-03-31T05:56:39.3826804Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_alias PASSED [ 81%] 2023-03-31T05:56:39.3827715Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_mutation PASSED [ 84%] 2023-03-31T05:56:39.3828560Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_real PASSED [ 87%] 2023-03-31T05:56:39.3829358Z functorch/test_control_flow.py::TestControlFlowTraced::test_map_symbolic PASSED [ 90%] 2023-03-31T05:56:39.3830179Z functorch/test_control_flow.py::TestControlFlowTraced::test_nested_cond_map_cond_symbolic PASSED [ 93%] 2023-03-31T05:56:39.3831052Z functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_real PASSED [ 96%] 2023-03-31T05:56:39.3831993Z functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_symbolic PASSED [100%] 2023-03-31T05:56:39.3832455Z 2023-03-31T05:56:39.3833352Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_control_flow/functorch.test_control_flow-d5e8f048f0857b56.xml - 2023-03-31T05:56:39.3834193Z ======================== 31 passed, 2 skipped in 3.02s ========================= 2023-03-31T05:56:39.3834883Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:39.3835327Z 2023-03-31T05:56:39.3835904Z ##[endgroup] 2023-03-31T05:56:39.3836889Z FINISHED PRINTING LOG FILE of functorch/test_control_flow (/var/lib/jenkins/workspace/test/test-reports/functorch-test_control_flow_rnmpf4z1.log) 2023-03-31T05:56:39.3837457Z 2023-03-31T05:56:41.4453458Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:41.4481187Z Ignoring disabled issues: [] 2023-03-31T05:56:41.4649002Z Running dynamo/test_subgraphs ... [2023-03-31 05:56:41.464512] 2023-03-31T05:56:41.4650673Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_subgraphs.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:41.464818] 2023-03-31T05:56:46.5220484Z 2023-03-31T05:56:46.5221055Z Expand the folded group to see the log file of dynamo/test_subgraphs 2023-03-31T05:56:46.5222225Z ##[group]PRINTING LOG FILE of dynamo/test_subgraphs (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_subgraphs_icrdd_w_.log) 2023-03-31T05:56:46.5223522Z Test results will be stored in test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-b95643fb9ab252db.xml 2023-03-31T05:56:46.5224093Z ============================= test session starts ============================== 2023-03-31T05:56:46.5224782Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:46.5225235Z cachedir: .pytest_cache 2023-03-31T05:56:46.5225945Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:46.5226586Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:46.5227443Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:46.5228076Z collecting ... collected 44 items 2023-03-31T05:56:46.5235294Z Running 44 items in this shard: test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call1, test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call2, test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call3, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow1, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow2, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow3, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow4, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow5, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_duck_size, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_kwarg, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_order_dependence, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_shapes, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_zero_inference, test/dynamo/test_subgraphs.py::SubGraphTests::test_enumerate_not_break_graph, test/dynamo/test_subgraphs.py::SubGraphTests::test_extended_args, test/dynamo/test_subgraphs.py::SubGraphTests::test_graph_break_on_item, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported1, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported2, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported3, test/dynamo/test_subgraphs.py::SubGraphTests::test_multigraph, test/dynamo/test_subgraphs.py::SubGraphTests::test_no_graph_break_on_item, test/dynamo/test_subgraphs.py::SubGraphTests::test_pop_after_resume, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_range, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_range_iter, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_state, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume1, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume2, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume3, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume4, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume5, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_freevars, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_paths_join, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_tuple_iterator, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad1, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad2, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad3, test/dynamo/test_subgraphs.py::SubGraphTests::test_stack_state1, test/dynamo/test_subgraphs.py::SubGraphTests::test_stack_state2, test/dynamo/test_subgraphs.py::SubGraphTests::test_start1, test/dynamo/test_subgraphs.py::SubGraphTests::test_start2, test/dynamo/test_subgraphs.py::SubGraphTests::test_start3, test/dynamo/test_subgraphs.py::SubGraphTests::test_start4, test/dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_mutate, test/dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_return 2023-03-31T05:56:46.5241676Z 2023-03-31T05:56:46.5242027Z dynamo/test_subgraphs.py::SubGraphTests::test_capi_call1 PASSED [ 2%] 2023-03-31T05:56:46.5242649Z dynamo/test_subgraphs.py::SubGraphTests::test_capi_call2 PASSED [ 4%] 2023-03-31T05:56:46.5243550Z dynamo/test_subgraphs.py::SubGraphTests::test_capi_call3 PASSED [ 6%] 2023-03-31T05:56:46.5244104Z dynamo/test_subgraphs.py::SubGraphTests::test_control_flow1 PASSED [ 9%] 2023-03-31T05:56:46.5244735Z dynamo/test_subgraphs.py::SubGraphTests::test_control_flow2 PASSED [ 11%] 2023-03-31T05:56:46.5245315Z dynamo/test_subgraphs.py::SubGraphTests::test_control_flow3 PASSED [ 13%] 2023-03-31T05:56:46.5245918Z dynamo/test_subgraphs.py::SubGraphTests::test_control_flow4 PASSED [ 15%] 2023-03-31T05:56:46.5246529Z dynamo/test_subgraphs.py::SubGraphTests::test_control_flow5 PASSED [ 18%] 2023-03-31T05:56:46.5247120Z dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_duck_size PASSED [ 20%] 2023-03-31T05:56:46.5247717Z dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_kwarg PASSED [ 22%] 2023-03-31T05:56:46.5248432Z dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_order_dependence PASSED [ 25%] 2023-03-31T05:56:46.5249125Z dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_shapes PASSED [ 27%] 2023-03-31T05:56:46.5249660Z dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_zero_inference PASSED [ 29%] 2023-03-31T05:56:46.5250293Z dynamo/test_subgraphs.py::SubGraphTests::test_enumerate_not_break_graph PASSED [ 31%] 2023-03-31T05:56:46.5250880Z dynamo/test_subgraphs.py::SubGraphTests::test_extended_args PASSED [ 34%] 2023-03-31T05:56:46.5251430Z dynamo/test_subgraphs.py::SubGraphTests::test_graph_break_on_item PASSED [ 36%] 2023-03-31T05:56:46.5251965Z dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported1 PASSED [ 38%] 2023-03-31T05:56:46.5252491Z dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported2 PASSED [ 40%] 2023-03-31T05:56:46.5253047Z dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported3 PASSED [ 43%] 2023-03-31T05:56:46.5253615Z dynamo/test_subgraphs.py::SubGraphTests::test_multigraph PASSED [ 45%] 2023-03-31T05:56:46.5254174Z dynamo/test_subgraphs.py::SubGraphTests::test_no_graph_break_on_item PASSED [ 47%] 2023-03-31T05:56:46.5254724Z dynamo/test_subgraphs.py::SubGraphTests::test_pop_after_resume PASSED [ 50%] 2023-03-31T05:56:46.5255309Z dynamo/test_subgraphs.py::SubGraphTests::test_restore_range PASSED [ 52%] 2023-03-31T05:56:46.5255847Z dynamo/test_subgraphs.py::SubGraphTests::test_restore_range_iter PASSED [ 54%] 2023-03-31T05:56:46.5256547Z dynamo/test_subgraphs.py::SubGraphTests::test_restore_state PASSED [ 56%] 2023-03-31T05:56:46.5257106Z dynamo/test_subgraphs.py::SubGraphTests::test_resume1 PASSED [ 59%] 2023-03-31T05:56:46.5257686Z dynamo/test_subgraphs.py::SubGraphTests::test_resume2 PASSED [ 61%] 2023-03-31T05:56:46.5258240Z dynamo/test_subgraphs.py::SubGraphTests::test_resume3 PASSED [ 63%] 2023-03-31T05:56:46.5258772Z dynamo/test_subgraphs.py::SubGraphTests::test_resume4 PASSED [ 65%] 2023-03-31T05:56:46.5259288Z dynamo/test_subgraphs.py::SubGraphTests::test_resume5 PASSED [ 68%] 2023-03-31T05:56:46.5259821Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_freevars PASSED [ 70%] 2023-03-31T05:56:46.5260356Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_paths_join PASSED [ 72%] 2023-03-31T05:56:46.5261132Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_tuple_iterator PASSED [ 75%] 2023-03-31T05:56:46.5261718Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad1 PASSED [ 77%] 2023-03-31T05:56:46.5262283Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad2 PASSED [ 79%] 2023-03-31T05:56:46.5262833Z dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad3 PASSED [ 81%] 2023-03-31T05:56:46.5263676Z dynamo/test_subgraphs.py::SubGraphTests::test_stack_state1 PASSED [ 84%] 2023-03-31T05:56:46.5264225Z dynamo/test_subgraphs.py::SubGraphTests::test_stack_state2 PASSED [ 86%] 2023-03-31T05:56:46.5264724Z dynamo/test_subgraphs.py::SubGraphTests::test_start1 PASSED [ 88%] 2023-03-31T05:56:46.5265252Z dynamo/test_subgraphs.py::SubGraphTests::test_start2 PASSED [ 90%] 2023-03-31T05:56:46.5265750Z dynamo/test_subgraphs.py::SubGraphTests::test_start3 PASSED [ 93%] 2023-03-31T05:56:46.5266208Z dynamo/test_subgraphs.py::SubGraphTests::test_start4 PASSED [ 95%] 2023-03-31T05:56:46.5266705Z dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_mutate SKIPPED (not working yet) [ 97%] 2023-03-31T05:56:46.5267275Z dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_return PASSED [100%] 2023-03-31T05:56:46.5267585Z 2023-03-31T05:56:46.5268273Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-b95643fb9ab252db.xml - 2023-03-31T05:56:46.5268910Z ======================== 43 passed, 1 skipped in 2.94s ========================= 2023-03-31T05:56:46.5269559Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:46.5270050Z 2023-03-31T05:56:46.5270587Z ##[endgroup] 2023-03-31T05:56:46.5271370Z FINISHED PRINTING LOG FILE of dynamo/test_subgraphs (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_subgraphs_icrdd_w_.log) 2023-03-31T05:56:46.5271825Z 2023-03-31T05:56:48.5850186Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:48.5878100Z Ignoring disabled issues: [] 2023-03-31T05:56:48.6048254Z Running test_maskedtensor ... [2023-03-31 05:56:48.604421] 2023-03-31T05:56:48.6049931Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_maskedtensor.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:48.604737] 2023-03-31T05:56:54.4834900Z 2023-03-31T05:56:54.4835505Z Expand the folded group to see the log file of test_maskedtensor 2023-03-31T05:56:54.4836549Z ##[group]PRINTING LOG FILE of test_maskedtensor (/var/lib/jenkins/workspace/test/test-reports/test_maskedtensor_6gkgygrf.log) 2023-03-31T05:56:54.4839459Z Test results will be stored in test-reports/python-pytest/test_maskedtensor/test_maskedtensor-53d6bd0ec99cdfe0.xml 2023-03-31T05:56:54.4840182Z ============================= test session starts ============================== 2023-03-31T05:56:54.4840935Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:56:54.4841666Z cachedir: .pytest_cache 2023-03-31T05:56:54.4842444Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:56:54.4843325Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:56:54.4844071Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:56:54.4844628Z collecting ... collected 832 items 2023-03-31T05:56:54.4955901Z Running 832 items in this shard: test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_abs_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_absolute_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_acos_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_acosh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arccos_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arccosh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arcsin_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arcsinh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arctan_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arctanh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_asin_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_asinh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_atan_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_atanh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_bitwise_not_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_ceil_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_clamp_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_clip_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_conj_physical_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_cos_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_cosh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_deg2rad_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_digamma_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erf_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erfc_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erfinv_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_exp2_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_exp_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_expm1_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_fix_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_floor_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_frac_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_i0_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_lgamma_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log10_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log1p_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log2_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_logit_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_nan_to_num_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_neg_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_negative_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_pow_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_rad2deg_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_reciprocal_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_round_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_rsqrt_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sgn_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sigmoid_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sign_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sin_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sinc_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sinh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sqrt_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_square_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_tan_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_tanh_, test/test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_trunc_, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_abs, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_absolute, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_acos, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_acosh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_angle, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arccos, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arccosh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arcsin, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arcsinh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arctan, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_arctanh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_asin, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_asinh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_atan, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_atanh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_bitwise_not, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_ceil, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_clamp, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_clip, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_conj_physical, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_cos, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_cosh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_deg2rad, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_digamma, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_erf, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_erfc, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_erfinv, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_exp, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_exp2, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_expm1, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_fix, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_floor, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_frac, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_i0, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_isnan, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_lgamma, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_log, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_log10, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_log1p, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_log2, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_logit, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_nan_to_num, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_neg, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_negative, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_positive, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_pow, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_rad2deg, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_reciprocal, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_round, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_rsqrt, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sgn, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sigmoid, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sign, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_signbit, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sin, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sinc, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sinh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_sqrt, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_square, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_tan, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_tanh, test/test_maskedtensor.py::TestUnary::test_unary_fn_aten_trunc, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_add, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_arctan2, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_atan2, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_and, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_left_shift, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_or, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_right_shift, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_xor, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_div, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_divide, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_eq, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_floor_divide, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmax, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmin, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmod, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_ge, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_greater, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_greater_equal, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_gt, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_le, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_less, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_less_equal, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_logaddexp, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_logaddexp2, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_lt, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_maximum, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_minimum, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_mul, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_multiply, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_ne, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_nextafter, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_not_equal, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_remainder, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_sub, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_subtract, test/test_maskedtensor.py::TestBinary::test_binary_fn_aten_true_divide, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_add_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_arctan2_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_atan2_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_and_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_left_shift_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_or_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_right_shift_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_xor_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_div_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_divide_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_eq_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_floor_divide_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_fmod_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_ge_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_greater_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_greater_equal_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_gt_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_le_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_less_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_less_equal_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_lt_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_mul_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_multiply_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_ne_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_nextafter_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_not_equal_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_remainder_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_sub_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_subtract_, test/test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_true_divide_, test/test_maskedtensor.py::TestBinary::test_masks_match_fn_name_add, test/test_maskedtensor.py::TestBinary::test_masks_match_fn_name_add_, test/test_maskedtensor.py::TestReductions::test_all, test/test_maskedtensor.py::TestReductions::test_amax, test/test_maskedtensor.py::TestReductions::test_amax_grad, test/test_maskedtensor.py::TestReductions::test_amin, test/test_maskedtensor.py::TestReductions::test_amin_grad, test/test_maskedtensor.py::TestReductions::test_grad_dtype, test/test_maskedtensor.py::TestReductions::test_max_not_implemented, test/test_maskedtensor.py::TestReductions::test_mean, test/test_maskedtensor.py::TestReductions::test_mean_dim_grad, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1a, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1b, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1c, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1d, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1e, test/test_maskedtensor.py::TestReductions::test_mean_grad_case_1f, test/test_maskedtensor.py::TestReductions::test_prod, test/test_maskedtensor.py::TestReductions::test_prod_grad, test/test_maskedtensor.py::TestReductions::test_sum, test/test_maskedtensor.py::TestReductions::test_sum_grad, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float16, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_coo_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_coo_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_csr_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_csr_cpu_float64, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_strided_cpu_float32, test/test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_strided_cpu_float64, test/test_maskedtensor.py::TestBasicsCPU::test_add_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_contiguous_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_diff_dim_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_diff_layouts_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_diff_sizes_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_grad_warning_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_coo_values_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_csr_values_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_layout_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_invalid_tensor_inputs_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_softmax_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_to_dense_and_sparse_coo_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_to_dense_and_sparse_csr_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_to_dense_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_to_sparse_cpu, test/test_maskedtensor.py::TestBasicsCPU::test_where_cpu 2023-03-31T05:56:54.5045698Z 2023-03-31T05:56:54.5046390Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_abs_ PASSED [ 0%] 2023-03-31T05:56:54.5046774Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_absolute_ PASSED [ 0%] 2023-03-31T05:56:54.5047116Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_acos_ PASSED [ 0%] 2023-03-31T05:56:54.5047458Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_acosh_ PASSED [ 0%] 2023-03-31T05:56:54.5047803Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arccos_ PASSED [ 0%] 2023-03-31T05:56:54.5048148Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arccosh_ PASSED [ 0%] 2023-03-31T05:56:54.5048477Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arcsin_ PASSED [ 0%] 2023-03-31T05:56:54.5049055Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arcsinh_ PASSED [ 0%] 2023-03-31T05:56:54.5049450Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arctan_ PASSED [ 1%] 2023-03-31T05:56:54.5049776Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_arctanh_ PASSED [ 1%] 2023-03-31T05:56:54.5050114Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_asin_ PASSED [ 1%] 2023-03-31T05:56:54.5050454Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_asinh_ PASSED [ 1%] 2023-03-31T05:56:54.5050789Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_atan_ PASSED [ 1%] 2023-03-31T05:56:54.5051109Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_atanh_ PASSED [ 1%] 2023-03-31T05:56:54.5051453Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_bitwise_not_ PASSED [ 1%] 2023-03-31T05:56:54.5051797Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_ceil_ PASSED [ 1%] 2023-03-31T05:56:54.5052121Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_clamp_ PASSED [ 2%] 2023-03-31T05:56:54.5052456Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_clip_ PASSED [ 2%] 2023-03-31T05:56:54.5052803Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_conj_physical_ PASSED [ 2%] 2023-03-31T05:56:54.5053153Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_cos_ PASSED [ 2%] 2023-03-31T05:56:54.5053516Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_cosh_ PASSED [ 2%] 2023-03-31T05:56:54.5053855Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_deg2rad_ PASSED [ 2%] 2023-03-31T05:56:54.5054198Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_digamma_ PASSED [ 2%] 2023-03-31T05:56:54.5054517Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erf_ PASSED [ 2%] 2023-03-31T05:56:54.5054844Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erfc_ PASSED [ 3%] 2023-03-31T05:56:54.5055182Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_erfinv_ PASSED [ 3%] 2023-03-31T05:56:54.5055521Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_exp2_ PASSED [ 3%] 2023-03-31T05:56:54.5055833Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_exp_ PASSED [ 3%] 2023-03-31T05:56:54.5056215Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_expm1_ PASSED [ 3%] 2023-03-31T05:56:54.5056550Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_fix_ PASSED [ 3%] 2023-03-31T05:56:54.5056876Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_floor_ PASSED [ 3%] 2023-03-31T05:56:54.5057208Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_frac_ PASSED [ 3%] 2023-03-31T05:56:54.5057533Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_i0_ PASSED [ 3%] 2023-03-31T05:56:54.5057865Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_lgamma_ PASSED [ 4%] 2023-03-31T05:56:54.5058193Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log10_ PASSED [ 4%] 2023-03-31T05:56:54.5058528Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log1p_ PASSED [ 4%] 2023-03-31T05:56:54.5058860Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log2_ PASSED [ 4%] 2023-03-31T05:56:54.5059176Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_log_ PASSED [ 4%] 2023-03-31T05:56:54.5059504Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_logit_ PASSED [ 4%] 2023-03-31T05:56:54.5059849Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_nan_to_num_ PASSED [ 4%] 2023-03-31T05:56:54.5060184Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_neg_ PASSED [ 4%] 2023-03-31T05:56:54.5060510Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_negative_ PASSED [ 5%] 2023-03-31T05:56:54.5060847Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_pow_ PASSED [ 5%] 2023-03-31T05:56:54.5061181Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_rad2deg_ PASSED [ 5%] 2023-03-31T05:56:54.5061546Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_reciprocal_ PASSED [ 5%] 2023-03-31T05:56:54.5061917Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_round_ PASSED [ 5%] 2023-03-31T05:56:54.5062252Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_rsqrt_ PASSED [ 5%] 2023-03-31T05:56:54.5062585Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sgn_ PASSED [ 5%] 2023-03-31T05:56:54.5062909Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sigmoid_ PASSED [ 5%] 2023-03-31T05:56:54.5063244Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sign_ PASSED [ 6%] 2023-03-31T05:56:54.5063658Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sin_ PASSED [ 6%] 2023-03-31T05:56:54.5063976Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sinc_ PASSED [ 6%] 2023-03-31T05:56:54.5064304Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sinh_ PASSED [ 6%] 2023-03-31T05:56:54.5064637Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_sqrt_ PASSED [ 6%] 2023-03-31T05:56:54.5064975Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_square_ PASSED [ 6%] 2023-03-31T05:56:54.5065294Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_tan_ PASSED [ 6%] 2023-03-31T05:56:54.5065623Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_tanh_ PASSED [ 6%] 2023-03-31T05:56:54.5065952Z test_maskedtensor.py::TestUnary::test_inplace_unary_fn_aten_trunc_ PASSED [ 6%] 2023-03-31T05:56:54.5066306Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_abs PASSED [ 7%] 2023-03-31T05:56:54.5066625Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_absolute PASSED [ 7%] 2023-03-31T05:56:54.5066945Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_acos PASSED [ 7%] 2023-03-31T05:56:54.5067269Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_acosh PASSED [ 7%] 2023-03-31T05:56:54.5067573Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_angle PASSED [ 7%] 2023-03-31T05:56:54.5067888Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arccos PASSED [ 7%] 2023-03-31T05:56:54.5068203Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arccosh PASSED [ 7%] 2023-03-31T05:56:54.5068505Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arcsin PASSED [ 7%] 2023-03-31T05:56:54.5068856Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arcsinh PASSED [ 8%] 2023-03-31T05:56:54.5069174Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arctan PASSED [ 8%] 2023-03-31T05:56:54.5069486Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_arctanh PASSED [ 8%] 2023-03-31T05:56:54.5069787Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_asin PASSED [ 8%] 2023-03-31T05:56:54.5070096Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_asinh PASSED [ 8%] 2023-03-31T05:56:54.5070405Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_atan PASSED [ 8%] 2023-03-31T05:56:54.5070708Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_atanh PASSED [ 8%] 2023-03-31T05:56:54.5071025Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_bitwise_not PASSED [ 8%] 2023-03-31T05:56:54.5071345Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_ceil PASSED [ 9%] 2023-03-31T05:56:54.5071660Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_clamp PASSED [ 9%] 2023-03-31T05:56:54.5071963Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_clip PASSED [ 9%] 2023-03-31T05:56:54.5072278Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_conj_physical PASSED [ 9%] 2023-03-31T05:56:54.5072592Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_cos PASSED [ 9%] 2023-03-31T05:56:54.5072887Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_cosh PASSED [ 9%] 2023-03-31T05:56:54.5073197Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_deg2rad PASSED [ 9%] 2023-03-31T05:56:54.5073542Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_digamma PASSED [ 9%] 2023-03-31T05:56:54.5073884Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_erf PASSED [ 9%] 2023-03-31T05:56:54.5074179Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_erfc PASSED [ 10%] 2023-03-31T05:56:54.5074491Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_erfinv PASSED [ 10%] 2023-03-31T05:56:54.5074802Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_exp PASSED [ 10%] 2023-03-31T05:56:54.5075104Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_exp2 PASSED [ 10%] 2023-03-31T05:56:54.5075415Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_expm1 PASSED [ 10%] 2023-03-31T05:56:54.5075726Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_fix PASSED [ 10%] 2023-03-31T05:56:54.5076022Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_floor PASSED [ 10%] 2023-03-31T05:56:54.5076336Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_frac PASSED [ 10%] 2023-03-31T05:56:54.5076645Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_i0 PASSED [ 11%] 2023-03-31T05:56:54.5076960Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_isnan PASSED [ 11%] 2023-03-31T05:56:54.5077264Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_lgamma PASSED [ 11%] 2023-03-31T05:56:54.5077576Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_log PASSED [ 11%] 2023-03-31T05:56:54.5077917Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_log10 PASSED [ 11%] 2023-03-31T05:56:54.5078218Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_log1p PASSED [ 11%] 2023-03-31T05:56:54.5078529Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_log2 PASSED [ 11%] 2023-03-31T05:56:54.5078836Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_logit PASSED [ 11%] 2023-03-31T05:56:54.5079146Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_nan_to_num PASSED [ 12%] 2023-03-31T05:56:54.5079448Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_neg PASSED [ 12%] 2023-03-31T05:56:54.5079765Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_negative PASSED [ 12%] 2023-03-31T05:56:54.5080084Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_positive PASSED [ 12%] 2023-03-31T05:56:54.5080420Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_pow PASSED [ 12%] 2023-03-31T05:56:54.5080733Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_rad2deg PASSED [ 12%] 2023-03-31T05:56:54.5081051Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_reciprocal PASSED [ 12%] 2023-03-31T05:56:54.5081369Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_round PASSED [ 12%] 2023-03-31T05:56:54.5081669Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_rsqrt PASSED [ 12%] 2023-03-31T05:56:54.5081977Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sgn PASSED [ 13%] 2023-03-31T05:56:54.5082289Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sigmoid PASSED [ 13%] 2023-03-31T05:56:54.5082589Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sign PASSED [ 13%] 2023-03-31T05:56:54.5082901Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_signbit PASSED [ 13%] 2023-03-31T05:56:54.5083428Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sin PASSED [ 13%] 2023-03-31T05:56:54.5083742Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sinc PASSED [ 13%] 2023-03-31T05:56:54.5084040Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sinh PASSED [ 13%] 2023-03-31T05:56:54.5084347Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_sqrt PASSED [ 13%] 2023-03-31T05:56:54.5084657Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_square PASSED [ 14%] 2023-03-31T05:56:54.5084952Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_tan PASSED [ 14%] 2023-03-31T05:56:54.5085259Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_tanh PASSED [ 14%] 2023-03-31T05:56:54.5085644Z test_maskedtensor.py::TestUnary::test_unary_fn_aten_trunc PASSED [ 14%] 2023-03-31T05:56:54.5085997Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_add PASSED [ 14%] 2023-03-31T05:56:54.5086305Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_arctan2 PASSED [ 14%] 2023-03-31T05:56:54.5086623Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_atan2 PASSED [ 14%] 2023-03-31T05:56:54.5086945Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_and PASSED [ 14%] 2023-03-31T05:56:54.5087272Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_left_shift PASSED [ 15%] 2023-03-31T05:56:54.5087612Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_or PASSED [ 15%] 2023-03-31T05:56:54.5087950Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_right_shift PASSED [ 15%] 2023-03-31T05:56:54.5088290Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_bitwise_xor PASSED [ 15%] 2023-03-31T05:56:54.5088597Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_div PASSED [ 15%] 2023-03-31T05:56:54.5088912Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_divide PASSED [ 15%] 2023-03-31T05:56:54.5089225Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_eq PASSED [ 15%] 2023-03-31T05:56:54.5089539Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_floor_divide PASSED [ 15%] 2023-03-31T05:56:54.5089864Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmax PASSED [ 15%] 2023-03-31T05:56:54.5090232Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmin PASSED [ 16%] 2023-03-31T05:56:54.5090546Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_fmod PASSED [ 16%] 2023-03-31T05:56:54.5090845Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_ge PASSED [ 16%] 2023-03-31T05:56:54.5091155Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_greater PASSED [ 16%] 2023-03-31T05:56:54.5091491Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_greater_equal PASSED [ 16%] 2023-03-31T05:56:54.5091806Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_gt PASSED [ 16%] 2023-03-31T05:56:54.5092119Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_le PASSED [ 16%] 2023-03-31T05:56:54.5092434Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_less PASSED [ 16%] 2023-03-31T05:56:54.5092794Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_less_equal PASSED [ 17%] 2023-03-31T05:56:54.5093105Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_logaddexp PASSED [ 17%] 2023-03-31T05:56:54.5093428Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_logaddexp2 PASSED [ 17%] 2023-03-31T05:56:54.5093745Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_lt PASSED [ 17%] 2023-03-31T05:56:54.5094048Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_maximum PASSED [ 17%] 2023-03-31T05:56:54.5094367Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_minimum PASSED [ 17%] 2023-03-31T05:56:54.5094683Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_mul PASSED [ 17%] 2023-03-31T05:56:54.5095000Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_multiply PASSED [ 17%] 2023-03-31T05:56:54.5095305Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_ne PASSED [ 18%] 2023-03-31T05:56:54.5095625Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_nextafter PASSED [ 18%] 2023-03-31T05:56:54.5095947Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_not_equal PASSED [ 18%] 2023-03-31T05:56:54.5096253Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_remainder PASSED [ 18%] 2023-03-31T05:56:54.5096569Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_sub PASSED [ 18%] 2023-03-31T05:56:54.5096894Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_subtract PASSED [ 18%] 2023-03-31T05:56:54.5097215Z test_maskedtensor.py::TestBinary::test_binary_fn_aten_true_divide PASSED [ 18%] 2023-03-31T05:56:54.5097567Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_add_ PASSED [ 18%] 2023-03-31T05:56:54.5097944Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_arctan2_ PASSED [ 18%] 2023-03-31T05:56:54.5098290Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_atan2_ PASSED [ 19%] 2023-03-31T05:56:54.5098627Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_and_ PASSED [ 19%] 2023-03-31T05:56:54.5098996Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_left_shift_ PASSED [ 19%] 2023-03-31T05:56:54.5099360Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_or_ PASSED [ 19%] 2023-03-31T05:56:54.5099726Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_right_shift_ PASSED [ 19%] 2023-03-31T05:56:54.5100076Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_bitwise_xor_ PASSED [ 19%] 2023-03-31T05:56:54.5100428Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_div_ PASSED [ 19%] 2023-03-31T05:56:54.5100775Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_divide_ PASSED [ 19%] 2023-03-31T05:56:54.5101104Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_eq_ PASSED [ 20%] 2023-03-31T05:56:54.5101450Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_floor_divide_ PASSED [ 20%] 2023-03-31T05:56:54.5101804Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_fmod_ PASSED [ 20%] 2023-03-31T05:56:54.5102177Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_ge_ PASSED [ 20%] 2023-03-31T05:56:54.5102507Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_greater_ PASSED [ 20%] 2023-03-31T05:56:54.5102863Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_greater_equal_ PASSED [ 20%] 2023-03-31T05:56:54.5103209Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_gt_ PASSED [ 20%] 2023-03-31T05:56:54.5103638Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_le_ PASSED [ 20%] 2023-03-31T05:56:54.5103979Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_less_ PASSED [ 21%] 2023-03-31T05:56:54.5104332Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_less_equal_ PASSED [ 21%] 2023-03-31T05:56:54.5104672Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_lt_ PASSED [ 21%] 2023-03-31T05:56:54.5105034Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_mul_ PASSED [ 21%] 2023-03-31T05:56:54.5105385Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_multiply_ PASSED [ 21%] 2023-03-31T05:56:54.5105726Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_ne_ PASSED [ 21%] 2023-03-31T05:56:54.5106058Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_nextafter_ PASSED [ 21%] 2023-03-31T05:56:54.5106411Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_not_equal_ PASSED [ 21%] 2023-03-31T05:56:54.5106760Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_remainder_ PASSED [ 21%] 2023-03-31T05:56:54.5107110Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_sub_ PASSED [ 22%] 2023-03-31T05:56:54.5107444Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_subtract_ PASSED [ 22%] 2023-03-31T05:56:54.5107797Z test_maskedtensor.py::TestBinary::test_inplace_binary_fn_aten_true_divide_ PASSED [ 22%] 2023-03-31T05:56:54.5108138Z test_maskedtensor.py::TestBinary::test_masks_match_fn_name_add PASSED [ 22%] 2023-03-31T05:56:54.5108467Z test_maskedtensor.py::TestBinary::test_masks_match_fn_name_add_ PASSED [ 22%] 2023-03-31T05:56:54.5108773Z test_maskedtensor.py::TestReductions::test_all PASSED [ 22%] 2023-03-31T05:56:54.5109080Z test_maskedtensor.py::TestReductions::test_amax PASSED [ 22%] 2023-03-31T05:56:54.5109393Z test_maskedtensor.py::TestReductions::test_amax_grad PASSED [ 22%] 2023-03-31T05:56:54.5109689Z test_maskedtensor.py::TestReductions::test_amin PASSED [ 23%] 2023-03-31T05:56:54.5110035Z test_maskedtensor.py::TestReductions::test_amin_grad PASSED [ 23%] 2023-03-31T05:56:54.5110382Z test_maskedtensor.py::TestReductions::test_grad_dtype PASSED [ 23%] 2023-03-31T05:56:54.5110693Z test_maskedtensor.py::TestReductions::test_max_not_implemented PASSED [ 23%] 2023-03-31T05:56:54.5111007Z test_maskedtensor.py::TestReductions::test_mean PASSED [ 23%] 2023-03-31T05:56:54.5111319Z test_maskedtensor.py::TestReductions::test_mean_dim_grad PASSED [ 23%] 2023-03-31T05:56:54.5111637Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1a PASSED [ 23%] 2023-03-31T05:56:54.5111943Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1b PASSED [ 23%] 2023-03-31T05:56:54.5112259Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1c PASSED [ 24%] 2023-03-31T05:56:54.5112585Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1d PASSED [ 24%] 2023-03-31T05:56:54.5112895Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1e PASSED [ 24%] 2023-03-31T05:56:54.5113220Z test_maskedtensor.py::TestReductions::test_mean_grad_case_1f PASSED [ 24%] 2023-03-31T05:56:54.5113537Z test_maskedtensor.py::TestReductions::test_prod PASSED [ 24%] 2023-03-31T05:56:54.5113847Z test_maskedtensor.py::TestReductions::test_prod_grad PASSED [ 24%] 2023-03-31T05:56:54.5114150Z test_maskedtensor.py::TestReductions::test_sum PASSED [ 24%] 2023-03-31T05:56:54.5114493Z test_maskedtensor.py::TestReductions::test_sum_grad PASSED [ 24%] 2023-03-31T05:56:54.5114863Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float16 PASSED [ 25%] 2023-03-31T05:56:54.5115279Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float32 PASSED [ 25%] 2023-03-31T05:56:54.5115671Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_coo_cpu_float64 PASSED [ 25%] 2023-03-31T05:56:54.5116070Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float16 PASSED [ 25%] 2023-03-31T05:56:54.5116488Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float32 PASSED [ 25%] 2023-03-31T05:56:54.5116882Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_sparse_csr_cpu_float64 PASSED [ 25%] 2023-03-31T05:56:54.5117325Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float16 PASSED [ 25%] 2023-03-31T05:56:54.5117736Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float32 PASSED [ 25%] 2023-03-31T05:56:54.5118134Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_add_layout_torch_strided_cpu_float64 PASSED [ 25%] 2023-03-31T05:56:54.5118523Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_coo_cpu_float32 PASSED [ 26%] 2023-03-31T05:56:54.5118933Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_coo_cpu_float64 PASSED [ 26%] 2023-03-31T05:56:54.5119343Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_csr_cpu_float32 PASSED [ 26%] 2023-03-31T05:56:54.5119746Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_sparse_csr_cpu_float64 PASSED [ 26%] 2023-03-31T05:56:54.5120146Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_strided_cpu_float32 PASSED [ 26%] 2023-03-31T05:56:54.5120554Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_atan2_layout_torch_strided_cpu_float64 PASSED [ 26%] 2023-03-31T05:56:54.5120977Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float16 PASSED [ 26%] 2023-03-31T05:56:54.5121418Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float32 PASSED [ 26%] 2023-03-31T05:56:54.5121876Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_coo_cpu_float64 PASSED [ 27%] 2023-03-31T05:56:54.5122344Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float16 PASSED [ 27%] 2023-03-31T05:56:54.5122776Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float32 PASSED [ 27%] 2023-03-31T05:56:54.5123397Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_sparse_csr_cpu_float64 PASSED [ 27%] 2023-03-31T05:56:54.5123822Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float16 PASSED [ 27%] 2023-03-31T05:56:54.5124259Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float32 PASSED [ 27%] 2023-03-31T05:56:54.5124688Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_floor_rounding_layout_torch_strided_cpu_float64 PASSED [ 27%] 2023-03-31T05:56:54.5125122Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float16 PASSED [ 27%] 2023-03-31T05:56:54.5125552Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float32 PASSED [ 28%] 2023-03-31T05:56:54.5125981Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_coo_cpu_float64 PASSED [ 28%] 2023-03-31T05:56:54.5126492Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float16 PASSED [ 28%] 2023-03-31T05:56:54.5126923Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float32 PASSED [ 28%] 2023-03-31T05:56:54.5127350Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_sparse_csr_cpu_float64 PASSED [ 28%] 2023-03-31T05:56:54.5127783Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float16 PASSED [ 28%] 2023-03-31T05:56:54.5128226Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float32 PASSED [ 28%] 2023-03-31T05:56:54.5128660Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_no_rounding_mode_layout_torch_strided_cpu_float64 PASSED [ 28%] 2023-03-31T05:56:54.5129124Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float16 PASSED [ 28%] 2023-03-31T05:56:54.5129560Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float32 PASSED [ 29%] 2023-03-31T05:56:54.5129996Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_coo_cpu_float64 PASSED [ 29%] 2023-03-31T05:56:54.5130411Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float16 PASSED [ 29%] 2023-03-31T05:56:54.5130847Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float32 PASSED [ 29%] 2023-03-31T05:56:54.5131281Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_sparse_csr_cpu_float64 PASSED [ 29%] 2023-03-31T05:56:54.5131712Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float16 PASSED [ 29%] 2023-03-31T05:56:54.5132134Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float32 PASSED [ 29%] 2023-03-31T05:56:54.5132562Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_div_trunc_rounding_layout_torch_strided_cpu_float64 PASSED [ 29%] 2023-03-31T05:56:54.5132980Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float16 PASSED [ 30%] 2023-03-31T05:56:54.5133383Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float32 PASSED [ 30%] 2023-03-31T05:56:54.5133829Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_coo_cpu_float64 PASSED [ 30%] 2023-03-31T05:56:54.5134270Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float16 PASSED [ 30%] 2023-03-31T05:56:54.5134671Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float32 PASSED [ 30%] 2023-03-31T05:56:54.5135072Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_sparse_csr_cpu_float64 PASSED [ 30%] 2023-03-31T05:56:54.5135463Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float16 PASSED [ 30%] 2023-03-31T05:56:54.5135862Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float32 PASSED [ 30%] 2023-03-31T05:56:54.5136256Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_eq_layout_torch_strided_cpu_float64 PASSED [ 31%] 2023-03-31T05:56:54.5136667Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float16 PASSED [ 31%] 2023-03-31T05:56:54.5137084Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float32 PASSED [ 31%] 2023-03-31T05:56:54.5137507Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_coo_cpu_float64 PASSED [ 31%] 2023-03-31T05:56:54.5137932Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float16 PASSED [ 31%] 2023-03-31T05:56:54.5138383Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float32 PASSED [ 31%] 2023-03-31T05:56:54.5138797Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_sparse_csr_cpu_float64 PASSED [ 31%] 2023-03-31T05:56:54.5139216Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float16 PASSED [ 31%] 2023-03-31T05:56:54.5139645Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float32 PASSED [ 31%] 2023-03-31T05:56:54.5140066Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_floor_divide_layout_torch_strided_cpu_float64 PASSED [ 32%] 2023-03-31T05:56:54.5140464Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float16 PASSED [ 32%] 2023-03-31T05:56:54.5140931Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float32 PASSED [ 32%] 2023-03-31T05:56:54.5156158Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_coo_cpu_float64 PASSED [ 32%] 2023-03-31T05:56:54.5156597Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float16 PASSED [ 32%] 2023-03-31T05:56:54.5157006Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float32 PASSED [ 32%] 2023-03-31T05:56:54.5157420Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_sparse_csr_cpu_float64 PASSED [ 32%] 2023-03-31T05:56:54.5157832Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float16 PASSED [ 32%] 2023-03-31T05:56:54.5158242Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float32 PASSED [ 33%] 2023-03-31T05:56:54.5158631Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmax_layout_torch_strided_cpu_float64 PASSED [ 33%] 2023-03-31T05:56:54.5159043Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float16 PASSED [ 33%] 2023-03-31T05:56:54.5159446Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float32 PASSED [ 33%] 2023-03-31T05:56:54.5159856Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_coo_cpu_float64 PASSED [ 33%] 2023-03-31T05:56:54.5160248Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float16 PASSED [ 33%] 2023-03-31T05:56:54.5160790Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float32 PASSED [ 33%] 2023-03-31T05:56:54.5161240Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_sparse_csr_cpu_float64 PASSED [ 33%] 2023-03-31T05:56:54.5161646Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float16 PASSED [ 34%] 2023-03-31T05:56:54.5162042Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float32 PASSED [ 34%] 2023-03-31T05:56:54.5162445Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmin_layout_torch_strided_cpu_float64 PASSED [ 34%] 2023-03-31T05:56:54.5162850Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float16 PASSED [ 34%] 2023-03-31T05:56:54.5163462Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float32 PASSED [ 34%] 2023-03-31T05:56:54.5163863Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_coo_cpu_float64 PASSED [ 34%] 2023-03-31T05:56:54.5164272Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float16 PASSED [ 34%] 2023-03-31T05:56:54.5164684Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float32 PASSED [ 34%] 2023-03-31T05:56:54.5165082Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_sparse_csr_cpu_float64 PASSED [ 34%] 2023-03-31T05:56:54.5165580Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float16 PASSED [ 35%] 2023-03-31T05:56:54.5165983Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float32 PASSED [ 35%] 2023-03-31T05:56:54.5166381Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_fmod_layout_torch_strided_cpu_float64 PASSED [ 35%] 2023-03-31T05:56:54.5166768Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float16 PASSED [ 35%] 2023-03-31T05:56:54.5167179Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float32 PASSED [ 35%] 2023-03-31T05:56:54.5167588Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_coo_cpu_float64 PASSED [ 35%] 2023-03-31T05:56:54.5167994Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float16 PASSED [ 35%] 2023-03-31T05:56:54.5168437Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float32 PASSED [ 35%] 2023-03-31T05:56:54.5168842Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_sparse_csr_cpu_float64 PASSED [ 36%] 2023-03-31T05:56:54.5169247Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float16 PASSED [ 36%] 2023-03-31T05:56:54.5169651Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float32 PASSED [ 36%] 2023-03-31T05:56:54.5170036Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ge_layout_torch_strided_cpu_float64 PASSED [ 36%] 2023-03-31T05:56:54.5170442Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float16 PASSED [ 36%] 2023-03-31T05:56:54.5170845Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float32 PASSED [ 36%] 2023-03-31T05:56:54.5171234Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_coo_cpu_float64 PASSED [ 36%] 2023-03-31T05:56:54.5171632Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float16 PASSED [ 36%] 2023-03-31T05:56:54.5172042Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float32 PASSED [ 37%] 2023-03-31T05:56:54.5172448Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_sparse_csr_cpu_float64 PASSED [ 37%] 2023-03-31T05:56:54.5172882Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float16 PASSED [ 37%] 2023-03-31T05:56:54.5173324Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float32 PASSED [ 37%] 2023-03-31T05:56:54.5173729Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_gt_layout_torch_strided_cpu_float64 PASSED [ 37%] 2023-03-31T05:56:54.5174132Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float16 PASSED [ 37%] 2023-03-31T05:56:54.5174528Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float32 PASSED [ 37%] 2023-03-31T05:56:54.5174931Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_coo_cpu_float64 PASSED [ 37%] 2023-03-31T05:56:54.5175342Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float16 PASSED [ 37%] 2023-03-31T05:56:54.5175745Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float32 PASSED [ 38%] 2023-03-31T05:56:54.5176138Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_sparse_csr_cpu_float64 PASSED [ 38%] 2023-03-31T05:56:54.5176545Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float16 PASSED [ 38%] 2023-03-31T05:56:54.5176949Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float32 PASSED [ 38%] 2023-03-31T05:56:54.5177335Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_le_layout_torch_strided_cpu_float64 PASSED [ 38%] 2023-03-31T05:56:54.5177787Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_coo_cpu_float32 PASSED [ 38%] 2023-03-31T05:56:54.5178216Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_coo_cpu_float64 PASSED [ 38%] 2023-03-31T05:56:54.5178639Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_csr_cpu_float32 PASSED [ 38%] 2023-03-31T05:56:54.5179054Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_sparse_csr_cpu_float64 PASSED [ 39%] 2023-03-31T05:56:54.5179481Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_strided_cpu_float32 PASSED [ 39%] 2023-03-31T05:56:54.5179904Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_logaddexp_layout_torch_strided_cpu_float64 PASSED [ 39%] 2023-03-31T05:56:54.5180367Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float16 PASSED [ 39%] 2023-03-31T05:56:54.5180761Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float32 PASSED [ 39%] 2023-03-31T05:56:54.5181162Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_coo_cpu_float64 PASSED [ 39%] 2023-03-31T05:56:54.5181561Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float16 PASSED [ 39%] 2023-03-31T05:56:54.5181962Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float32 PASSED [ 39%] 2023-03-31T05:56:54.5182345Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_sparse_csr_cpu_float64 PASSED [ 40%] 2023-03-31T05:56:54.5182750Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float16 PASSED [ 40%] 2023-03-31T05:56:54.5183152Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float32 PASSED [ 40%] 2023-03-31T05:56:54.5183632Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_lt_layout_torch_strided_cpu_float64 PASSED [ 40%] 2023-03-31T05:56:54.5184026Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float16 PASSED [ 40%] 2023-03-31T05:56:54.5184441Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float32 PASSED [ 40%] 2023-03-31T05:56:54.5184854Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_coo_cpu_float64 PASSED [ 40%] 2023-03-31T05:56:54.5185303Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float16 PASSED [ 40%] 2023-03-31T05:56:54.5185742Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float32 PASSED [ 40%] 2023-03-31T05:56:54.5186159Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_sparse_csr_cpu_float64 PASSED [ 41%] 2023-03-31T05:56:54.5186572Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float16 PASSED [ 41%] 2023-03-31T05:56:54.5186971Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float32 PASSED [ 41%] 2023-03-31T05:56:54.5187376Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_maximum_layout_torch_strided_cpu_float64 PASSED [ 41%] 2023-03-31T05:56:54.5187793Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float16 PASSED [ 41%] 2023-03-31T05:56:54.5188208Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float32 PASSED [ 41%] 2023-03-31T05:56:54.5188606Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_coo_cpu_float64 PASSED [ 41%] 2023-03-31T05:56:54.5189016Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float16 PASSED [ 41%] 2023-03-31T05:56:54.5189463Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float32 PASSED [ 42%] 2023-03-31T05:56:54.5189875Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_sparse_csr_cpu_float64 PASSED [ 42%] 2023-03-31T05:56:54.5190269Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float16 PASSED [ 42%] 2023-03-31T05:56:54.5190672Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float32 PASSED [ 42%] 2023-03-31T05:56:54.5191080Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_minimum_layout_torch_strided_cpu_float64 PASSED [ 42%] 2023-03-31T05:56:54.5191488Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float16 PASSED [ 42%] 2023-03-31T05:56:54.5191881Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float32 PASSED [ 42%] 2023-03-31T05:56:54.5192316Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_coo_cpu_float64 PASSED [ 42%] 2023-03-31T05:56:54.5192719Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float16 PASSED [ 43%] 2023-03-31T05:56:54.5193121Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float32 PASSED [ 43%] 2023-03-31T05:56:54.5193509Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_sparse_csr_cpu_float64 PASSED [ 43%] 2023-03-31T05:56:54.5193912Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float16 PASSED [ 43%] 2023-03-31T05:56:54.5194315Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float32 PASSED [ 43%] 2023-03-31T05:56:54.5194700Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_mul_layout_torch_strided_cpu_float64 PASSED [ 43%] 2023-03-31T05:56:54.5195100Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float16 PASSED [ 43%] 2023-03-31T05:56:54.5195503Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float32 PASSED [ 43%] 2023-03-31T05:56:54.5195902Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_coo_cpu_float64 PASSED [ 43%] 2023-03-31T05:56:54.5196289Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float16 PASSED [ 44%] 2023-03-31T05:56:54.5196686Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float32 PASSED [ 44%] 2023-03-31T05:56:54.5197119Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_sparse_csr_cpu_float64 PASSED [ 44%] 2023-03-31T05:56:54.5197553Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float16 PASSED [ 44%] 2023-03-31T05:56:54.5197938Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float32 PASSED [ 44%] 2023-03-31T05:56:54.5198343Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_ne_layout_torch_strided_cpu_float64 PASSED [ 44%] 2023-03-31T05:56:54.5198747Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_coo_cpu_float32 PASSED [ 44%] 2023-03-31T05:56:54.5199167Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_coo_cpu_float64 PASSED [ 44%] 2023-03-31T05:56:54.5199570Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_csr_cpu_float32 PASSED [ 45%] 2023-03-31T05:56:54.5199994Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_sparse_csr_cpu_float64 PASSED [ 45%] 2023-03-31T05:56:54.5200410Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_strided_cpu_float32 PASSED [ 45%] 2023-03-31T05:56:54.5200824Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_nextafter_layout_torch_strided_cpu_float64 PASSED [ 45%] 2023-03-31T05:56:54.5201271Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float16 PASSED [ 45%] 2023-03-31T05:56:54.5201688Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float32 PASSED [ 45%] 2023-03-31T05:56:54.5202103Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_coo_cpu_float64 PASSED [ 45%] 2023-03-31T05:56:54.5202500Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float16 PASSED [ 45%] 2023-03-31T05:56:54.5202909Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float32 PASSED [ 46%] 2023-03-31T05:56:54.5203539Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_sparse_csr_cpu_float64 PASSED [ 46%] 2023-03-31T05:56:54.5203962Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float16 PASSED [ 46%] 2023-03-31T05:56:54.5204426Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float32 PASSED [ 46%] 2023-03-31T05:56:54.5204847Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_remainder_layout_torch_strided_cpu_float64 PASSED [ 46%] 2023-03-31T05:56:54.5205257Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float16 PASSED [ 46%] 2023-03-31T05:56:54.5205668Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float32 PASSED [ 46%] 2023-03-31T05:56:54.5206069Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_coo_cpu_float64 PASSED [ 46%] 2023-03-31T05:56:54.5206472Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float16 PASSED [ 46%] 2023-03-31T05:56:54.5206875Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float32 PASSED [ 47%] 2023-03-31T05:56:54.5207286Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_sparse_csr_cpu_float64 PASSED [ 47%] 2023-03-31T05:56:54.5207678Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float16 PASSED [ 47%] 2023-03-31T05:56:54.5208078Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float32 PASSED [ 47%] 2023-03-31T05:56:54.5208472Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_sub_layout_torch_strided_cpu_float64 PASSED [ 47%] 2023-03-31T05:56:54.5208880Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float16 PASSED [ 47%] 2023-03-31T05:56:54.5209333Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float32 PASSED [ 47%] 2023-03-31T05:56:54.5209794Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_coo_cpu_float64 PASSED [ 47%] 2023-03-31T05:56:54.5210212Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float16 PASSED [ 48%] 2023-03-31T05:56:54.5210618Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float32 PASSED [ 48%] 2023-03-31T05:56:54.5211035Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_sparse_csr_cpu_float64 PASSED [ 48%] 2023-03-31T05:56:54.5211453Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float16 PASSED [ 48%] 2023-03-31T05:56:54.5211873Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float32 PASSED [ 48%] 2023-03-31T05:56:54.5212277Z test_maskedtensor.py::TestOperatorsCPU::test_binary_core_true_divide_layout_torch_strided_cpu_float64 PASSED [ 48%] 2023-03-31T05:56:54.5212692Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float16 PASSED [ 48%] 2023-03-31T05:56:54.5213109Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float32 PASSED [ 48%] 2023-03-31T05:56:54.5213573Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_coo_cpu_float64 PASSED [ 49%] 2023-03-31T05:56:54.5213974Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float16 PASSED [ 49%] 2023-03-31T05:56:54.5214381Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float32 PASSED [ 49%] 2023-03-31T05:56:54.5214789Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_sparse_csr_cpu_float64 PASSED [ 49%] 2023-03-31T05:56:54.5215200Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float16 PASSED [ 49%] 2023-03-31T05:56:54.5215594Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float32 PASSED [ 49%] 2023-03-31T05:56:54.5215998Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amax_layout_torch_strided_cpu_float64 PASSED [ 49%] 2023-03-31T05:56:54.5216444Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float16 PASSED [ 49%] 2023-03-31T05:56:54.5216851Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float32 PASSED [ 50%] 2023-03-31T05:56:54.5217248Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_coo_cpu_float64 PASSED [ 50%] 2023-03-31T05:56:54.5217657Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float16 PASSED [ 50%] 2023-03-31T05:56:54.5218066Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float32 PASSED [ 50%] 2023-03-31T05:56:54.5218476Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_sparse_csr_cpu_float64 PASSED [ 50%] 2023-03-31T05:56:54.5218872Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float16 PASSED [ 50%] 2023-03-31T05:56:54.5219283Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float32 PASSED [ 50%] 2023-03-31T05:56:54.5219687Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_amin_layout_torch_strided_cpu_float64 PASSED [ 50%] 2023-03-31T05:56:54.5220086Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float16 PASSED [ 50%] 2023-03-31T05:56:54.5220497Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float32 PASSED [ 51%] 2023-03-31T05:56:54.5220949Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_coo_cpu_float64 PASSED [ 51%] 2023-03-31T05:56:54.5221391Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float16 PASSED [ 51%] 2023-03-31T05:56:54.5221795Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float32 PASSED [ 51%] 2023-03-31T05:56:54.5222207Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_sparse_csr_cpu_float64 PASSED [ 51%] 2023-03-31T05:56:54.5222627Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float16 PASSED [ 51%] 2023-03-31T05:56:54.5223036Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float32 PASSED [ 51%] 2023-03-31T05:56:54.5223522Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmax_layout_torch_strided_cpu_float64 PASSED [ 51%] 2023-03-31T05:56:54.5223940Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float16 PASSED [ 52%] 2023-03-31T05:56:54.5224356Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float32 PASSED [ 52%] 2023-03-31T05:56:54.5224773Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_coo_cpu_float64 PASSED [ 52%] 2023-03-31T05:56:54.5225178Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float16 PASSED [ 52%] 2023-03-31T05:56:54.5225639Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float32 PASSED [ 52%] 2023-03-31T05:56:54.5226054Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_sparse_csr_cpu_float64 PASSED [ 52%] 2023-03-31T05:56:54.5226466Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float16 PASSED [ 52%] 2023-03-31T05:56:54.5226866Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float32 PASSED [ 52%] 2023-03-31T05:56:54.5227276Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_argmin_layout_torch_strided_cpu_float64 PASSED [ 53%] 2023-03-31T05:56:54.5227696Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_coo_cpu_float32 PASSED [ 53%] 2023-03-31T05:56:54.5228132Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_coo_cpu_float64 PASSED [ 53%] 2023-03-31T05:56:54.5228548Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_csr_cpu_float32 PASSED [ 53%] 2023-03-31T05:56:54.5228961Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_sparse_csr_cpu_float64 PASSED [ 53%] 2023-03-31T05:56:54.5229374Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_strided_cpu_float32 PASSED [ 53%] 2023-03-31T05:56:54.5229774Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_prod_layout_torch_strided_cpu_float64 PASSED [ 53%] 2023-03-31T05:56:54.5230183Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float16 PASSED [ 53%] 2023-03-31T05:56:54.5230611Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float32 SKIPPED (Skipped!) [ 53%] 2023-03-31T05:56:54.5231039Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_coo_cpu_float64 PASSED [ 54%] 2023-03-31T05:56:54.5231434Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float16 PASSED [ 54%] 2023-03-31T05:56:54.5231853Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float32 SKIPPED (Skipped!) [ 54%] 2023-03-31T05:56:54.5232271Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_sparse_csr_cpu_float64 PASSED [ 54%] 2023-03-31T05:56:54.5232678Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float16 PASSED [ 54%] 2023-03-31T05:56:54.5233127Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float32 SKIPPED (Skipped!) [ 54%] 2023-03-31T05:56:54.5233579Z test_maskedtensor.py::TestOperatorsCPU::test_reduction_all_sum_layout_torch_strided_cpu_float64 PASSED [ 54%] 2023-03-31T05:56:54.5233988Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float16 PASSED [ 54%] 2023-03-31T05:56:54.5234393Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float32 PASSED [ 55%] 2023-03-31T05:56:54.5234778Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_coo_cpu_float64 PASSED [ 55%] 2023-03-31T05:56:54.5235184Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float16 PASSED [ 55%] 2023-03-31T05:56:54.5235589Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float32 PASSED [ 55%] 2023-03-31T05:56:54.5235993Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_sparse_csr_cpu_float64 PASSED [ 55%] 2023-03-31T05:56:54.5236387Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float16 PASSED [ 55%] 2023-03-31T05:56:54.5236793Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float32 PASSED [ 55%] 2023-03-31T05:56:54.5237193Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_abs_layout_torch_strided_cpu_float64 PASSED [ 55%] 2023-03-31T05:56:54.5237613Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_coo_cpu_float32 PASSED [ 56%] 2023-03-31T05:56:54.5238023Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_coo_cpu_float64 PASSED [ 56%] 2023-03-31T05:56:54.5238431Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_csr_cpu_float32 PASSED [ 56%] 2023-03-31T05:56:54.5238839Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_sparse_csr_cpu_float64 PASSED [ 56%] 2023-03-31T05:56:54.5239237Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_strided_cpu_float32 PASSED [ 56%] 2023-03-31T05:56:54.5239642Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acos_layout_torch_strided_cpu_float64 PASSED [ 56%] 2023-03-31T05:56:54.5240086Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_coo_cpu_float32 PASSED [ 56%] 2023-03-31T05:56:54.5240505Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_coo_cpu_float64 PASSED [ 56%] 2023-03-31T05:56:54.5240903Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_csr_cpu_float32 PASSED [ 56%] 2023-03-31T05:56:54.5241305Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_sparse_csr_cpu_float64 PASSED [ 57%] 2023-03-31T05:56:54.5241721Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_strided_cpu_float32 PASSED [ 57%] 2023-03-31T05:56:54.5242132Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_acosh_layout_torch_strided_cpu_float64 PASSED [ 57%] 2023-03-31T05:56:54.5242524Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float16 PASSED [ 57%] 2023-03-31T05:56:54.5242931Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float32 PASSED [ 57%] 2023-03-31T05:56:54.5243545Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_coo_cpu_float64 PASSED [ 57%] 2023-03-31T05:56:54.5243939Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float16 PASSED [ 57%] 2023-03-31T05:56:54.5244349Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float32 PASSED [ 57%] 2023-03-31T05:56:54.5244747Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_sparse_csr_cpu_float64 PASSED [ 58%] 2023-03-31T05:56:54.5245257Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float16 PASSED [ 58%] 2023-03-31T05:56:54.5245690Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float32 PASSED [ 58%] 2023-03-31T05:56:54.5246091Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_angle_layout_torch_strided_cpu_float64 PASSED [ 58%] 2023-03-31T05:56:54.5246502Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_coo_cpu_float32 PASSED [ 58%] 2023-03-31T05:56:54.5246910Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_coo_cpu_float64 PASSED [ 58%] 2023-03-31T05:56:54.5247304Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_csr_cpu_float32 PASSED [ 58%] 2023-03-31T05:56:54.5247707Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_sparse_csr_cpu_float64 PASSED [ 58%] 2023-03-31T05:56:54.5248115Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_strided_cpu_float32 PASSED [ 59%] 2023-03-31T05:56:54.5248518Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asin_layout_torch_strided_cpu_float64 PASSED [ 59%] 2023-03-31T05:56:54.5248906Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_coo_cpu_float32 PASSED [ 59%] 2023-03-31T05:56:54.5249316Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_coo_cpu_float64 PASSED [ 59%] 2023-03-31T05:56:54.5249762Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_csr_cpu_float32 PASSED [ 59%] 2023-03-31T05:56:54.5250165Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_sparse_csr_cpu_float64 PASSED [ 59%] 2023-03-31T05:56:54.5250553Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_strided_cpu_float32 PASSED [ 59%] 2023-03-31T05:56:54.5250953Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_asinh_layout_torch_strided_cpu_float64 PASSED [ 59%] 2023-03-31T05:56:54.5251354Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_coo_cpu_float32 PASSED [ 59%] 2023-03-31T05:56:54.5251749Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_coo_cpu_float64 PASSED [ 60%] 2023-03-31T05:56:54.5252152Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_csr_cpu_float32 PASSED [ 60%] 2023-03-31T05:56:54.5252594Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_sparse_csr_cpu_float64 PASSED [ 60%] 2023-03-31T05:56:54.5252996Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_strided_cpu_float32 PASSED [ 60%] 2023-03-31T05:56:54.5253384Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atan_layout_torch_strided_cpu_float64 PASSED [ 60%] 2023-03-31T05:56:54.5253783Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_coo_cpu_float32 PASSED [ 60%] 2023-03-31T05:56:54.5254189Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_coo_cpu_float64 PASSED [ 60%] 2023-03-31T05:56:54.5254593Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_csr_cpu_float32 PASSED [ 60%] 2023-03-31T05:56:54.5254980Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_sparse_csr_cpu_float64 PASSED [ 61%] 2023-03-31T05:56:54.5255381Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_strided_cpu_float32 PASSED [ 61%] 2023-03-31T05:56:54.5255785Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_atanh_layout_torch_strided_cpu_float64 PASSED [ 61%] 2023-03-31T05:56:54.5256188Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_coo_cpu_float32 PASSED [ 61%] 2023-03-31T05:56:54.5256577Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_coo_cpu_float64 PASSED [ 61%] 2023-03-31T05:56:54.5257010Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_csr_cpu_float32 PASSED [ 61%] 2023-03-31T05:56:54.5257439Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_sparse_csr_cpu_float64 PASSED [ 61%] 2023-03-31T05:56:54.5257828Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_strided_cpu_float32 PASSED [ 61%] 2023-03-31T05:56:54.5258233Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_ceil_layout_torch_strided_cpu_float64 PASSED [ 62%] 2023-03-31T05:56:54.5258650Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float16 PASSED [ 62%] 2023-03-31T05:56:54.5259075Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float32 PASSED [ 62%] 2023-03-31T05:56:54.5259489Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_coo_cpu_float64 PASSED [ 62%] 2023-03-31T05:56:54.5259912Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float16 PASSED [ 62%] 2023-03-31T05:56:54.5260340Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float32 PASSED [ 62%] 2023-03-31T05:56:54.5260759Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_sparse_csr_cpu_float64 PASSED [ 62%] 2023-03-31T05:56:54.5261173Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float16 PASSED [ 62%] 2023-03-31T05:56:54.5261620Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float32 PASSED [ 62%] 2023-03-31T05:56:54.5262034Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_conj_physical_layout_torch_strided_cpu_float64 PASSED [ 63%] 2023-03-31T05:56:54.5262449Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_coo_cpu_float32 PASSED [ 63%] 2023-03-31T05:56:54.5262841Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_coo_cpu_float64 PASSED [ 63%] 2023-03-31T05:56:54.5263237Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_csr_cpu_float32 PASSED [ 63%] 2023-03-31T05:56:54.5263725Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_sparse_csr_cpu_float64 PASSED [ 63%] 2023-03-31T05:56:54.5264165Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_strided_cpu_float32 PASSED [ 63%] 2023-03-31T05:56:54.5264552Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cos_layout_torch_strided_cpu_float64 PASSED [ 63%] 2023-03-31T05:56:54.5264954Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_coo_cpu_float32 PASSED [ 63%] 2023-03-31T05:56:54.5265363Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_coo_cpu_float64 PASSED [ 64%] 2023-03-31T05:56:54.5265769Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_csr_cpu_float32 PASSED [ 64%] 2023-03-31T05:56:54.5266162Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_sparse_csr_cpu_float64 PASSED [ 64%] 2023-03-31T05:56:54.5266568Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_strided_cpu_float32 PASSED [ 64%] 2023-03-31T05:56:54.5266970Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_cosh_layout_torch_strided_cpu_float64 PASSED [ 64%] 2023-03-31T05:56:54.5267370Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float16 PASSED [ 64%] 2023-03-31T05:56:54.5267779Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float32 PASSED [ 64%] 2023-03-31T05:56:54.5268189Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_coo_cpu_float64 PASSED [ 64%] 2023-03-31T05:56:54.5268597Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float16 PASSED [ 65%] 2023-03-31T05:56:54.5269043Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float32 PASSED [ 65%] 2023-03-31T05:56:54.5269485Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_sparse_csr_cpu_float64 PASSED [ 65%] 2023-03-31T05:56:54.5269892Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float16 PASSED [ 65%] 2023-03-31T05:56:54.5270297Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float32 PASSED [ 65%] 2023-03-31T05:56:54.5270688Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_deg2rad_layout_torch_strided_cpu_float64 PASSED [ 65%] 2023-03-31T05:56:54.5271091Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_coo_cpu_float32 PASSED [ 65%] 2023-03-31T05:56:54.5271504Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_coo_cpu_float64 PASSED [ 65%] 2023-03-31T05:56:54.5271916Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_csr_cpu_float32 PASSED [ 65%] 2023-03-31T05:56:54.5272316Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_sparse_csr_cpu_float64 PASSED [ 66%] 2023-03-31T05:56:54.5272721Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_strided_cpu_float32 PASSED [ 66%] 2023-03-31T05:56:54.5273132Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_digamma_layout_torch_strided_cpu_float64 PASSED [ 66%] 2023-03-31T05:56:54.5273574Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_coo_cpu_float32 PASSED [ 66%] 2023-03-31T05:56:54.5273966Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_coo_cpu_float64 PASSED [ 66%] 2023-03-31T05:56:54.5274367Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_csr_cpu_float32 PASSED [ 66%] 2023-03-31T05:56:54.5274765Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_sparse_csr_cpu_float64 PASSED [ 66%] 2023-03-31T05:56:54.5275150Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_strided_cpu_float32 PASSED [ 66%] 2023-03-31T05:56:54.5275550Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erf_layout_torch_strided_cpu_float64 PASSED [ 67%] 2023-03-31T05:56:54.5275956Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_coo_cpu_float32 PASSED [ 67%] 2023-03-31T05:56:54.5276412Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_coo_cpu_float64 PASSED [ 67%] 2023-03-31T05:56:54.5276804Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_csr_cpu_float32 PASSED [ 67%] 2023-03-31T05:56:54.5277206Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_sparse_csr_cpu_float64 PASSED [ 67%] 2023-03-31T05:56:54.5277606Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_strided_cpu_float32 PASSED [ 67%] 2023-03-31T05:56:54.5278007Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfc_layout_torch_strided_cpu_float64 PASSED [ 67%] 2023-03-31T05:56:54.5278401Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_coo_cpu_float32 PASSED [ 67%] 2023-03-31T05:56:54.5278817Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_coo_cpu_float64 PASSED [ 68%] 2023-03-31T05:56:54.5279229Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_csr_cpu_float32 PASSED [ 68%] 2023-03-31T05:56:54.5279640Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_sparse_csr_cpu_float64 PASSED [ 68%] 2023-03-31T05:56:54.5280036Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_strided_cpu_float32 PASSED [ 68%] 2023-03-31T05:56:54.5280437Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_erfinv_layout_torch_strided_cpu_float64 PASSED [ 68%] 2023-03-31T05:56:54.5280876Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float16 PASSED [ 68%] 2023-03-31T05:56:54.5281310Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float32 PASSED [ 68%] 2023-03-31T05:56:54.5281702Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_coo_cpu_float64 PASSED [ 68%] 2023-03-31T05:56:54.5282107Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float16 PASSED [ 68%] 2023-03-31T05:56:54.5282509Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float32 PASSED [ 69%] 2023-03-31T05:56:54.5282898Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_sparse_csr_cpu_float64 PASSED [ 69%] 2023-03-31T05:56:54.5283492Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float16 PASSED [ 69%] 2023-03-31T05:56:54.5283898Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float32 PASSED [ 69%] 2023-03-31T05:56:54.5284301Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp2_layout_torch_strided_cpu_float64 PASSED [ 69%] 2023-03-31T05:56:54.5284689Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_coo_cpu_float32 PASSED [ 69%] 2023-03-31T05:56:54.5285095Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_coo_cpu_float64 PASSED [ 69%] 2023-03-31T05:56:54.5285562Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_csr_cpu_float32 PASSED [ 69%] 2023-03-31T05:56:54.5285966Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_sparse_csr_cpu_float64 PASSED [ 70%] 2023-03-31T05:56:54.5286353Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_strided_cpu_float32 PASSED [ 70%] 2023-03-31T05:56:54.5286754Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_exp_layout_torch_strided_cpu_float64 PASSED [ 70%] 2023-03-31T05:56:54.5287163Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_coo_cpu_float32 PASSED [ 70%] 2023-03-31T05:56:54.5287574Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_coo_cpu_float64 PASSED [ 70%] 2023-03-31T05:56:54.5287968Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_csr_cpu_float32 PASSED [ 70%] 2023-03-31T05:56:54.5288417Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_sparse_csr_cpu_float64 PASSED [ 70%] 2023-03-31T05:56:54.5288822Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_strided_cpu_float32 PASSED [ 70%] 2023-03-31T05:56:54.5289212Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_expm1_layout_torch_strided_cpu_float64 PASSED [ 71%] 2023-03-31T05:56:54.5289613Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_coo_cpu_float32 PASSED [ 71%] 2023-03-31T05:56:54.5290023Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_coo_cpu_float64 PASSED [ 71%] 2023-03-31T05:56:54.5290435Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_csr_cpu_float32 PASSED [ 71%] 2023-03-31T05:56:54.5290827Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_sparse_csr_cpu_float64 PASSED [ 71%] 2023-03-31T05:56:54.5291238Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_strided_cpu_float32 PASSED [ 71%] 2023-03-31T05:56:54.5291643Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_floor_layout_torch_strided_cpu_float64 PASSED [ 71%] 2023-03-31T05:56:54.5292051Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float16 PASSED [ 71%] 2023-03-31T05:56:54.5292449Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float32 PASSED [ 71%] 2023-03-31T05:56:54.5292850Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_coo_cpu_float64 PASSED [ 72%] 2023-03-31T05:56:54.5293292Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float16 PASSED [ 72%] 2023-03-31T05:56:54.5293737Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float32 PASSED [ 72%] 2023-03-31T05:56:54.5294131Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_sparse_csr_cpu_float64 PASSED [ 72%] 2023-03-31T05:56:54.5294541Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float16 PASSED [ 72%] 2023-03-31T05:56:54.5294949Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float32 PASSED [ 72%] 2023-03-31T05:56:54.5295351Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_frac_layout_torch_strided_cpu_float64 PASSED [ 72%] 2023-03-31T05:56:54.5295739Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_coo_cpu_float32 PASSED [ 72%] 2023-03-31T05:56:54.5296149Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_coo_cpu_float64 PASSED [ 73%] 2023-03-31T05:56:54.5296549Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_csr_cpu_float32 PASSED [ 73%] 2023-03-31T05:56:54.5296929Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_sparse_csr_cpu_float64 PASSED [ 73%] 2023-03-31T05:56:54.5297325Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_strided_cpu_float32 PASSED [ 73%] 2023-03-31T05:56:54.5297754Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_i0_layout_torch_strided_cpu_float64 PASSED [ 73%] 2023-03-31T05:56:54.5298161Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float16 PASSED [ 73%] 2023-03-31T05:56:54.5298557Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float32 PASSED [ 73%] 2023-03-31T05:56:54.5298965Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_coo_cpu_float64 PASSED [ 73%] 2023-03-31T05:56:54.5299378Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float16 PASSED [ 74%] 2023-03-31T05:56:54.5299787Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float32 PASSED [ 74%] 2023-03-31T05:56:54.5300214Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_sparse_csr_cpu_float64 PASSED [ 74%] 2023-03-31T05:56:54.5300618Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float16 PASSED [ 74%] 2023-03-31T05:56:54.5301026Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float32 PASSED [ 74%] 2023-03-31T05:56:54.5301430Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_isnan_layout_torch_strided_cpu_float64 PASSED [ 74%] 2023-03-31T05:56:54.5301821Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_coo_cpu_float32 PASSED [ 74%] 2023-03-31T05:56:54.5302233Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_coo_cpu_float64 PASSED [ 74%] 2023-03-31T05:56:54.5302642Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_csr_cpu_float32 PASSED [ 75%] 2023-03-31T05:56:54.5303037Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_sparse_csr_cpu_float64 PASSED [ 75%] 2023-03-31T05:56:54.5303536Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_strided_cpu_float32 PASSED [ 75%] 2023-03-31T05:56:54.5303945Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_lgamma_layout_torch_strided_cpu_float64 PASSED [ 75%] 2023-03-31T05:56:54.5304351Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_coo_cpu_float32 PASSED [ 75%] 2023-03-31T05:56:54.5304741Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_coo_cpu_float64 PASSED [ 75%] 2023-03-31T05:56:54.5305187Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_csr_cpu_float32 PASSED [ 75%] 2023-03-31T05:56:54.5305614Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_sparse_csr_cpu_float64 PASSED [ 75%] 2023-03-31T05:56:54.5306020Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_strided_cpu_float32 PASSED [ 75%] 2023-03-31T05:56:54.5306408Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log10_layout_torch_strided_cpu_float64 PASSED [ 76%] 2023-03-31T05:56:54.5306811Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_coo_cpu_float32 PASSED [ 76%] 2023-03-31T05:56:54.5307217Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_coo_cpu_float64 PASSED [ 76%] 2023-03-31T05:56:54.5307621Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_csr_cpu_float32 PASSED [ 76%] 2023-03-31T05:56:54.5308010Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_sparse_csr_cpu_float64 PASSED [ 76%] 2023-03-31T05:56:54.5308413Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_strided_cpu_float32 PASSED [ 76%] 2023-03-31T05:56:54.5308814Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log1p_layout_torch_strided_cpu_float64 PASSED [ 76%] 2023-03-31T05:56:54.5309216Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_coo_cpu_float32 PASSED [ 76%] 2023-03-31T05:56:54.5309660Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_coo_cpu_float64 PASSED [ 77%] 2023-03-31T05:56:54.5310061Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_csr_cpu_float32 PASSED [ 77%] 2023-03-31T05:56:54.5310460Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_sparse_csr_cpu_float64 PASSED [ 77%] 2023-03-31T05:56:54.5310849Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_strided_cpu_float32 PASSED [ 77%] 2023-03-31T05:56:54.5311251Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log2_layout_torch_strided_cpu_float64 PASSED [ 77%] 2023-03-31T05:56:54.5311652Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_coo_cpu_float32 PASSED [ 77%] 2023-03-31T05:56:54.5312051Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_coo_cpu_float64 PASSED [ 77%] 2023-03-31T05:56:54.5312477Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_csr_cpu_float32 PASSED [ 77%] 2023-03-31T05:56:54.5312878Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_sparse_csr_cpu_float64 PASSED [ 78%] 2023-03-31T05:56:54.5313277Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_strided_cpu_float32 PASSED [ 78%] 2023-03-31T05:56:54.5313676Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_log_layout_torch_strided_cpu_float64 PASSED [ 78%] 2023-03-31T05:56:54.5314066Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_coo_cpu_float32 PASSED [ 78%] 2023-03-31T05:56:54.5314473Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_coo_cpu_float64 PASSED [ 78%] 2023-03-31T05:56:54.5314880Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_csr_cpu_float32 PASSED [ 78%] 2023-03-31T05:56:54.5315283Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_sparse_csr_cpu_float64 PASSED [ 78%] 2023-03-31T05:56:54.5315675Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_strided_cpu_float32 PASSED [ 78%] 2023-03-31T05:56:54.5316075Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_logit_layout_torch_strided_cpu_float64 PASSED [ 78%] 2023-03-31T05:56:54.5316483Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float16 PASSED [ 79%] 2023-03-31T05:56:54.5316917Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float32 PASSED [ 79%] 2023-03-31T05:56:54.5317360Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_coo_cpu_float64 PASSED [ 79%] 2023-03-31T05:56:54.5317769Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float16 PASSED [ 79%] 2023-03-31T05:56:54.5318189Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float32 PASSED [ 79%] 2023-03-31T05:56:54.5318590Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_sparse_csr_cpu_float64 PASSED [ 79%] 2023-03-31T05:56:54.5319000Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float16 PASSED [ 79%] 2023-03-31T05:56:54.5319406Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float32 PASSED [ 79%] 2023-03-31T05:56:54.5319818Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_nan_to_num_layout_torch_strided_cpu_float64 PASSED [ 80%] 2023-03-31T05:56:54.5320213Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float16 PASSED [ 80%] 2023-03-31T05:56:54.5320616Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float32 PASSED [ 80%] 2023-03-31T05:56:54.5321018Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_coo_cpu_float64 PASSED [ 80%] 2023-03-31T05:56:54.5321447Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float16 PASSED [ 80%] 2023-03-31T05:56:54.5321838Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float32 PASSED [ 80%] 2023-03-31T05:56:54.5322237Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_sparse_csr_cpu_float64 PASSED [ 80%] 2023-03-31T05:56:54.5322635Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float16 PASSED [ 80%] 2023-03-31T05:56:54.5323206Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float32 PASSED [ 81%] 2023-03-31T05:56:54.5323633Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_neg_layout_torch_strided_cpu_float64 PASSED [ 81%] 2023-03-31T05:56:54.5324045Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float16 PASSED [ 81%] 2023-03-31T05:56:54.5324528Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float32 PASSED [ 81%] 2023-03-31T05:56:54.5324928Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_coo_cpu_float64 PASSED [ 81%] 2023-03-31T05:56:54.5325333Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float16 PASSED [ 81%] 2023-03-31T05:56:54.5325752Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float32 PASSED [ 81%] 2023-03-31T05:56:54.5326170Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_sparse_csr_cpu_float64 PASSED [ 81%] 2023-03-31T05:56:54.5326574Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float16 PASSED [ 81%] 2023-03-31T05:56:54.5326988Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float32 PASSED [ 82%] 2023-03-31T05:56:54.5327400Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_positive_layout_torch_strided_cpu_float64 PASSED [ 82%] 2023-03-31T05:56:54.5327819Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float16 PASSED [ 82%] 2023-03-31T05:56:54.5328222Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float32 PASSED [ 82%] 2023-03-31T05:56:54.5328630Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_coo_cpu_float64 PASSED [ 82%] 2023-03-31T05:56:54.5329081Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float16 PASSED [ 82%] 2023-03-31T05:56:54.5329527Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float32 PASSED [ 82%] 2023-03-31T05:56:54.5329920Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_sparse_csr_cpu_float64 PASSED [ 82%] 2023-03-31T05:56:54.5330328Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float16 PASSED [ 83%] 2023-03-31T05:56:54.5330734Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float32 PASSED [ 83%] 2023-03-31T05:56:54.5331140Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rad2deg_layout_torch_strided_cpu_float64 PASSED [ 83%] 2023-03-31T05:56:54.5331542Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float16 PASSED [ 83%] 2023-03-31T05:56:54.5331966Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float32 PASSED [ 83%] 2023-03-31T05:56:54.5332389Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_coo_cpu_float64 PASSED [ 83%] 2023-03-31T05:56:54.5332805Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float16 PASSED [ 83%] 2023-03-31T05:56:54.5333206Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float32 PASSED [ 83%] 2023-03-31T05:56:54.5333663Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_sparse_csr_cpu_float64 PASSED [ 84%] 2023-03-31T05:56:54.5334080Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float16 PASSED [ 84%] 2023-03-31T05:56:54.5334486Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float32 PASSED [ 84%] 2023-03-31T05:56:54.5334896Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_reciprocal_layout_torch_strided_cpu_float64 PASSED [ 84%] 2023-03-31T05:56:54.5335326Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_coo_cpu_float32 PASSED [ 84%] 2023-03-31T05:56:54.5335765Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_coo_cpu_float64 PASSED [ 84%] 2023-03-31T05:56:54.5336262Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_csr_cpu_float32 PASSED [ 84%] 2023-03-31T05:56:54.5336688Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_sparse_csr_cpu_float64 PASSED [ 84%] 2023-03-31T05:56:54.5337120Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_strided_cpu_float32 PASSED [ 84%] 2023-03-31T05:56:54.5337544Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_0_layout_torch_strided_cpu_float64 PASSED [ 85%] 2023-03-31T05:56:54.5337959Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_coo_cpu_float32 PASSED [ 85%] 2023-03-31T05:56:54.5338389Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_coo_cpu_float64 PASSED [ 85%] 2023-03-31T05:56:54.5338812Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_csr_cpu_float32 PASSED [ 85%] 2023-03-31T05:56:54.5339237Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_sparse_csr_cpu_float64 PASSED [ 85%] 2023-03-31T05:56:54.5339645Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_strided_cpu_float32 PASSED [ 85%] 2023-03-31T05:56:54.5340069Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_3_layout_torch_strided_cpu_float64 PASSED [ 85%] 2023-03-31T05:56:54.5340499Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_coo_cpu_float32 PASSED [ 85%] 2023-03-31T05:56:54.5340972Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_coo_cpu_float64 PASSED [ 86%] 2023-03-31T05:56:54.5341423Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_csr_cpu_float32 PASSED [ 86%] 2023-03-31T05:56:54.5341853Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_sparse_csr_cpu_float64 PASSED [ 86%] 2023-03-31T05:56:54.5342289Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_strided_cpu_float32 PASSED [ 86%] 2023-03-31T05:56:54.5342717Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_decimals_neg_3_layout_torch_strided_cpu_float64 PASSED [ 86%] 2023-03-31T05:56:54.5343122Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_coo_cpu_float32 PASSED [ 86%] 2023-03-31T05:56:54.5343616Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_coo_cpu_float64 PASSED [ 86%] 2023-03-31T05:56:54.5344027Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_csr_cpu_float32 PASSED [ 86%] 2023-03-31T05:56:54.5344435Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_sparse_csr_cpu_float64 PASSED [ 87%] 2023-03-31T05:56:54.5344826Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_strided_cpu_float32 PASSED [ 87%] 2023-03-31T05:56:54.5345234Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_round_layout_torch_strided_cpu_float64 PASSED [ 87%] 2023-03-31T05:56:54.5345685Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_coo_cpu_float32 PASSED [ 87%] 2023-03-31T05:56:54.5346095Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_coo_cpu_float64 PASSED [ 87%] 2023-03-31T05:56:54.5346485Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_csr_cpu_float32 PASSED [ 87%] 2023-03-31T05:56:54.5346886Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_sparse_csr_cpu_float64 PASSED [ 87%] 2023-03-31T05:56:54.5347292Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_strided_cpu_float32 PASSED [ 87%] 2023-03-31T05:56:54.5347677Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_rsqrt_layout_torch_strided_cpu_float64 PASSED [ 87%] 2023-03-31T05:56:54.5348110Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float16 PASSED [ 88%] 2023-03-31T05:56:54.5348516Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float32 PASSED [ 88%] 2023-03-31T05:56:54.5348917Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_coo_cpu_float64 PASSED [ 88%] 2023-03-31T05:56:54.5349304Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float16 PASSED [ 88%] 2023-03-31T05:56:54.5349708Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float32 PASSED [ 88%] 2023-03-31T05:56:54.5350109Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_sparse_csr_cpu_float64 PASSED [ 88%] 2023-03-31T05:56:54.5350510Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float16 PASSED [ 88%] 2023-03-31T05:56:54.5350899Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float32 PASSED [ 88%] 2023-03-31T05:56:54.5351298Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sgn_layout_torch_strided_cpu_float64 PASSED [ 89%] 2023-03-31T05:56:54.5351700Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_coo_cpu_float32 PASSED [ 89%] 2023-03-31T05:56:54.5352111Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_coo_cpu_float64 PASSED [ 89%] 2023-03-31T05:56:54.5352508Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_csr_cpu_float32 PASSED [ 89%] 2023-03-31T05:56:54.5352967Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_sparse_csr_cpu_float64 PASSED [ 89%] 2023-03-31T05:56:54.5353404Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_strided_cpu_float32 PASSED [ 89%] 2023-03-31T05:56:54.5353797Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sigmoid_layout_torch_strided_cpu_float64 PASSED [ 89%] 2023-03-31T05:56:54.5354203Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float16 PASSED [ 89%] 2023-03-31T05:56:54.5354608Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float32 PASSED [ 90%] 2023-03-31T05:56:54.5355018Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_coo_cpu_float64 PASSED [ 90%] 2023-03-31T05:56:54.5355412Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float16 PASSED [ 90%] 2023-03-31T05:56:54.5355821Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float32 PASSED [ 90%] 2023-03-31T05:56:54.5356228Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_sparse_csr_cpu_float64 PASSED [ 90%] 2023-03-31T05:56:54.5356638Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float16 PASSED [ 90%] 2023-03-31T05:56:54.5357030Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float32 PASSED [ 90%] 2023-03-31T05:56:54.5357478Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sign_layout_torch_strided_cpu_float64 PASSED [ 90%] 2023-03-31T05:56:54.5357885Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float16 PASSED [ 90%] 2023-03-31T05:56:54.5358302Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float32 PASSED [ 91%] 2023-03-31T05:56:54.5358701Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_coo_cpu_float64 PASSED [ 91%] 2023-03-31T05:56:54.5359114Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float16 PASSED [ 91%] 2023-03-31T05:56:54.5359526Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float32 PASSED [ 91%] 2023-03-31T05:56:54.5359972Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_sparse_csr_cpu_float64 PASSED [ 91%] 2023-03-31T05:56:54.5360371Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float16 PASSED [ 91%] 2023-03-31T05:56:54.5360785Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float32 PASSED [ 91%] 2023-03-31T05:56:54.5361195Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_signbit_layout_torch_strided_cpu_float64 PASSED [ 91%] 2023-03-31T05:56:54.5361589Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_coo_cpu_float32 PASSED [ 92%] 2023-03-31T05:56:54.5361997Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_coo_cpu_float64 PASSED [ 92%] 2023-03-31T05:56:54.5362406Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_csr_cpu_float32 PASSED [ 92%] 2023-03-31T05:56:54.5362808Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_sparse_csr_cpu_float64 PASSED [ 92%] 2023-03-31T05:56:54.5363353Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_strided_cpu_float32 PASSED [ 92%] 2023-03-31T05:56:54.5363753Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sin_layout_torch_strided_cpu_float64 PASSED [ 92%] 2023-03-31T05:56:54.5364156Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_coo_cpu_float32 PASSED [ 92%] 2023-03-31T05:56:54.5364568Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_coo_cpu_float64 PASSED [ 92%] 2023-03-31T05:56:54.5365012Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_csr_cpu_float32 PASSED [ 93%] 2023-03-31T05:56:54.5365451Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_sparse_csr_cpu_float64 PASSED [ 93%] 2023-03-31T05:56:54.5365857Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_strided_cpu_float32 PASSED [ 93%] 2023-03-31T05:56:54.5366260Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinc_layout_torch_strided_cpu_float64 PASSED [ 93%] 2023-03-31T05:56:54.5366650Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_coo_cpu_float32 PASSED [ 93%] 2023-03-31T05:56:54.5367055Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_coo_cpu_float64 PASSED [ 93%] 2023-03-31T05:56:54.5367459Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_csr_cpu_float32 PASSED [ 93%] 2023-03-31T05:56:54.5367865Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_sparse_csr_cpu_float64 PASSED [ 93%] 2023-03-31T05:56:54.5368256Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_strided_cpu_float32 PASSED [ 93%] 2023-03-31T05:56:54.5368660Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sinh_layout_torch_strided_cpu_float64 PASSED [ 94%] 2023-03-31T05:56:54.5369058Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_coo_cpu_float32 PASSED [ 94%] 2023-03-31T05:56:54.5369478Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_coo_cpu_float64 PASSED [ 94%] 2023-03-31T05:56:54.5369885Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_csr_cpu_float32 PASSED [ 94%] 2023-03-31T05:56:54.5370279Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_sparse_csr_cpu_float64 PASSED [ 94%] 2023-03-31T05:56:54.5370676Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_strided_cpu_float32 PASSED [ 94%] 2023-03-31T05:56:54.5371061Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_sqrt_layout_torch_strided_cpu_float64 PASSED [ 94%] 2023-03-31T05:56:54.5371467Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float16 PASSED [ 94%] 2023-03-31T05:56:54.5371877Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float32 PASSED [ 95%] 2023-03-31T05:56:54.5372330Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_coo_cpu_float64 PASSED [ 95%] 2023-03-31T05:56:54.5372725Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float16 PASSED [ 95%] 2023-03-31T05:56:54.5373126Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float32 PASSED [ 95%] 2023-03-31T05:56:54.5373530Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_sparse_csr_cpu_float64 PASSED [ 95%] 2023-03-31T05:56:54.5373938Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float16 PASSED [ 95%] 2023-03-31T05:56:54.5374329Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float32 PASSED [ 95%] 2023-03-31T05:56:54.5374730Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_square_layout_torch_strided_cpu_float64 PASSED [ 95%] 2023-03-31T05:56:54.5375133Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_coo_cpu_float32 PASSED [ 96%] 2023-03-31T05:56:54.5375532Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_coo_cpu_float64 PASSED [ 96%] 2023-03-31T05:56:54.5375919Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_csr_cpu_float32 PASSED [ 96%] 2023-03-31T05:56:54.5376316Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_sparse_csr_cpu_float64 PASSED [ 96%] 2023-03-31T05:56:54.5376721Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_strided_cpu_float32 PASSED [ 96%] 2023-03-31T05:56:54.5377138Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tan_layout_torch_strided_cpu_float64 PASSED [ 96%] 2023-03-31T05:56:54.5377569Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_coo_cpu_float32 PASSED [ 96%] 2023-03-31T05:56:54.5377975Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_coo_cpu_float64 PASSED [ 96%] 2023-03-31T05:56:54.5378380Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_csr_cpu_float32 PASSED [ 96%] 2023-03-31T05:56:54.5378770Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_sparse_csr_cpu_float64 PASSED [ 97%] 2023-03-31T05:56:54.5379169Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_strided_cpu_float32 PASSED [ 97%] 2023-03-31T05:56:54.5379567Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_tanh_layout_torch_strided_cpu_float64 PASSED [ 97%] 2023-03-31T05:56:54.5379967Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_coo_cpu_float32 PASSED [ 97%] 2023-03-31T05:56:54.5380365Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_coo_cpu_float64 PASSED [ 97%] 2023-03-31T05:56:54.5380768Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_csr_cpu_float32 PASSED [ 97%] 2023-03-31T05:56:54.5381177Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_sparse_csr_cpu_float64 PASSED [ 97%] 2023-03-31T05:56:54.5381611Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_strided_cpu_float32 PASSED [ 97%] 2023-03-31T05:56:54.5382005Z test_maskedtensor.py::TestOperatorsCPU::test_unary_core_trunc_layout_torch_strided_cpu_float64 PASSED [ 98%] 2023-03-31T05:56:54.5382360Z test_maskedtensor.py::TestBasicsCPU::test_add_cpu PASSED [ 98%] 2023-03-31T05:56:54.5382686Z test_maskedtensor.py::TestBasicsCPU::test_contiguous_cpu PASSED [ 98%] 2023-03-31T05:56:54.5383010Z test_maskedtensor.py::TestBasicsCPU::test_diff_dim_cpu PASSED [ 98%] 2023-03-31T05:56:54.5383412Z test_maskedtensor.py::TestBasicsCPU::test_diff_layouts_cpu PASSED [ 98%] 2023-03-31T05:56:54.5383748Z test_maskedtensor.py::TestBasicsCPU::test_diff_sizes_cpu PASSED [ 98%] 2023-03-31T05:56:54.5384121Z test_maskedtensor.py::TestBasicsCPU::test_grad_warning_cpu PASSED [ 98%] 2023-03-31T05:56:54.5384454Z test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_coo_values_cpu PASSED [ 98%] 2023-03-31T05:56:54.5384807Z test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_csr_values_cpu PASSED [ 99%] 2023-03-31T05:56:54.5385153Z test_maskedtensor.py::TestBasicsCPU::test_invalid_sparse_layout_cpu PASSED [ 99%] 2023-03-31T05:56:54.5385488Z test_maskedtensor.py::TestBasicsCPU::test_invalid_tensor_inputs_cpu PASSED [ 99%] 2023-03-31T05:56:54.5385825Z test_maskedtensor.py::TestBasicsCPU::test_softmax_cpu PASSED [ 99%] 2023-03-31T05:56:54.5386164Z test_maskedtensor.py::TestBasicsCPU::test_to_dense_and_sparse_coo_cpu PASSED [ 99%] 2023-03-31T05:56:54.5386511Z test_maskedtensor.py::TestBasicsCPU::test_to_dense_and_sparse_csr_cpu PASSED [ 99%] 2023-03-31T05:56:54.5386837Z test_maskedtensor.py::TestBasicsCPU::test_to_dense_cpu PASSED [ 99%] 2023-03-31T05:56:54.5387164Z test_maskedtensor.py::TestBasicsCPU::test_to_sparse_cpu PASSED [ 99%] 2023-03-31T05:56:54.5387490Z test_maskedtensor.py::TestBasicsCPU::test_where_cpu PASSED [100%] 2023-03-31T05:56:54.5387675Z 2023-03-31T05:56:54.5388168Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-53d6bd0ec99cdfe0.xml - 2023-03-31T05:56:54.5388561Z ======================== 829 passed, 3 skipped in 2.80s ======================== 2023-03-31T05:56:54.5388885Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:56:54.5389090Z 2023-03-31T05:56:54.5389477Z ##[endgroup] 2023-03-31T05:56:54.5389931Z FINISHED PRINTING LOG FILE of test_maskedtensor (/var/lib/jenkins/workspace/test/test-reports/test_maskedtensor_6gkgygrf.log) 2023-03-31T05:56:54.5390204Z 2023-03-31T05:56:56.5609967Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:56:56.5637474Z Ignoring disabled issues: [] 2023-03-31T05:56:56.5806029Z Running test_package ... [2023-03-31 05:56:56.580231] 2023-03-31T05:56:56.5807685Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_package.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:56:56.580541] 2023-03-31T05:57:01.0754390Z 2023-03-31T05:57:01.0755153Z Expand the folded group to see the log file of test_package 2023-03-31T05:57:01.0756198Z ##[group]PRINTING LOG FILE of test_package (/var/lib/jenkins/workspace/test/test-reports/test_package__axiw8pm.log) 2023-03-31T05:57:01.0757301Z Test results will be stored in test-reports/python-pytest/test_package/test_package-ee02c66f15df73e0.xml 2023-03-31T05:57:01.0757846Z ============================= test session starts ============================== 2023-03-31T05:57:01.0758534Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:01.0758966Z cachedir: .pytest_cache 2023-03-31T05:57:01.0759697Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:01.0760615Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:01.0761417Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:01.0761959Z collecting ... collected 133 items 2023-03-31T05:57:01.0782616Z Running 133 items in this shard: test/test_package.py::TestResources::test_importer_access, test/test_package.py::TestResources::test_package_resource_access, test/test_package.py::TestResources::test_resource_access_by_path, test/test_package.py::TestResources::test_resource_reader, test/test_package.py::ModelTest::test_model_save, test/test_package.py::ModelTest::test_resnet, test/test_package.py::ModelTest::test_script_resnet, test/test_package.py::TestDependencyAPI::test_allow_empty_with_error, test/test_package.py::TestDependencyAPI::test_broken_dependency, test/test_package.py::TestDependencyAPI::test_deny, test/test_package.py::TestDependencyAPI::test_deny_glob, test/test_package.py::TestDependencyAPI::test_extern, test/test_package.py::TestDependencyAPI::test_extern_glob, test/test_package.py::TestDependencyAPI::test_extern_glob_allow_empty, test/test_package.py::TestDependencyAPI::test_externing_c_extension, test/test_package.py::TestDependencyAPI::test_implicit_intern, test/test_package.py::TestDependencyAPI::test_intern_error, test/test_package.py::TestDependencyAPI::test_invalid_import, test/test_package.py::TestDependencyAPI::test_mock, test/test_package.py::TestDependencyAPI::test_mock_glob, test/test_package.py::TestDependencyAPI::test_mock_glob_allow_empty, test/test_package.py::TestDependencyAPI::test_pickle_mocked, test/test_package.py::TestDependencyAPI::test_pickle_mocked_all, test/test_package.py::TestDependencyAPI::test_repackage_mocked_module, test/test_package.py::TestMangling::test_demangle_base, test/test_package.py::TestMangling::test_demangler_multiple_manglers, test/test_package.py::TestMangling::test_is_mangled, test/test_package.py::TestMangling::test_mangle_empty_errors, test/test_package.py::TestMangling::test_mangle_prefix, test/test_package.py::TestMangling::test_mangler_is_consistent, test/test_package.py::TestMangling::test_package_mangler, test/test_package.py::TestMangling::test_roundtrip_mangling, test/test_package.py::TestMangling::test_unique_manglers, test/test_package.py::TestMangling::test_unique_module_names, test/test_package.py::TestMisc::test_dunder_package_present, test/test_package.py::TestMisc::test_dunder_package_works_from_package, test/test_package.py::TestMisc::test_exporter_content_lists, test/test_package.py::TestMisc::test_file_structure, test/test_package.py::TestMisc::test_file_structure_has_file, test/test_package.py::TestMisc::test_inspect_class, test/test_package.py::TestMisc::test_is_from_package, test/test_package.py::TestMisc::test_load_python_version_from_package, test/test_package.py::TestMisc::test_loaders_that_remap_files_work_ok, test/test_package.py::TestMisc::test_python_version, test/test_package.py::TestMisc::test_std_lib_sys_hackery_checks, test/test_package.py::DirectoryReaderTest::test_importer_access, test/test_package.py::DirectoryReaderTest::test_loading_has_record, test/test_package.py::DirectoryReaderTest::test_loading_module, test/test_package.py::DirectoryReaderTest::test_loading_pickle, test/test_package.py::DirectoryReaderTest::test_package_resource_access, test/test_package.py::DirectoryReaderTest::test_resource_access_by_path, test/test_package.py::DirectoryReaderTest::test_resource_reader, test/test_package.py::DirectoryReaderTest::test_scriptobject_failure_message, test/test_package.py::TestImporter::test_ordered_importer_basic, test/test_package.py::TestImporter::test_ordered_importer_whichmodule, test/test_package.py::TestImporter::test_package_importer_whichmodule_no_dunder_module, test/test_package.py::TestImporter::test_single_ordered_importer, test/test_package.py::TestImporter::test_sys_importer, test/test_package.py::TestImporter::test_sys_importer_roundtrip, test/test_package.py::TestGlobGroup::test_exclude, test/test_package.py::TestGlobGroup::test_exclude_from_all, test/test_package.py::TestGlobGroup::test_invalid_raw, test/test_package.py::TestGlobGroup::test_list_include_exclude, test/test_package.py::TestGlobGroup::test_one_star, test/test_package.py::TestGlobGroup::test_one_star_middle, test/test_package.py::TestGlobGroup::test_one_star_multiple_in_component, test/test_package.py::TestGlobGroup::test_one_star_partial, test/test_package.py::TestGlobGroup::test_one_star_partial_extension, test/test_package.py::TestGlobGroup::test_raw_two_star, test/test_package.py::TestGlobGroup::test_two_star, test/test_package.py::TestGlobGroup::test_two_star_end, test/test_package.py::TestGlobGroup::test_two_star_middle, test/test_package.py::TestGlobGroup::test_two_star_multiple, test/test_package.py::TestPackageScript::test_different_package_interface, test/test_package.py::TestPackageScript::test_different_package_script_class, test/test_package.py::TestPackageScript::test_load_shared_scriptmodules, test/test_package.py::TestPackageScript::test_load_shared_tensors, test/test_package.py::TestPackageScript::test_load_shared_tensors_repackaged, test/test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules, test/test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules_shared_code, test/test_package.py::TestPackageScript::test_package_interface, test/test_package.py::TestPackageScript::test_package_script_class, test/test_package.py::TestPackageScript::test_package_script_class_referencing_self, test/test_package.py::TestPackageScript::test_save_eager_mods_sharing_scriptmodule, test/test_package.py::TestPackageScript::test_save_independent_scriptmodules, test/test_package.py::TestPackageScript::test_save_repeat_scriptmodules, test/test_package.py::TestPackageScript::test_save_scriptmodule, test/test_package.py::TestPackageScript::test_save_scriptmodule_file, test/test_package.py::TestPackageScript::test_save_scriptmodule_only_necessary_code, test/test_package.py::TestPackageScript::test_save_scriptmodule_with_submods, test/test_package.py::TestPackageScript::test_save_scriptmodules_in_container, test/test_package.py::TestPackageScript::test_save_scriptmodules_submod_redefinition, test/test_package.py::TestPackageScript::test_save_shared_tensors, test/test_package.py::TestPackageScript::test_saving_and_scripting_packaged_mod, test/test_package.py::TestPackageScript::test_scriptmodules_repeat_save, test/test_package.py::TestPackageScript::test_tensor_sharing_pickle, test/test_package.py::TestSaveLoad::test_bad_dunder_imports, test/test_package.py::TestSaveLoad::test_dunder_imports, test/test_package.py::TestSaveLoad::test_exporting_mismatched_code, test/test_package.py::TestSaveLoad::test_pickle, test/test_package.py::TestSaveLoad::test_save_imported_module, test/test_package.py::TestSaveLoad::test_save_imported_module_using_package_importer, test/test_package.py::TestSaveLoad::test_save_module, test/test_package.py::TestSaveLoad::test_save_module_binary, test/test_package.py::TestSaveLoad::test_saving_source, test/test_package.py::TestSaveLoad::test_saving_string, test/test_package.py::TestRepackage::test_repackage_import_indirectly_via_parent_module, test/test_package.py::TestPackageFX::test_package_fx_custom_tracer, test/test_package.py::TestPackageFX::test_package_fx_package, test/test_package.py::TestPackageFX::test_package_fx_simple, test/test_package.py::TestPackageFX::test_package_fx_with_imports, test/test_package.py::TestPackageFX::test_package_then_fx, test/test_package.py::TestDependencyHooks::test_extern_and_mock_hook, test/test_package.py::TestDependencyHooks::test_multiple_extern_hooks, test/test_package.py::TestDependencyHooks::test_multiple_mock_hooks, test/test_package.py::TestDependencyHooks::test_remove_hooks, test/test_package.py::TestDependencyHooks::test_single_hook, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_fx_module, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_nn_module, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_torchscript_module, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestDiGraph::test_all_paths, test/test_package.py::TestDiGraph::test_contains, test/test_package.py::TestDiGraph::test_contains_non_hashable, test/test_package.py::TestDiGraph::test_edges, test/test_package.py::TestDiGraph::test_forward_closure, test/test_package.py::TestDiGraph::test_iter, test/test_package.py::TestDiGraph::test_node_attr_update, test/test_package.py::TestDiGraph::test_node_attrs, test/test_package.py::TestDiGraph::test_predecessor_not_in_graph, test/test_package.py::TestDiGraph::test_predecessors, test/test_package.py::TestDiGraph::test_successor_not_in_graph, test/test_package.py::TestDiGraph::test_successors 2023-03-31T05:57:01.0801162Z 2023-03-31T05:57:01.0801819Z test_package.py::TestResources::test_importer_access <- test/package/test_resources.py PASSED [ 0%] 2023-03-31T05:57:01.0803246Z test_package.py::TestResources::test_package_resource_access <- test/package/test_resources.py PASSED [ 1%] 2023-03-31T05:57:01.0804330Z test_package.py::TestResources::test_resource_access_by_path <- test/package/test_resources.py PASSED [ 2%] 2023-03-31T05:57:01.0805157Z test_package.py::TestResources::test_resource_reader <- test/package/test_resources.py PASSED [ 3%] 2023-03-31T05:57:01.0806365Z test_package.py::ModelTest::test_model_save <- test/package/test_model.py SKIPPED (Does not work with recent torchvision, see https://github.com/pytorch/pytorch/issues/81115) [ 3%] 2023-03-31T05:57:01.0807762Z test_package.py::ModelTest::test_resnet <- test/package/test_model.py SKIPPED (Does not work with recent torchvision, see https://github.com/pytorch/pytorch/issues/81115) [ 4%] 2023-03-31T05:57:01.0808855Z test_package.py::ModelTest::test_script_resnet <- test/package/test_model.py SKIPPED (Does not work with recent torchvision, see https://github.com/pytorch/pytorch/issues/81115) [ 5%] 2023-03-31T05:57:01.0809901Z test_package.py::TestDependencyAPI::test_allow_empty_with_error <- test/package/test_dependency_api.py PASSED [ 6%] 2023-03-31T05:57:01.0810957Z test_package.py::TestDependencyAPI::test_broken_dependency <- test/package/test_dependency_api.py PASSED [ 6%] 2023-03-31T05:57:01.0811957Z test_package.py::TestDependencyAPI::test_deny <- test/package/test_dependency_api.py PASSED [ 7%] 2023-03-31T05:57:01.0812937Z test_package.py::TestDependencyAPI::test_deny_glob <- test/package/test_dependency_api.py PASSED [ 8%] 2023-03-31T05:57:01.0813996Z test_package.py::TestDependencyAPI::test_extern <- test/package/test_dependency_api.py PASSED [ 9%] 2023-03-31T05:57:01.0814963Z test_package.py::TestDependencyAPI::test_extern_glob <- test/package/test_dependency_api.py PASSED [ 9%] 2023-03-31T05:57:01.0816068Z test_package.py::TestDependencyAPI::test_extern_glob_allow_empty <- test/package/test_dependency_api.py PASSED [ 10%] 2023-03-31T05:57:01.0817048Z test_package.py::TestDependencyAPI::test_externing_c_extension <- test/package/test_dependency_api.py PASSED [ 11%] 2023-03-31T05:57:01.0817933Z test_package.py::TestDependencyAPI::test_implicit_intern <- test/package/test_dependency_api.py PASSED [ 12%] 2023-03-31T05:57:01.0818823Z test_package.py::TestDependencyAPI::test_intern_error <- test/package/test_dependency_api.py PASSED [ 12%] 2023-03-31T05:57:01.0819733Z test_package.py::TestDependencyAPI::test_invalid_import <- test/package/test_dependency_api.py PASSED [ 13%] 2023-03-31T05:57:01.0820569Z test_package.py::TestDependencyAPI::test_mock <- test/package/test_dependency_api.py PASSED [ 14%] 2023-03-31T05:57:01.0821568Z test_package.py::TestDependencyAPI::test_mock_glob <- test/package/test_dependency_api.py PASSED [ 15%] 2023-03-31T05:57:01.0822438Z test_package.py::TestDependencyAPI::test_mock_glob_allow_empty <- test/package/test_dependency_api.py PASSED [ 15%] 2023-03-31T05:57:01.0823634Z test_package.py::TestDependencyAPI::test_pickle_mocked <- test/package/test_dependency_api.py PASSED [ 16%] 2023-03-31T05:57:01.0824945Z test_package.py::TestDependencyAPI::test_pickle_mocked_all <- test/package/test_dependency_api.py PASSED [ 17%] 2023-03-31T05:57:01.0825936Z test_package.py::TestDependencyAPI::test_repackage_mocked_module <- test/package/test_dependency_api.py PASSED [ 18%] 2023-03-31T05:57:01.0826789Z test_package.py::TestMangling::test_demangle_base <- test/package/test_mangling.py PASSED [ 18%] 2023-03-31T05:57:01.0827608Z test_package.py::TestMangling::test_demangler_multiple_manglers <- test/package/test_mangling.py PASSED [ 19%] 2023-03-31T05:57:01.0828610Z test_package.py::TestMangling::test_is_mangled <- test/package/test_mangling.py PASSED [ 20%] 2023-03-31T05:57:01.0829506Z test_package.py::TestMangling::test_mangle_empty_errors <- test/package/test_mangling.py PASSED [ 21%] 2023-03-31T05:57:01.0830288Z test_package.py::TestMangling::test_mangle_prefix <- test/package/test_mangling.py PASSED [ 21%] 2023-03-31T05:57:01.0831493Z test_package.py::TestMangling::test_mangler_is_consistent <- test/package/test_mangling.py PASSED [ 22%] 2023-03-31T05:57:01.0832392Z test_package.py::TestMangling::test_package_mangler <- test/package/test_mangling.py PASSED [ 23%] 2023-03-31T05:57:01.0833227Z test_package.py::TestMangling::test_roundtrip_mangling <- test/package/test_mangling.py PASSED [ 24%] 2023-03-31T05:57:01.0834043Z test_package.py::TestMangling::test_unique_manglers <- test/package/test_mangling.py PASSED [ 24%] 2023-03-31T05:57:01.0834883Z test_package.py::TestMangling::test_unique_module_names <- test/package/test_mangling.py PASSED [ 25%] 2023-03-31T05:57:01.0835726Z test_package.py::TestMisc::test_dunder_package_present <- test/package/test_misc.py PASSED [ 26%] 2023-03-31T05:57:01.0836578Z test_package.py::TestMisc::test_dunder_package_works_from_package <- test/package/test_misc.py PASSED [ 27%] 2023-03-31T05:57:01.0837411Z test_package.py::TestMisc::test_exporter_content_lists <- test/package/test_misc.py PASSED [ 27%] 2023-03-31T05:57:01.0838210Z test_package.py::TestMisc::test_file_structure <- test/package/test_misc.py PASSED [ 28%] 2023-03-31T05:57:01.0838939Z test_package.py::TestMisc::test_file_structure_has_file <- test/package/test_misc.py PASSED [ 29%] 2023-03-31T05:57:01.0839597Z test_package.py::TestMisc::test_inspect_class <- test/package/test_misc.py PASSED [ 30%] 2023-03-31T05:57:01.0840352Z test_package.py::TestMisc::test_is_from_package <- test/package/test_misc.py PASSED [ 30%] 2023-03-31T05:57:01.0841193Z test_package.py::TestMisc::test_load_python_version_from_package <- test/package/test_misc.py PASSED [ 31%] 2023-03-31T05:57:01.0842162Z test_package.py::TestMisc::test_loaders_that_remap_files_work_ok <- test/package/test_misc.py PASSED [ 32%] 2023-03-31T05:57:01.0843342Z test_package.py::TestMisc::test_python_version <- test/package/test_misc.py PASSED [ 33%] 2023-03-31T05:57:01.0844205Z test_package.py::TestMisc::test_std_lib_sys_hackery_checks <- test/package/test_misc.py PASSED [ 33%] 2023-03-31T05:57:01.0845039Z test_package.py::DirectoryReaderTest::test_importer_access <- test/package/test_directory_reader.py PASSED [ 34%] 2023-03-31T05:57:01.0846156Z test_package.py::DirectoryReaderTest::test_loading_has_record <- test/package/test_directory_reader.py PASSED [ 35%] 2023-03-31T05:57:01.0847054Z test_package.py::DirectoryReaderTest::test_loading_module <- test/package/test_directory_reader.py PASSED [ 36%] 2023-03-31T05:57:01.0848445Z test_package.py::DirectoryReaderTest::test_loading_pickle <- test/package/test_directory_reader.py SKIPPED (Does not work with latest TorchVision, see https://github.com/pytorch/pytorch/issues/81115) [ 36%] 2023-03-31T05:57:01.0849554Z test_package.py::DirectoryReaderTest::test_package_resource_access <- test/package/test_directory_reader.py PASSED [ 37%] 2023-03-31T05:57:01.0850679Z test_package.py::DirectoryReaderTest::test_resource_access_by_path <- test/package/test_directory_reader.py PASSED [ 38%] 2023-03-31T05:57:01.0851665Z test_package.py::DirectoryReaderTest::test_resource_reader <- test/package/test_directory_reader.py PASSED [ 39%] 2023-03-31T05:57:01.0852824Z test_package.py::DirectoryReaderTest::test_scriptobject_failure_message <- test/package/test_directory_reader.py PASSED [ 39%] 2023-03-31T05:57:01.0853962Z test_package.py::TestImporter::test_ordered_importer_basic <- test/package/test_importer.py PASSED [ 40%] 2023-03-31T05:57:01.0854792Z test_package.py::TestImporter::test_ordered_importer_whichmodule <- test/package/test_importer.py PASSED [ 41%] 2023-03-31T05:57:01.0855826Z test_package.py::TestImporter::test_package_importer_whichmodule_no_dunder_module <- test/package/test_importer.py PASSED [ 42%] 2023-03-31T05:57:01.0856850Z test_package.py::TestImporter::test_single_ordered_importer <- test/package/test_importer.py PASSED [ 42%] 2023-03-31T05:57:01.0857704Z test_package.py::TestImporter::test_sys_importer <- test/package/test_importer.py PASSED [ 43%] 2023-03-31T05:57:01.0858552Z test_package.py::TestImporter::test_sys_importer_roundtrip <- test/package/test_importer.py PASSED [ 44%] 2023-03-31T05:57:01.0859519Z test_package.py::TestGlobGroup::test_exclude <- test/package/test_glob_group.py PASSED [ 45%] 2023-03-31T05:57:01.0860498Z test_package.py::TestGlobGroup::test_exclude_from_all <- test/package/test_glob_group.py PASSED [ 45%] 2023-03-31T05:57:01.0861290Z test_package.py::TestGlobGroup::test_invalid_raw <- test/package/test_glob_group.py PASSED [ 46%] 2023-03-31T05:57:01.0862169Z test_package.py::TestGlobGroup::test_list_include_exclude <- test/package/test_glob_group.py PASSED [ 47%] 2023-03-31T05:57:01.0863145Z test_package.py::TestGlobGroup::test_one_star <- test/package/test_glob_group.py PASSED [ 48%] 2023-03-31T05:57:01.0864111Z test_package.py::TestGlobGroup::test_one_star_middle <- test/package/test_glob_group.py PASSED [ 48%] 2023-03-31T05:57:01.0864970Z test_package.py::TestGlobGroup::test_one_star_multiple_in_component <- test/package/test_glob_group.py PASSED [ 49%] 2023-03-31T05:57:01.0865961Z test_package.py::TestGlobGroup::test_one_star_partial <- test/package/test_glob_group.py PASSED [ 50%] 2023-03-31T05:57:01.0866890Z test_package.py::TestGlobGroup::test_one_star_partial_extension <- test/package/test_glob_group.py PASSED [ 51%] 2023-03-31T05:57:01.0867747Z test_package.py::TestGlobGroup::test_raw_two_star <- test/package/test_glob_group.py PASSED [ 51%] 2023-03-31T05:57:01.0868697Z test_package.py::TestGlobGroup::test_two_star <- test/package/test_glob_group.py PASSED [ 52%] 2023-03-31T05:57:01.0869536Z test_package.py::TestGlobGroup::test_two_star_end <- test/package/test_glob_group.py PASSED [ 53%] 2023-03-31T05:57:01.0870517Z test_package.py::TestGlobGroup::test_two_star_middle <- test/package/test_glob_group.py PASSED [ 54%] 2023-03-31T05:57:01.0871357Z test_package.py::TestGlobGroup::test_two_star_multiple <- test/package/test_glob_group.py PASSED [ 54%] 2023-03-31T05:57:01.0872220Z test_package.py::TestPackageScript::test_different_package_interface <- test/package/test_package_script.py PASSED [ 55%] 2023-03-31T05:57:01.0873150Z test_package.py::TestPackageScript::test_different_package_script_class <- test/package/test_package_script.py PASSED [ 56%] 2023-03-31T05:57:01.0874080Z test_package.py::TestPackageScript::test_load_shared_scriptmodules <- test/package/test_package_script.py PASSED [ 57%] 2023-03-31T05:57:01.0874916Z test_package.py::TestPackageScript::test_load_shared_tensors <- test/package/test_package_script.py PASSED [ 57%] 2023-03-31T05:57:01.0875871Z test_package.py::TestPackageScript::test_load_shared_tensors_repackaged <- test/package/test_package_script.py PASSED [ 58%] 2023-03-31T05:57:01.0876885Z test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules <- test/package/test_package_script.py PASSED [ 59%] 2023-03-31T05:57:01.0877910Z test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules_shared_code <- test/package/test_package_script.py PASSED [ 60%] 2023-03-31T05:57:01.0878827Z test_package.py::TestPackageScript::test_package_interface <- test/package/test_package_script.py PASSED [ 60%] 2023-03-31T05:57:01.0879742Z test_package.py::TestPackageScript::test_package_script_class <- test/package/test_package_script.py PASSED [ 61%] 2023-03-31T05:57:01.0880795Z test_package.py::TestPackageScript::test_package_script_class_referencing_self <- test/package/test_package_script.py PASSED [ 62%] 2023-03-31T05:57:01.0881804Z test_package.py::TestPackageScript::test_save_eager_mods_sharing_scriptmodule <- test/package/test_package_script.py PASSED [ 63%] 2023-03-31T05:57:01.0882996Z test_package.py::TestPackageScript::test_save_independent_scriptmodules <- test/package/test_package_script.py PASSED [ 63%] 2023-03-31T05:57:01.0884196Z test_package.py::TestPackageScript::test_save_repeat_scriptmodules <- test/package/test_package_script.py PASSED [ 64%] 2023-03-31T05:57:01.0885259Z test_package.py::TestPackageScript::test_save_scriptmodule <- test/package/test_package_script.py PASSED [ 65%] 2023-03-31T05:57:01.0886269Z test_package.py::TestPackageScript::test_save_scriptmodule_file <- test/package/test_package_script.py PASSED [ 66%] 2023-03-31T05:57:01.0887414Z test_package.py::TestPackageScript::test_save_scriptmodule_only_necessary_code <- test/package/test_package_script.py PASSED [ 66%] 2023-03-31T05:57:01.0888463Z test_package.py::TestPackageScript::test_save_scriptmodule_with_submods <- test/package/test_package_script.py PASSED [ 67%] 2023-03-31T05:57:01.0889528Z test_package.py::TestPackageScript::test_save_scriptmodules_in_container <- test/package/test_package_script.py PASSED [ 68%] 2023-03-31T05:57:01.0890584Z test_package.py::TestPackageScript::test_save_scriptmodules_submod_redefinition <- test/package/test_package_script.py PASSED [ 69%] 2023-03-31T05:57:01.0891682Z test_package.py::TestPackageScript::test_save_shared_tensors <- test/package/test_package_script.py PASSED [ 69%] 2023-03-31T05:57:01.0892605Z test_package.py::TestPackageScript::test_saving_and_scripting_packaged_mod <- test/package/test_package_script.py PASSED [ 70%] 2023-03-31T05:57:01.0893764Z test_package.py::TestPackageScript::test_scriptmodules_repeat_save <- test/package/test_package_script.py PASSED [ 71%] 2023-03-31T05:57:01.0894803Z test_package.py::TestPackageScript::test_tensor_sharing_pickle <- test/package/test_package_script.py PASSED [ 72%] 2023-03-31T05:57:01.0895664Z test_package.py::TestSaveLoad::test_bad_dunder_imports <- test/package/test_save_load.py PASSED [ 72%] 2023-03-31T05:57:01.0896590Z test_package.py::TestSaveLoad::test_dunder_imports <- test/package/test_save_load.py PASSED [ 73%] 2023-03-31T05:57:01.0897570Z test_package.py::TestSaveLoad::test_exporting_mismatched_code <- test/package/test_save_load.py PASSED [ 74%] 2023-03-31T05:57:01.0898524Z test_package.py::TestSaveLoad::test_pickle <- test/package/test_save_load.py PASSED [ 75%] 2023-03-31T05:57:01.0899501Z test_package.py::TestSaveLoad::test_save_imported_module <- test/package/test_save_load.py PASSED [ 75%] 2023-03-31T05:57:01.0900499Z test_package.py::TestSaveLoad::test_save_imported_module_using_package_importer <- test/package/test_save_load.py PASSED [ 76%] 2023-03-31T05:57:01.0901341Z test_package.py::TestSaveLoad::test_save_module <- test/package/test_save_load.py PASSED [ 77%] 2023-03-31T05:57:01.0902254Z test_package.py::TestSaveLoad::test_save_module_binary <- test/package/test_save_load.py PASSED [ 78%] 2023-03-31T05:57:01.0903322Z test_package.py::TestSaveLoad::test_saving_source <- test/package/test_save_load.py PASSED [ 78%] 2023-03-31T05:57:01.0904059Z test_package.py::TestSaveLoad::test_saving_string <- test/package/test_save_load.py PASSED [ 79%] 2023-03-31T05:57:01.0904959Z test_package.py::TestRepackage::test_repackage_import_indirectly_via_parent_module <- test/package/test_repackage.py PASSED [ 80%] 2023-03-31T05:57:01.0905864Z test_package.py::TestPackageFX::test_package_fx_custom_tracer <- test/package/test_package_fx.py PASSED [ 81%] 2023-03-31T05:57:01.0906858Z test_package.py::TestPackageFX::test_package_fx_package <- test/package/test_package_fx.py PASSED [ 81%] 2023-03-31T05:57:01.0907788Z test_package.py::TestPackageFX::test_package_fx_simple <- test/package/test_package_fx.py PASSED [ 82%] 2023-03-31T05:57:01.0908805Z test_package.py::TestPackageFX::test_package_fx_with_imports <- test/package/test_package_fx.py PASSED [ 83%] 2023-03-31T05:57:01.0909846Z test_package.py::TestPackageFX::test_package_then_fx <- test/package/test_package_fx.py PASSED [ 84%] 2023-03-31T05:57:01.0910774Z test_package.py::TestDependencyHooks::test_extern_and_mock_hook <- test/package/test_dependency_hooks.py PASSED [ 84%] 2023-03-31T05:57:01.0911752Z test_package.py::TestDependencyHooks::test_multiple_extern_hooks <- test/package/test_dependency_hooks.py PASSED [ 85%] 2023-03-31T05:57:01.0912842Z test_package.py::TestDependencyHooks::test_multiple_mock_hooks <- test/package/test_dependency_hooks.py PASSED [ 86%] 2023-03-31T05:57:01.0913718Z test_package.py::TestDependencyHooks::test_remove_hooks <- test/package/test_dependency_hooks.py PASSED [ 87%] 2023-03-31T05:57:01.0914767Z test_package.py::TestDependencyHooks::test_single_hook <- test/package/test_dependency_hooks.py PASSED [ 87%] 2023-03-31T05:57:01.0915836Z test_package.py::TestLoadBCPackages::test_load_bc_packages_fx_module <- test/package/test_load_bc_packages.py PASSED [ 88%] 2023-03-31T05:57:01.0916871Z test_package.py::TestLoadBCPackages::test_load_bc_packages_nn_module <- test/package/test_load_bc_packages.py PASSED [ 89%] 2023-03-31T05:57:01.0918061Z test_package.py::TestLoadBCPackages::test_load_bc_packages_torchscript_module <- test/package/test_load_bc_packages.py PASSED [ 90%] 2023-03-31T05:57:01.0919972Z test_package.py::TestAnalyze::test_trace_dependencies <- test/package/test_analyze.py SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/81213 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 90%] 2023-03-31T05:57:01.0921510Z test_package.py::TestDiGraph::test_all_paths <- test/package/test_digraph.py PASSED [ 91%] 2023-03-31T05:57:01.0922320Z test_package.py::TestDiGraph::test_contains <- test/package/test_digraph.py PASSED [ 92%] 2023-03-31T05:57:01.0923361Z test_package.py::TestDiGraph::test_contains_non_hashable <- test/package/test_digraph.py PASSED [ 93%] 2023-03-31T05:57:01.0924120Z test_package.py::TestDiGraph::test_edges <- test/package/test_digraph.py PASSED [ 93%] 2023-03-31T05:57:01.0925038Z test_package.py::TestDiGraph::test_forward_closure <- test/package/test_digraph.py PASSED [ 94%] 2023-03-31T05:57:01.0925871Z test_package.py::TestDiGraph::test_iter <- test/package/test_digraph.py PASSED [ 95%] 2023-03-31T05:57:01.0926806Z test_package.py::TestDiGraph::test_node_attr_update <- test/package/test_digraph.py PASSED [ 96%] 2023-03-31T05:57:01.0927832Z test_package.py::TestDiGraph::test_node_attrs <- test/package/test_digraph.py PASSED [ 96%] 2023-03-31T05:57:01.0928713Z test_package.py::TestDiGraph::test_predecessor_not_in_graph <- test/package/test_digraph.py PASSED [ 97%] 2023-03-31T05:57:01.0929531Z test_package.py::TestDiGraph::test_predecessors <- test/package/test_digraph.py PASSED [ 98%] 2023-03-31T05:57:01.0930359Z test_package.py::TestDiGraph::test_successor_not_in_graph <- test/package/test_digraph.py PASSED [ 99%] 2023-03-31T05:57:01.0932067Z test_package.py::TestDiGraph::test_successors <- test/package/test_digraph.py PASSED [100%]/var/lib/jenkins/workspace/test/package/package_a/test_all_leaf_modules_tracer.py:6: PytestCollectionWarning: cannot collect test class 'TestAllLeafModulesTracer' because it has a __init__ constructor (from: test/test_package.py) 2023-03-31T05:57:01.0933031Z class TestAllLeafModulesTracer(Tracer): 2023-03-31T05:57:01.0934310Z /var/lib/jenkins/workspace/test/package/package_a/test_nn_module.py:6: PytestCollectionWarning: cannot collect test class 'TestNnModule' because it has a __init__ constructor (from: test/test_package.py) 2023-03-31T05:57:01.0935083Z class TestNnModule(torch.nn.Module): 2023-03-31T05:57:01.0935309Z 2023-03-31T05:57:01.0935316Z 2023-03-31T05:57:01.0935976Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_package/test_package-ee02c66f15df73e0.xml - 2023-03-31T05:57:01.0936579Z ======================== 128 passed, 5 skipped in 2.36s ======================== 2023-03-31T05:57:01.0937125Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:01.0937475Z 2023-03-31T05:57:01.0937982Z ##[endgroup] 2023-03-31T05:57:01.0938840Z FINISHED PRINTING LOG FILE of test_package (/var/lib/jenkins/workspace/test/test-reports/test_package__axiw8pm.log) 2023-03-31T05:57:01.0939286Z 2023-03-31T05:57:03.1520569Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:03.1547952Z Ignoring disabled issues: [] 2023-03-31T05:57:03.1716235Z Running functorch/test_vmap_registrations ... [2023-03-31 05:57:03.171287] 2023-03-31T05:57:03.1718853Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'functorch/test_vmap_registrations.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:03.171605] 2023-03-31T05:57:08.0523697Z 2023-03-31T05:57:08.0524308Z Expand the folded group to see the log file of functorch/test_vmap_registrations 2023-03-31T05:57:08.0525317Z ##[group]PRINTING LOG FILE of functorch/test_vmap_registrations (/var/lib/jenkins/workspace/test/test-reports/functorch-test_vmap_registrations_w8ccd5em.log) 2023-03-31T05:57:08.0531470Z Test results will be stored in test-reports/python-pytest/functorch.test_vmap_registrations/functorch.test_vmap_registrations-1e483d8efa4950e7.xml 2023-03-31T05:57:08.0533173Z ============================= test session starts ============================== 2023-03-31T05:57:08.0533943Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:08.0534403Z cachedir: .pytest_cache 2023-03-31T05:57:08.0535152Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:08.0535794Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:08.0536536Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:08.0537084Z collecting ... collected 1720 items 2023-03-31T05:57:08.1018538Z Running 1720 items in this shard: test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::cat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::get_first], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::leaky_relu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__and__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__and__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__iand__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__iand__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ior__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ior__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ixor__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ixor__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__or__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__or__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__xor__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__xor__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_add_batch_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_assert_tensor_metadata], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_autocast_to_full_precision], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_autocast_to_reduced_precision], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_batch_norm_impl_index], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_batch_norm_impl_index_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Byte], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Char], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Double], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Half], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Long], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Short], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_choose_qparams_per_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution.deprecated], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution_double_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_clear_plan_cache], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_get_plan_cache_max_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_get_plan_cache_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_set_plan_cache_max_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_debug_has_internal_overlap], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_dim_arange], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_embedding_bag_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_embedding_bag_sparse_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_gather_sparse_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_grid_sampler_2d_cpu_fallback_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_has_compatible_shallow_copy_type], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_is_zerotensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_lu_with_info], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_nnpack_available], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pack_padded_sequence_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_circular], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_enum], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_packed_sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_propagate_xla_data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_remove_batch_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_reshape_from_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_rowwise_prune], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_saturate_weight_to_fp16], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_scaled_dot_product_attention], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_scaled_dot_product_attention_math], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_shape_as_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_draw], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_ff_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_initialize_state_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_scramble_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_bsc_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_bsr_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_compressed_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_coo_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_csc_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_csr_tensor_unsafe], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_log_softmax.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_log_softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_mm.reduce], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_mm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_softmax.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum.dim_dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum.dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_ambiguous_defaults.a], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_ambiguous_defaults.b], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_autograd_multiple_dispatch.ntonly], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_check_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_serialization_subcmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_string_default], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_differentiable_gru_cell_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_differentiable_lstm_cell_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_fused_lstm_cell_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_to_cpu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_unpack_dual], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_bicubic2d_aa.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_bilinear2d_aa.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact1d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact3d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_use_cudnn_rnn_flatten_weight], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_bsc_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_bsr_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_compressed_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_coo_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_csc_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_csr_tensor_args], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_version], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_weight_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_weight_norm_differentiable_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_max_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adjoint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::affine_grid_generator_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_to.ellipsis_idx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_to], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::all.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::all.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::alpha_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::alpha_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::any.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::any.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argsort.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argsort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argwhere], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_1d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_2d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_3d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::batch_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bilinear], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_and_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_or.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_or_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_xor.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_xor_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::broadcast_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::broadcast_to], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::can_cast], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cartesian_prod], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cat.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cat.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cdist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chain_matmul.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chain_matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chalf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::choose_qparams_optimized], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chunk], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::coalesce], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::column_stack.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::column_stack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::combinations], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conj_physical], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::contiguous], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv1d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv2d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv3d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_tbc_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose2d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose3d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::corrcoef], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cosine_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cosine_similarity], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cov], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross_entropy_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ctc_loss.IntList], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ctc_loss.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cudnn_is_acceptable], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummax.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummax.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummaxmin_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummin.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummin.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod_.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum_.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumulative_trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumulative_trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diag.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diagflat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diagonal.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diff.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diff], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.out_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dstack.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::einsum], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_bag.padding_idx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_bag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_sparse_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::empty.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::expand_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_channel_affine], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_channel_affine_cachemask_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine.tensor_qparams], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine_cachemask_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_fp16_weight], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_fp16_weight_fp32_activation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_int8_weight], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_int8_weight_fp32_activation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_quantize_weight], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_gemm_matrix_fp16], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_quantized_matrix.KN], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_quantized_matrix], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_alpha_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_alpha_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfftn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fill_diagonal_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.DimnameList], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.named_out_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.using_ints], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.using_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten_dense_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fliplr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flipud], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::floor_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::floor_divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::frobenius_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::frobenius_norm.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fused_moving_avg_obs_fake_quant], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ger.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ger], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::get_gradients], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalararray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarrayarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.tensorarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.tensorarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::grid_sampler], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::group_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hinge_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd.TensorList_bins], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd.int_bins], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hstack.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::imag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_add.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_copy.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_copy_.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill.Dimname_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill.Dimname_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill_.Dimname_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill_.Dimname_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::infinitely_differentiable_gelu_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inner.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inner], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::instance_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inverse.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_complex], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_distributed], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_floating_point], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_inference], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_leaf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_neg], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_nonzero], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_signed], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_vulkan_available], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isclose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isfinite], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isreal], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::istft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::item], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kl_div], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kron.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kron], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kthvalue.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kthvalue.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::l1_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::layer_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cholesky.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cholesky], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.p_str], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.p_str_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_det.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_diagonal], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigh.eigvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvals.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvalsh.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvalsh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_inv.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_inv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_ldl_factor.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_ldl_factor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_lu_factor.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_lu_factor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matmul.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.str_ord], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.str_ord_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_power.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_float_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.out_tol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.tol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_multi_dot.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_multi_dot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.ord_str], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.ord_str_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.atol_rtol_float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.atol_rtol_float_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.out_rcond_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.rcond_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_slogdet.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_slogdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve_ex.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve_ex], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svd.U], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svdvals.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svdvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorinv.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorsolve.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorsolve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vander], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vecdot.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vecdot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linear], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_sigmoid.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_sigmoid], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_softmax.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logcumsumexp.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logcumsumexp.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logsumexp.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logsumexp.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lu_solve.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lu_solve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mH], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mT], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::margin_ranking_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::masked_select_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matmul.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_H], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_exp], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_exp_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_power.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.names_dim_max], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool1d_with_indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mean.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::median.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::median.names_dim_values], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::meshgrid.indexing], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::meshgrid], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.names_dim_min], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mish_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mode.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mode.dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::moveaxis.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::moveaxis.intlist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::movedim.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::movedim.intlist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::msort.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::msort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multilabel_margin_loss.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multilabel_margin_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmean.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmedian.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmedian.names_dim_values], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::narrow.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::narrow], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::native_channel_shuffle], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nested_to_padded_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss2d.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss_nd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nonzero_numpy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_ScalarOpt_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_ScalarOpt_dim_dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_dtype_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm_except_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Tensor_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.dim_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::numpy_T], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::one_hot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::orgqr.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::orgqr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::outer.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::outer], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::output_nr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pad_sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pairwise_distance], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pdist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pin_memory], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pinverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::poisson_nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::positive], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prod.Dimname_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prod.dim_Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::promote_types], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::qr.Q], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::qr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_gru_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_lstm_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_rnn_relu_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_rnn_tanh_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rand.generator_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::randn.generator_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::randn.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ravel], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::real], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::refine_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::relu6], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::relu6_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rename], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rename_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::repeat_interleave.self_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::repeat_interleave.self_int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::requires_grad_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::reshape], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::reshape_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::resolve_conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::resolve_neg], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::retain_grad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::retains_grad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::row_stack.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::row_stack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rrelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rrelu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scaled_dot_product_attention], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter.dimname_src], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter.dimname_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter_add.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::select.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::selu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::selu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::set_.source_Tensor_storage_offset], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::set_data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::silu_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::size.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::size.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slogdet.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slogdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slow_conv3d.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slow_conv3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::smm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::softmax.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_stable], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_values], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_values_stable], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsc_tensor.ccol_row_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsc_tensor.ccol_row_value_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsr_tensor.crow_col_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsr_tensor.crow_col_value_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_compressed_tensor.comp_plain_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_compressed_tensor.comp_plain_value_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_coo_tensor.indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_coo_tensor.indices_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csc_tensor.ccol_row_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csc_tensor.ccol_row_value_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csr_tensor.crow_col_value], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csr_tensor.crow_col_value_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_digamma.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_digamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erf.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfc.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfinv.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_exp2.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_exp2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expit.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expm1.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expm1], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammainc.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammainc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaincc.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaincc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaln.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_i0.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_i0], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log1p.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log1p], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logit.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logsumexp.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logsumexp], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_multigammaln.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_multigammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_ndtr.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_ndtr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_polygamma.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_polygamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_psi.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_psi], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_round.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_round], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.x_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_sinc.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_sinc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.other_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.other_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.self_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.self_scalar_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::split.sizes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::squeeze.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::squeeze_.dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sspaddmm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.correction_names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stft.center], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stride.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stride.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum.DimnameList_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum.dim_DimnameList], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum_to_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::svd.U], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapaxes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapaxes_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapdims], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapdims_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::take_along_dim.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::take_along_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.sections], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.tensor_indices_or_sections], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensordot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::thnn_conv2d.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::thnn_conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.device], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.dtype_layout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_dense], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_dense_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_mkldnn_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trace_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::transpose.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapz.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapz.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::triplet_margin_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::type_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unbind.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten.Dimname], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten_dense_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unsafe_chunk], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_bicubic2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_bilinear2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_linear1d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest1d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest3d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_trilinear3d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::value_selecting_reduction_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vander], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.correction_names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.names_out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::view_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vstack.out], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.ScalarOther], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.ScalarSelf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[nvprims::var_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[nvprims::view.shape], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[prepacked::unpack_prepacked_sizes_conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[prepacked::unpack_prepacked_sizes_linear], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_enter], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_enter_new], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_exit._RecordFunction], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_exit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv1d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_dilation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_groups], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_output_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_stride], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_transpose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_unpack_sizes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_dilation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_groups], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_output_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_stride], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_transpose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose1d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_dilation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_groups], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_output_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_stride], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_transpose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_dilation], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_groups], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_output_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_stride], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_transpose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::embedding_bag_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::linear_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::linear_unpack_fp16], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::make_quantized_cell_params_fp16], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[sparse::qlinear_unpack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__and__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__and__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__iand__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__iand__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ior__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ior__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ixor__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ixor__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__or__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__or__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__xor__.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__xor__.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_batch_norm_impl_index], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_convolution_double_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_convolution_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_has_compatible_shallow_copy_type], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_lu_with_info], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_pad_circular], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_test_check_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::absolute], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_max_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adjoint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arccos], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arccosh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arcsin], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arcsinh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctan2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctan], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctanh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::argsort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_1d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_2d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_3d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::batch_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_and_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_or.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_or_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_xor.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_xor_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::broadcast_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::broadcast_to], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cartesian_prod], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cdist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::chunk], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::clip.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::clip], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::combinations], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::concat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::concatenate], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conj_physical], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::contiguous], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv1d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv2d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv3d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose2d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose3d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::corrcoef], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cosine_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cosine_similarity], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cov], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cross], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cross_entropy_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumprod_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumulative_trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumulative_trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diagonal_copy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diff], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::einsum], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::embedding_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::expand_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ihfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fix], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fliplr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::flipud], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Tensor_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Tensor_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::floor_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::frobenius_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gather_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ger], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalararray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarrayarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.tensorarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.tensorarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::grid_sampler], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::group_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hinge_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::index_select_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::inner], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::instance_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::inverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::is_complex], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::is_same_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::kron], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::l1_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::layer_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ldexp.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_cholesky], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_cond], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_diagonal], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_eigvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_eigvalsh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_inv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_ldl_factor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_norm.str_ord], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_rank.atol_rtol_float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_rank.atol_rtol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_multi_dot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_norm.ord_str], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_solve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_solve_ex], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_svdvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_tensorinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_vander], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_vecdot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::mH], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::mT], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matrix_H], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max_pool1d_with_indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::meshgrid.indexing], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::meshgrid], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::min.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::moveaxis.intlist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::movedim.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::msort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nanmean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::narrow], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::negative], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss_nd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::not_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::not_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nuclear_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nuclear_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::numpy_T], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::orgqr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::outer], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pairwise_distance], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pinverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::poisson_nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::positive], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::prelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::qr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ravel], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::relu6], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::relu6_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::repeat_interleave.self_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::repeat_interleave.self_int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::reshape], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::reshape_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::resolve_conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::resolve_neg], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::row_stack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::rrelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::rrelu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::scaled_dot_product_attention], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::selu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::selu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::size.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::slogdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_digamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erfc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erfinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_exp2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_expit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_expm1], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammainc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammaincc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_i0], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_log1p], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_log_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_logit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_logsumexp], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_multigammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_ndtr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_polygamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_psi], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_round], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_sinc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::split.sizes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::square], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::subtract.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::sum_to_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapaxes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapaxes_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapdims], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapdims_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::take_along_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::tensordot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::tile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapz.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapz.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::type_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::unflatten.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::unfold_copy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::value_selecting_reduction_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::view_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.ScalarOther], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.ScalarSelf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::absolute], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::absolute_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_max_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adjoint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::affine_grid_generator_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_to.ellipsis_idx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_to], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::alpha_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::alpha_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccos], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccos_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccosh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccosh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsin], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsin_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsinh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsinh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan2_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctanh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctanh_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::argsort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::argwhere], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_1d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_2d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_3d.Sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::avg_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::batch_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bilinear], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_and_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_or.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_or_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_xor.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_xor_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::broadcast_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::broadcast_to], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::can_cast], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cartesian_prod], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cat.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cdist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chain_matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chalf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::choose_qparams_optimized], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chunk], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::coalesce], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::column_stack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::combinations], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concat.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concatenate.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concatenate], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conj_physical], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::contiguous], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv1d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv2d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv3d.padding], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_tbc_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose2d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose3d.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::corrcoef], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cosine_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cosine_similarity], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cov], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cross], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cross_entropy_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ctc_loss.IntList], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ctc_loss.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cudnn_is_acceptable], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cummaxmin_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumprod_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumulative_trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumulative_trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diagflat], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diff], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.out_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Scalar_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Tensor_mode], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::einsum], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_bag.padding_idx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_bag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::expand_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_alpha_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_alpha_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_dropout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_dropout_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifftshift], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfft2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfftn], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fill_diagonal_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fix], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fix_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.named_out_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.using_ints], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.using_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten_dense_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fliplr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flipud], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Tensor_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Tensor_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::floor_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::floor_divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::frobenius_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fused_moving_avg_obs_fake_quant], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gather_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ger], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::get_gradients], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalararray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarrayarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.tensorarray], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.tensorarrayint], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::grid_sampler], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::group_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hinge_embedding_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd.TensorList_bins], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd.int_bins], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::imag], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::index_select_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::infinitely_differentiable_gelu_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::inner], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::instance_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::inverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isclose], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isfinite], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isreal], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::istft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::item], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::kl_div], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::kron], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::l1_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::layer_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ldexp.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ldexp_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cholesky], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cond.p_str], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cond], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_det], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_diagonal], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigh.eigvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigvalsh], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_inv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_ldl_factor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_lu_factor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_norm.str_ord], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.atol_rtol_float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.atol_rtol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.out_tol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.tol_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_multi_dot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_norm.ord_str], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.atol_rtol_float], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.out_rcond_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.rcond_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_slogdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_solve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_solve_ex], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svd.U], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svdvals], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_tensorinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_tensorsolve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_vander], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_vecdot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linear], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::log_sigmoid], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::log_softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::logdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::logsumexp.names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lu_solve], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mH], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mT], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::margin_ranking_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::masked_select_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matmul], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_H], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_exp], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_exp_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_power], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.names_dim_max], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool1d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool1d_with_indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::median.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::median.names_dim_values], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::meshgrid.indexing], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::meshgrid], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.names_dim_min], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mish_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::moveaxis.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::moveaxis.intlist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::movedim.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::movedim.intlist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::msort], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multilabel_margin_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmedian.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmedian.names_dim_values], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanquantile.scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanquantile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::narrow.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::narrow], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::native_channel_shuffle], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::negative], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::negative_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nested_to_padded_tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss_nd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nonzero_numpy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm.names_ScalarOpt_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm.names_ScalarOpt_dim_dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm_except_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nuclear_norm.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nuclear_norm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::numpy_T], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::one_hot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::orgqr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::outer], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::output_nr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pad_sequence], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pairwise_distance], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pdist], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pin_memory], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pinverse], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::poisson_nll_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::positive], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::prelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::promote_types], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::qr.Q], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::qr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::quantile.scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::quantile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ravel], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::real], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::refine_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::relu6], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::relu6_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rename], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rename_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::repeat_interleave.self_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::repeat_interleave.self_int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::requires_grad_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::reshape], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::reshape_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::resolve_conj], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::resolve_neg], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar_Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar_Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::retain_grad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::retains_grad], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh.data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh.input], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh_cell], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::row_stack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rrelu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rrelu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::scaled_dot_product_attention], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::selu], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::selu_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::set_.source_Tensor_storage_offset], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::set_data], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::silu_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::size.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::slogdet], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::slow_conv3d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::smm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::softmax.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_t.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_t.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_u.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_u.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_v.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_v.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_w.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_w.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_digamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erfc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erfinv], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_exp2], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_expit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_expm1], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammainc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammaincc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_h.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_h.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_he.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_he.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_i0], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_laguerre_polynomial_l.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_laguerre_polynomial_l.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_legendre_polynomial_p.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_legendre_polynomial_p.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_log1p], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_log_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_logit], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_logsumexp], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_multigammaln], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_ndtr], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_polygamma], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_psi], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_round], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_t.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_t.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_u.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_u.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_v.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_v.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_w.n_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_w.x_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_sinc], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_softmax], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy.other_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy.self_scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::split.sizes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::square], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::square_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::sspaddmm], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stft.center], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stft], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stride.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::sum_to_size], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::svd.U], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::svd], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapaxes], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapaxes_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapdims], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapdims_], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::take_along_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.indices], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.sections], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.tensor_indices_or_sections], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensordot], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::thnn_conv2d], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tile], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.device], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.dtype], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.dtype_layout], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.other], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_dense], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_dense_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_mkldnn_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trace_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapezoid.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapezoid.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapz.dx], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapz.x], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::triplet_margin_loss], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide_.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide_.Tensor], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::type_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unflatten.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unflatten_dense_tensors], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unsafe_chunk], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_bicubic2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_bilinear2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_linear1d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest1d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest2d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest3d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_trilinear3d.vec], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::value_selecting_reduction_backward], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vander], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.correction_names], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.names_dim], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::view_as], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vsplit.array], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vsplit.int], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vstack], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.ScalarOther], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.ScalarSelf], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.Scalar], test/functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where] 2023-03-31T05:57:08.1311025Z 2023-03-31T05:57:08.1311654Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::cat] PASSED [ 0%] 2023-03-31T05:57:08.1313319Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::get_first] PASSED [ 0%] 2023-03-31T05:57:08.1313849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[_test::leaky_relu] PASSED [ 0%] 2023-03-31T05:57:08.1314380Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__and__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1314902Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__and__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1315436Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__iand__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1315979Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__iand__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1316504Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ior__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1317022Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ior__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1317686Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ixor__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1318264Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__ixor__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1318793Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__or__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1319318Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__or__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1319824Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__xor__.Scalar] PASSED [ 0%] 2023-03-31T05:57:08.1320343Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::__xor__.Tensor] PASSED [ 0%] 2023-03-31T05:57:08.1320869Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_add_batch_dim] PASSED [ 0%] 2023-03-31T05:57:08.1321401Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_assert_tensor_metadata] PASSED [ 0%] 2023-03-31T05:57:08.1322037Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_autocast_to_full_precision] PASSED [ 1%] 2023-03-31T05:57:08.1322579Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_autocast_to_reduced_precision] PASSED [ 1%] 2023-03-31T05:57:08.1323229Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_backward] PASSED [ 1%] 2023-03-31T05:57:08.1323767Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_batch_norm_impl_index] PASSED [ 1%] 2023-03-31T05:57:08.1324310Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_batch_norm_impl_index_backward] PASSED [ 1%] 2023-03-31T05:57:08.1324883Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Byte] PASSED [ 1%] 2023-03-31T05:57:08.1325395Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Char] PASSED [ 1%] 2023-03-31T05:57:08.1325913Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Double] PASSED [ 1%] 2023-03-31T05:57:08.1326437Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Float] PASSED [ 1%] 2023-03-31T05:57:08.1326955Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Half] PASSED [ 1%] 2023-03-31T05:57:08.1327459Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Int] PASSED [ 1%] 2023-03-31T05:57:08.1327971Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Long] PASSED [ 1%] 2023-03-31T05:57:08.1328485Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cast_Short] PASSED [ 1%] 2023-03-31T05:57:08.1329054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_choose_qparams_per_tensor] PASSED [ 1%] 2023-03-31T05:57:08.1329626Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution.deprecated] PASSED [ 1%] 2023-03-31T05:57:08.1330185Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution_double_backward] PASSED [ 1%] 2023-03-31T05:57:08.1330733Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_convolution_mode] PASSED [ 1%] 2023-03-31T05:57:08.1331262Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_clear_plan_cache] PASSED [ 2%] 2023-03-31T05:57:08.1331806Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_get_plan_cache_max_size] PASSED [ 2%] 2023-03-31T05:57:08.1332337Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_get_plan_cache_size] PASSED [ 2%] 2023-03-31T05:57:08.1332872Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_cufft_set_plan_cache_max_size] PASSED [ 2%] 2023-03-31T05:57:08.1333451Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_debug_has_internal_overlap] PASSED [ 2%] 2023-03-31T05:57:08.1333986Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_dim_arange] PASSED [ 2%] 2023-03-31T05:57:08.1334505Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_embedding_bag_backward] PASSED [ 2%] 2023-03-31T05:57:08.1335054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_embedding_bag_sparse_backward] PASSED [ 2%] 2023-03-31T05:57:08.1335598Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_gather_sparse_backward] PASSED [ 2%] 2023-03-31T05:57:08.1336187Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_grid_sampler_2d_cpu_fallback_backward] PASSED [ 2%] 2023-03-31T05:57:08.1336757Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_has_compatible_shallow_copy_type] PASSED [ 2%] 2023-03-31T05:57:08.1337289Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_is_zerotensor] PASSED [ 2%] 2023-03-31T05:57:08.1337813Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_lu_with_info] PASSED [ 2%] 2023-03-31T05:57:08.1338339Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_nnpack_available] PASSED [ 2%] 2023-03-31T05:57:08.1338885Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pack_padded_sequence_backward] PASSED [ 2%] 2023-03-31T05:57:08.1339414Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_circular] PASSED [ 2%] 2023-03-31T05:57:08.1339931Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_enum] PASSED [ 2%] 2023-03-31T05:57:08.1340488Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_pad_packed_sequence] PASSED [ 3%] 2023-03-31T05:57:08.1341051Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_propagate_xla_data] PASSED [ 3%] 2023-03-31T05:57:08.1341585Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_remove_batch_dim] PASSED [ 3%] 2023-03-31T05:57:08.1342100Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_reshape_from_tensor] PASSED [ 3%] 2023-03-31T05:57:08.1342627Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_rowwise_prune] PASSED [ 3%] 2023-03-31T05:57:08.1343255Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_saturate_weight_to_fp16] PASSED [ 3%] 2023-03-31T05:57:08.1343810Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_scaled_dot_product_attention] PASSED [ 3%] 2023-03-31T05:57:08.1344400Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_scaled_dot_product_attention_math] PASSED [ 3%] 2023-03-31T05:57:08.1344947Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_shape_as_tensor] PASSED [ 3%] 2023-03-31T05:57:08.1345475Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_draw] PASSED [ 3%] 2023-03-31T05:57:08.1346011Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_ff_] PASSED [ 3%] 2023-03-31T05:57:08.1346555Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_initialize_state_] PASSED [ 3%] 2023-03-31T05:57:08.1347130Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sobol_engine_scramble_] PASSED [ 3%] 2023-03-31T05:57:08.1347672Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_bsc_tensor_unsafe] PASSED [ 3%] 2023-03-31T05:57:08.1348216Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_bsr_tensor_unsafe] PASSED [ 3%] 2023-03-31T05:57:08.1348767Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_compressed_tensor_unsafe] PASSED [ 3%] 2023-03-31T05:57:08.1349305Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_coo_tensor_unsafe] PASSED [ 3%] 2023-03-31T05:57:08.1349849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_csc_tensor_unsafe] PASSED [ 4%] 2023-03-31T05:57:08.1350394Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_csr_tensor_unsafe] PASSED [ 4%] 2023-03-31T05:57:08.1350940Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_log_softmax.Dimname] PASSED [ 4%] 2023-03-31T05:57:08.1351522Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_log_softmax.int] PASSED [ 4%] 2023-03-31T05:57:08.1352072Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_mm.reduce] PASSED [ 4%] 2023-03-31T05:57:08.1352592Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_mm] PASSED [ 4%] 2023-03-31T05:57:08.1353127Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_softmax.Dimname] PASSED [ 4%] 2023-03-31T05:57:08.1353666Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_softmax.int] PASSED [ 4%] 2023-03-31T05:57:08.1354195Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum.dim_dtype] PASSED [ 4%] 2023-03-31T05:57:08.1354730Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum.dtype] PASSED [ 4%] 2023-03-31T05:57:08.1355249Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_sparse_sum] PASSED [ 4%] 2023-03-31T05:57:08.1355813Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_ambiguous_defaults.a] PASSED [ 4%] 2023-03-31T05:57:08.1356361Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_ambiguous_defaults.b] PASSED [ 4%] 2023-03-31T05:57:08.1356916Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_autograd_multiple_dispatch.ntonly] PASSED [ 4%] 2023-03-31T05:57:08.1357472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_check_tensor] PASSED [ 4%] 2023-03-31T05:57:08.1358012Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_serialization_subcmul] PASSED [ 4%] 2023-03-31T05:57:08.1358594Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_test_string_default] PASSED [ 4%] 2023-03-31T05:57:08.1359137Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_differentiable_gru_cell_backward] PASSED [ 5%] 2023-03-31T05:57:08.1359719Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_differentiable_lstm_cell_backward] PASSED [ 5%] 2023-03-31T05:57:08.1360285Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_thnn_fused_lstm_cell_backward] PASSED [ 5%] 2023-03-31T05:57:08.1360808Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_to_cpu] PASSED [ 5%] 2023-03-31T05:57:08.1361329Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_unpack_dual] PASSED [ 5%] 2023-03-31T05:57:08.1361848Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_bicubic2d_aa.vec] PASSED [ 5%] 2023-03-31T05:57:08.1362399Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_bilinear2d_aa.vec] PASSED [ 5%] 2023-03-31T05:57:08.1363000Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact1d.vec] PASSED [ 5%] 2023-03-31T05:57:08.1363689Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact2d.vec] PASSED [ 5%] 2023-03-31T05:57:08.1364256Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_upsample_nearest_exact3d.vec] PASSED [ 5%] 2023-03-31T05:57:08.1364794Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_use_cudnn_rnn_flatten_weight] PASSED [ 5%] 2023-03-31T05:57:08.1365346Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_bsc_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1365903Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_bsr_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1366480Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_compressed_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1367091Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_coo_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1367640Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_csc_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1368195Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_validate_sparse_csr_tensor_args] PASSED [ 5%] 2023-03-31T05:57:08.1368730Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_version] PASSED [ 5%] 2023-03-31T05:57:08.1369243Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_weight_norm] PASSED [ 6%] 2023-03-31T05:57:08.1369829Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::_weight_norm_differentiable_backward] PASSED [ 6%] 2023-03-31T05:57:08.1370377Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute.out] PASSED [ 6%] 2023-03-31T05:57:08.1370893Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute] PASSED [ 6%] 2023-03-31T05:57:08.1371404Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::absolute_] PASSED [ 6%] 2023-03-31T05:57:08.1371921Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool1d] PASSED [ 6%] 2023-03-31T05:57:08.1372457Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool2d] PASSED [ 6%] 2023-03-31T05:57:08.1372990Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_avg_pool3d] PASSED [ 6%] 2023-03-31T05:57:08.1373518Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adaptive_max_pool1d] PASSED [ 6%] 2023-03-31T05:57:08.1374082Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::adjoint] PASSED [ 6%] 2023-03-31T05:57:08.1374638Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::affine_grid_generator_backward] PASSED [ 6%] 2023-03-31T05:57:08.1375174Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_as] PASSED [ 6%] 2023-03-31T05:57:08.1375693Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_tensors] PASSED [ 6%] 2023-03-31T05:57:08.1376225Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_to.ellipsis_idx] PASSED [ 6%] 2023-03-31T05:57:08.1376744Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::align_to] PASSED [ 6%] 2023-03-31T05:57:08.1377261Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::all.dimname] PASSED [ 6%] 2023-03-31T05:57:08.1377783Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::all.dimname_out] PASSED [ 6%] 2023-03-31T05:57:08.1378349Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::alpha_dropout] PASSED [ 7%] 2023-03-31T05:57:08.1378879Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::alpha_dropout_] PASSED [ 7%] 2023-03-31T05:57:08.1379386Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::any.dimname] PASSED [ 7%] 2023-03-31T05:57:08.1379909Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::any.dimname_out] PASSED [ 7%] 2023-03-31T05:57:08.1380433Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos.out] PASSED [ 7%] 2023-03-31T05:57:08.1380979Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos] PASSED [ 7%] 2023-03-31T05:57:08.1381469Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccos_] PASSED [ 7%] 2023-03-31T05:57:08.1381977Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh.out] PASSED [ 7%] 2023-03-31T05:57:08.1382493Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh] PASSED [ 7%] 2023-03-31T05:57:08.1383112Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arccosh_] PASSED [ 7%] 2023-03-31T05:57:08.1383619Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin.out] PASSED [ 7%] 2023-03-31T05:57:08.1384133Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin] PASSED [ 7%] 2023-03-31T05:57:08.1384643Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsin_] PASSED [ 7%] 2023-03-31T05:57:08.1385158Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh.out] PASSED [ 7%] 2023-03-31T05:57:08.1385719Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh] PASSED [ 7%] 2023-03-31T05:57:08.1386248Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arcsinh_] PASSED [ 7%] 2023-03-31T05:57:08.1386769Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan.out] PASSED [ 7%] 2023-03-31T05:57:08.1387286Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2.out] PASSED [ 8%] 2023-03-31T05:57:08.1387799Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2] PASSED [ 8%] 2023-03-31T05:57:08.1388299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan2_] PASSED [ 8%] 2023-03-31T05:57:08.1388808Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan] PASSED [ 8%] 2023-03-31T05:57:08.1389316Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctan_] PASSED [ 8%] 2023-03-31T05:57:08.1389870Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh.out] PASSED [ 8%] 2023-03-31T05:57:08.1390376Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh] PASSED [ 8%] 2023-03-31T05:57:08.1390887Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::arctanh_] PASSED [ 8%] 2023-03-31T05:57:08.1391415Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argsort.dimname] PASSED [ 8%] 2023-03-31T05:57:08.1391939Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argsort] PASSED [ 8%] 2023-03-31T05:57:08.1392481Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::argwhere] PASSED [ 8%] 2023-03-31T05:57:08.1392999Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_1d.Sequence] PASSED [ 8%] 2023-03-31T05:57:08.1393528Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_1d] PASSED [ 8%] 2023-03-31T05:57:08.1394054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_2d.Sequence] PASSED [ 8%] 2023-03-31T05:57:08.1394584Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_2d] PASSED [ 8%] 2023-03-31T05:57:08.1395097Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_3d.Sequence] PASSED [ 8%] 2023-03-31T05:57:08.1395630Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::atleast_3d] PASSED [ 8%] 2023-03-31T05:57:08.1396144Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::avg_pool1d] PASSED [ 9%] 2023-03-31T05:57:08.1396656Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::batch_norm] PASSED [ 9%] 2023-03-31T05:57:08.1397198Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bilinear] PASSED [ 9%] 2023-03-31T05:57:08.1397734Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_and_.Scalar] PASSED [ 9%] 2023-03-31T05:57:08.1398273Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_or.Scalar] PASSED [ 9%] 2023-03-31T05:57:08.1398807Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_or_.Scalar] PASSED [ 9%] 2023-03-31T05:57:08.1399331Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_xor.Scalar] PASSED [ 9%] 2023-03-31T05:57:08.1399842Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::bitwise_xor_.Scalar] PASSED [ 9%] 2023-03-31T05:57:08.1400371Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::broadcast_tensors] PASSED [ 9%] 2023-03-31T05:57:08.1400892Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::broadcast_to] PASSED [ 9%] 2023-03-31T05:57:08.1401452Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::can_cast] PASSED [ 9%] 2023-03-31T05:57:08.1401956Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cartesian_prod] PASSED [ 9%] 2023-03-31T05:57:08.1402477Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cat.names] PASSED [ 9%] 2023-03-31T05:57:08.1402999Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cat.names_out] PASSED [ 9%] 2023-03-31T05:57:08.1403643Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cdist] PASSED [ 9%] 2023-03-31T05:57:08.1404262Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chain_matmul.out] PASSED [ 9%] 2023-03-31T05:57:08.1404769Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chain_matmul] PASSED [ 9%] 2023-03-31T05:57:08.1405279Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chalf] PASSED [ 10%] 2023-03-31T05:57:08.1405811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::choose_qparams_optimized] PASSED [ 10%] 2023-03-31T05:57:08.1406346Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::chunk] PASSED [ 10%] 2023-03-31T05:57:08.1406854Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.Tensor] PASSED [ 10%] 2023-03-31T05:57:08.1407376Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.Tensor_out] PASSED [ 10%] 2023-03-31T05:57:08.1407897Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip.out] PASSED [ 10%] 2023-03-31T05:57:08.1408442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip] PASSED [ 10%] 2023-03-31T05:57:08.1408973Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip_.Tensor] PASSED [ 10%] 2023-03-31T05:57:08.1409484Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::clip_] PASSED [ 10%] 2023-03-31T05:57:08.1409992Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::coalesce] PASSED [ 10%] 2023-03-31T05:57:08.1410510Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::column_stack.out] PASSED [ 10%] 2023-03-31T05:57:08.1411026Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::column_stack] PASSED [ 10%] 2023-03-31T05:57:08.1411533Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::combinations] PASSED [ 10%] 2023-03-31T05:57:08.1412052Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.names] PASSED [ 10%] 2023-03-31T05:57:08.1412615Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.names_out] PASSED [ 10%] 2023-03-31T05:57:08.1413137Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat.out] PASSED [ 10%] 2023-03-31T05:57:08.1413631Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concat] PASSED [ 10%] 2023-03-31T05:57:08.1414154Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.names] PASSED [ 10%] 2023-03-31T05:57:08.1414695Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.names_out] PASSED [ 11%] 2023-03-31T05:57:08.1415260Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate.out] PASSED [ 11%] 2023-03-31T05:57:08.1415790Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::concatenate] PASSED [ 11%] 2023-03-31T05:57:08.1416288Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conj] PASSED [ 11%] 2023-03-31T05:57:08.1416800Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conj_physical] PASSED [ 11%] 2023-03-31T05:57:08.1417321Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::contiguous] PASSED [ 11%] 2023-03-31T05:57:08.1417842Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv1d.padding] PASSED [ 11%] 2023-03-31T05:57:08.1418347Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv1d] PASSED [ 11%] 2023-03-31T05:57:08.1418859Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv2d.padding] PASSED [ 11%] 2023-03-31T05:57:08.1419371Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv2d] PASSED [ 11%] 2023-03-31T05:57:08.1419914Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv3d.padding] PASSED [ 11%] 2023-03-31T05:57:08.1420439Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv3d] PASSED [ 11%] 2023-03-31T05:57:08.1420957Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_tbc_backward] PASSED [ 11%] 2023-03-31T05:57:08.1421484Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose1d] PASSED [ 11%] 2023-03-31T05:57:08.1422022Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose2d.input] PASSED [ 11%] 2023-03-31T05:57:08.1422561Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::conv_transpose3d.input] PASSED [ 11%] 2023-03-31T05:57:08.1423158Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::corrcoef] PASSED [ 11%] 2023-03-31T05:57:08.1423690Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cosine_embedding_loss] PASSED [ 12%] 2023-03-31T05:57:08.1424265Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cosine_similarity] PASSED [ 12%] 2023-03-31T05:57:08.1424779Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cov] PASSED [ 12%] 2023-03-31T05:57:08.1425273Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross.out] PASSED [ 12%] 2023-03-31T05:57:08.1425783Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross] PASSED [ 12%] 2023-03-31T05:57:08.1426303Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cross_entropy_loss] PASSED [ 12%] 2023-03-31T05:57:08.1426865Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ctc_loss.IntList] PASSED [ 12%] 2023-03-31T05:57:08.1427393Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ctc_loss.Tensor] PASSED [ 12%] 2023-03-31T05:57:08.1427907Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cudnn_is_acceptable] PASSED [ 12%] 2023-03-31T05:57:08.1428436Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummax.dimname] PASSED [ 12%] 2023-03-31T05:57:08.1428968Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummax.dimname_out] PASSED [ 12%] 2023-03-31T05:57:08.1429499Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummaxmin_backward] PASSED [ 12%] 2023-03-31T05:57:08.1430007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummin.dimname] PASSED [ 12%] 2023-03-31T05:57:08.1430532Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cummin.dimname_out] PASSED [ 12%] 2023-03-31T05:57:08.1431086Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod.dimname] PASSED [ 12%] 2023-03-31T05:57:08.1431648Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod.dimname_out] PASSED [ 12%] 2023-03-31T05:57:08.1432174Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod_.dimname] PASSED [ 12%] 2023-03-31T05:57:08.1432692Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumprod_backward] PASSED [ 13%] 2023-03-31T05:57:08.1433213Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum.dimname] PASSED [ 13%] 2023-03-31T05:57:08.1433736Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum.dimname_out] PASSED [ 13%] 2023-03-31T05:57:08.1434268Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumsum_.dimname] PASSED [ 13%] 2023-03-31T05:57:08.1434791Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumulative_trapezoid.dx] PASSED [ 13%] 2023-03-31T05:57:08.1435366Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::cumulative_trapezoid.x] PASSED [ 13%] 2023-03-31T05:57:08.1435883Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::data] PASSED [ 13%] 2023-03-31T05:57:08.1436384Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::det] PASSED [ 13%] 2023-03-31T05:57:08.1436882Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diag.out] PASSED [ 13%] 2023-03-31T05:57:08.1437390Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diag] PASSED [ 13%] 2023-03-31T05:57:08.1437925Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diagflat] PASSED [ 13%] 2023-03-31T05:57:08.1438451Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diagonal.Dimname] PASSED [ 13%] 2023-03-31T05:57:08.1438972Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diff.out] PASSED [ 13%] 2023-03-31T05:57:08.1439464Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::diff] PASSED [ 13%] 2023-03-31T05:57:08.1439975Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Scalar] PASSED [ 13%] 2023-03-31T05:57:08.1440506Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Scalar_mode] PASSED [ 13%] 2023-03-31T05:57:08.1441037Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Tensor] PASSED [ 13%] 2023-03-31T05:57:08.1441549Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.Tensor_mode] PASSED [ 14%] 2023-03-31T05:57:08.1442070Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.out] PASSED [ 14%] 2023-03-31T05:57:08.1442632Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide.out_mode] PASSED [ 14%] 2023-03-31T05:57:08.1443389Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Scalar] PASSED [ 14%] 2023-03-31T05:57:08.1443915Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Scalar_mode] PASSED [ 14%] 2023-03-31T05:57:08.1444445Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Tensor] PASSED [ 14%] 2023-03-31T05:57:08.1444974Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::divide_.Tensor_mode] PASSED [ 14%] 2023-03-31T05:57:08.1445497Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dropout] PASSED [ 14%] 2023-03-31T05:57:08.1446017Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dropout_] PASSED [ 14%] 2023-03-31T05:57:08.1446526Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dsplit.array] PASSED [ 14%] 2023-03-31T05:57:08.1447109Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dsplit.int] PASSED [ 14%] 2023-03-31T05:57:08.1447626Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dstack.out] PASSED [ 14%] 2023-03-31T05:57:08.1448144Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::dstack] PASSED [ 14%] 2023-03-31T05:57:08.1448643Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::einsum] PASSED [ 14%] 2023-03-31T05:57:08.1449170Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_backward] PASSED [ 14%] 2023-03-31T05:57:08.1449763Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_bag.padding_idx] PASSED [ 14%] 2023-03-31T05:57:08.1450299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_bag] PASSED [ 14%] 2023-03-31T05:57:08.1450842Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::embedding_sparse_backward] PASSED [ 15%] 2023-03-31T05:57:08.1451364Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::empty.out] PASSED [ 15%] 2023-03-31T05:57:08.1451877Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::expand_as] PASSED [ 15%] 2023-03-31T05:57:08.1452409Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_channel_affine] PASSED [ 15%] 2023-03-31T05:57:08.1452989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_channel_affine_cachemask_backward] PASSED [ 15%] 2023-03-31T05:57:08.1453570Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine.tensor_qparams] PASSED [ 15%] 2023-03-31T05:57:08.1454180Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine] PASSED [ 15%] 2023-03-31T05:57:08.1454799Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fake_quantize_per_tensor_affine_cachemask_backward] PASSED [ 15%] 2023-03-31T05:57:08.1455380Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_fp16_weight] PASSED [ 15%] 2023-03-31T05:57:08.1455939Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_fp16_weight_fp32_activation] PASSED [ 15%] 2023-03-31T05:57:08.1456496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_int8_weight] PASSED [ 15%] 2023-03-31T05:57:08.1457056Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_int8_weight_fp32_activation] PASSED [ 15%] 2023-03-31T05:57:08.1457624Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_linear_quantize_weight] PASSED [ 15%] 2023-03-31T05:57:08.1458211Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_gemm_matrix_fp16] PASSED [ 15%] 2023-03-31T05:57:08.1458767Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_quantized_matrix.KN] PASSED [ 15%] 2023-03-31T05:57:08.1459306Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fbgemm_pack_quantized_matrix] PASSED [ 15%] 2023-03-31T05:57:08.1459851Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_alpha_dropout] PASSED [ 15%] 2023-03-31T05:57:08.1460397Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_alpha_dropout_] PASSED [ 15%] 2023-03-31T05:57:08.1460963Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_dropout] PASSED [ 16%] 2023-03-31T05:57:08.1461480Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::feature_dropout_] PASSED [ 16%] 2023-03-31T05:57:08.1462002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft.out] PASSED [ 16%] 2023-03-31T05:57:08.1462526Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft2.out] PASSED [ 16%] 2023-03-31T05:57:08.1463117Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft2] PASSED [ 16%] 2023-03-31T05:57:08.1463634Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fft] PASSED [ 16%] 2023-03-31T05:57:08.1464136Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftn.out] PASSED [ 16%] 2023-03-31T05:57:08.1464650Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftn] PASSED [ 16%] 2023-03-31T05:57:08.1465198Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_fftshift] PASSED [ 16%] 2023-03-31T05:57:08.1465750Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft.out] PASSED [ 16%] 2023-03-31T05:57:08.1466261Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft2.out] PASSED [ 16%] 2023-03-31T05:57:08.1466783Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft2] PASSED [ 16%] 2023-03-31T05:57:08.1467301Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfft] PASSED [ 16%] 2023-03-31T05:57:08.1467820Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfftn.out] PASSED [ 16%] 2023-03-31T05:57:08.1468331Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_hfftn] PASSED [ 16%] 2023-03-31T05:57:08.1468853Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft.out] PASSED [ 16%] 2023-03-31T05:57:08.1469377Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft2.out] PASSED [ 16%] 2023-03-31T05:57:08.1469936Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft2] PASSED [ 17%] 2023-03-31T05:57:08.1470450Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifft] PASSED [ 17%] 2023-03-31T05:57:08.1470956Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftn.out] PASSED [ 17%] 2023-03-31T05:57:08.1471476Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftn] PASSED [ 17%] 2023-03-31T05:57:08.1471997Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ifftshift] PASSED [ 17%] 2023-03-31T05:57:08.1472559Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft.out] PASSED [ 17%] 2023-03-31T05:57:08.1473068Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft2.out] PASSED [ 17%] 2023-03-31T05:57:08.1473587Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft2] PASSED [ 17%] 2023-03-31T05:57:08.1474105Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfft] PASSED [ 17%] 2023-03-31T05:57:08.1474632Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfftn.out] PASSED [ 17%] 2023-03-31T05:57:08.1475147Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_ihfftn] PASSED [ 17%] 2023-03-31T05:57:08.1475667Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft.out] PASSED [ 17%] 2023-03-31T05:57:08.1476186Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft2.out] PASSED [ 17%] 2023-03-31T05:57:08.1476735Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft2] PASSED [ 17%] 2023-03-31T05:57:08.1477276Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfft] PASSED [ 17%] 2023-03-31T05:57:08.1477780Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfftn.out] PASSED [ 17%] 2023-03-31T05:57:08.1478301Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_irfftn] PASSED [ 17%] 2023-03-31T05:57:08.1478813Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft.out] PASSED [ 18%] 2023-03-31T05:57:08.1479334Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft2.out] PASSED [ 18%] 2023-03-31T05:57:08.1479840Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft2] PASSED [ 18%] 2023-03-31T05:57:08.1480350Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfft] PASSED [ 18%] 2023-03-31T05:57:08.1480917Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfftn.out] PASSED [ 18%] 2023-03-31T05:57:08.1481432Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fft_rfftn] PASSED [ 18%] 2023-03-31T05:57:08.1481944Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fill_diagonal_] PASSED [ 18%] 2023-03-31T05:57:08.1482446Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix.out] PASSED [ 18%] 2023-03-31T05:57:08.1482954Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix] PASSED [ 18%] 2023-03-31T05:57:08.1483578Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fix_] PASSED [ 18%] 2023-03-31T05:57:08.1484161Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.DimnameList] PASSED [ 18%] 2023-03-31T05:57:08.1484692Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.named_out_dim] PASSED [ 18%] 2023-03-31T05:57:08.1485229Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.using_ints] XFAIL [ 18%] 2023-03-31T05:57:08.1485762Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten.using_names] PASSED [ 18%] 2023-03-31T05:57:08.1486299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flatten_dense_tensors] PASSED [ 18%] 2023-03-31T05:57:08.1486811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fliplr] PASSED [ 18%] 2023-03-31T05:57:08.1487315Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::flipud] PASSED [ 18%] 2023-03-31T05:57:08.1487831Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Scalar] PASSED [ 19%] 2023-03-31T05:57:08.1488407Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Scalar_out] PASSED [ 19%] 2023-03-31T05:57:08.1488989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Scalar] PASSED [ 19%] 2023-03-31T05:57:08.1489534Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Scalar_out] PASSED [ 19%] 2023-03-31T05:57:08.1490084Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Tensor] PASSED [ 19%] 2023-03-31T05:57:08.1490633Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power.Tensor_Tensor_out] PASSED [ 19%] 2023-03-31T05:57:08.1491178Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power_.Scalar] PASSED [ 19%] 2023-03-31T05:57:08.1491695Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::float_power_.Tensor] PASSED [ 19%] 2023-03-31T05:57:08.1492303Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::floor_divide.Scalar] PASSED [ 19%] 2023-03-31T05:57:08.1492839Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::floor_divide_.Scalar] PASSED [ 19%] 2023-03-31T05:57:08.1493373Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::frobenius_norm.dim] PASSED [ 19%] 2023-03-31T05:57:08.1493909Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::frobenius_norm.out] PASSED [ 19%] 2023-03-31T05:57:08.1494437Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::fused_moving_avg_obs_fake_quant] PASSED [ 19%] 2023-03-31T05:57:08.1495018Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather.dimname] PASSED [ 19%] 2023-03-31T05:57:08.1495552Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather.dimname_out] PASSED [ 19%] 2023-03-31T05:57:08.1496077Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gather_backward] PASSED [ 19%] 2023-03-31T05:57:08.1496586Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ger.out] PASSED [ 19%] 2023-03-31T05:57:08.1497090Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ger] PASSED [ 20%] 2023-03-31T05:57:08.1497605Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::get_gradients] PASSED [ 20%] 2023-03-31T05:57:08.1498133Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.array] PASSED [ 20%] 2023-03-31T05:57:08.1498673Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalararray] PASSED [ 20%] 2023-03-31T05:57:08.1499204Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarint] PASSED [ 20%] 2023-03-31T05:57:08.1499796Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarrayarray] PASSED [ 20%] 2023-03-31T05:57:08.1500379Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.scalarrayint] PASSED [ 20%] 2023-03-31T05:57:08.1500936Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.tensorarray] PASSED [ 20%] 2023-03-31T05:57:08.1501482Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gradient.tensorarrayint] PASSED [ 20%] 2023-03-31T05:57:08.1502023Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Scalar] PASSED [ 20%] 2023-03-31T05:57:08.1502550Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Scalar_out] PASSED [ 20%] 2023-03-31T05:57:08.1503156Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Tensor] PASSED [ 20%] 2023-03-31T05:57:08.1503695Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater.Tensor_out] PASSED [ 20%] 2023-03-31T05:57:08.1504256Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_.Scalar] PASSED [ 20%] 2023-03-31T05:57:08.1504781Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_.Tensor] PASSED [ 20%] 2023-03-31T05:57:08.1505319Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Scalar] PASSED [ 20%] 2023-03-31T05:57:08.1505860Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Scalar_out] PASSED [ 20%] 2023-03-31T05:57:08.1506390Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Tensor] PASSED [ 20%] 2023-03-31T05:57:08.1506971Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal.Tensor_out] PASSED [ 21%] 2023-03-31T05:57:08.1507508Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal_.Scalar] PASSED [ 21%] 2023-03-31T05:57:08.1508043Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::greater_equal_.Tensor] PASSED [ 21%] 2023-03-31T05:57:08.1508572Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::grid_sampler] PASSED [ 21%] 2023-03-31T05:57:08.1509075Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::group_norm] PASSED [ 21%] 2023-03-31T05:57:08.1509595Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru.data] PASSED [ 21%] 2023-03-31T05:57:08.1510111Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru.input] PASSED [ 21%] 2023-03-31T05:57:08.1510617Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::gru_cell] PASSED [ 21%] 2023-03-31T05:57:08.1511168Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hinge_embedding_loss] PASSED [ 21%] 2023-03-31T05:57:08.1511749Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd.TensorList_bins] PASSED [ 21%] 2023-03-31T05:57:08.1512302Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd.int_bins] PASSED [ 21%] 2023-03-31T05:57:08.1512833Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::histogramdd] PASSED [ 21%] 2023-03-31T05:57:08.1513354Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hsplit.array] PASSED [ 21%] 2023-03-31T05:57:08.1513864Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hsplit.int] PASSED [ 21%] 2023-03-31T05:57:08.1514378Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hstack.out] PASSED [ 21%] 2023-03-31T05:57:08.1514893Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::hstack] PASSED [ 21%] 2023-03-31T05:57:08.1515429Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::imag] XFAIL [ 21%] 2023-03-31T05:57:08.1515935Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_add.dimname] PASSED [ 22%] 2023-03-31T05:57:08.1516469Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_copy.dimname] PASSED [ 22%] 2023-03-31T05:57:08.1517001Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_copy_.dimname] PASSED [ 22%] 2023-03-31T05:57:08.1517541Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill.Dimname_Scalar] PASSED [ 22%] 2023-03-31T05:57:08.1518106Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill.Dimname_Tensor] PASSED [ 22%] 2023-03-31T05:57:08.1518645Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill_.Dimname_Scalar] PASSED [ 22%] 2023-03-31T05:57:08.1519189Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_fill_.Dimname_Tensor] PASSED [ 22%] 2023-03-31T05:57:08.1519736Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select.dimname] PASSED [ 22%] 2023-03-31T05:57:08.1520275Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select.dimname_out] PASSED [ 22%] 2023-03-31T05:57:08.1520806Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::index_select_backward] PASSED [ 22%] 2023-03-31T05:57:08.1521369Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::infinitely_differentiable_gelu_backward] PASSED [ 22%] 2023-03-31T05:57:08.1521915Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inner.out] PASSED [ 22%] 2023-03-31T05:57:08.1522470Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inner] PASSED [ 22%] 2023-03-31T05:57:08.1523000Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::instance_norm] PASSED [ 22%] 2023-03-31T05:57:08.1523661Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inverse.out] PASSED [ 22%] 2023-03-31T05:57:08.1524179Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::inverse] PASSED [ 22%] 2023-03-31T05:57:08.1524690Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_complex] PASSED [ 22%] 2023-03-31T05:57:08.1525207Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_conj] PASSED [ 23%] 2023-03-31T05:57:08.1525712Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_distributed] PASSED [ 23%] 2023-03-31T05:57:08.1526243Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_floating_point] PASSED [ 23%] 2023-03-31T05:57:08.1526827Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_inference] PASSED [ 23%] 2023-03-31T05:57:08.1527342Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_leaf] PASSED [ 23%] 2023-03-31T05:57:08.1527838Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_neg] PASSED [ 23%] 2023-03-31T05:57:08.1528342Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_nonzero] XFAIL [ 23%] 2023-03-31T05:57:08.1528862Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_signed] PASSED [ 23%] 2023-03-31T05:57:08.1529425Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::is_vulkan_available] PASSED [ 23%] 2023-03-31T05:57:08.1529935Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isclose] PASSED [ 23%] 2023-03-31T05:57:08.1530442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isfinite] XFAIL [ 23%] 2023-03-31T05:57:08.1530947Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::isreal] XFAIL [ 23%] 2023-03-31T05:57:08.1531449Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::istft] PASSED [ 23%] 2023-03-31T05:57:08.1531950Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::item] XFAIL [ 23%] 2023-03-31T05:57:08.1532446Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kl_div] PASSED [ 23%] 2023-03-31T05:57:08.1532953Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kron.out] PASSED [ 23%] 2023-03-31T05:57:08.1533457Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kron] PASSED [ 23%] 2023-03-31T05:57:08.1534019Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kthvalue.dimname] PASSED [ 24%] 2023-03-31T05:57:08.1534581Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::kthvalue.dimname_out] PASSED [ 24%] 2023-03-31T05:57:08.1535103Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::l1_loss] PASSED [ 24%] 2023-03-31T05:57:08.1535613Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::layer_norm] PASSED [ 24%] 2023-03-31T05:57:08.1536132Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp.Tensor] PASSED [ 24%] 2023-03-31T05:57:08.1536642Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp.out] PASSED [ 24%] 2023-03-31T05:57:08.1537152Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ldexp_] PASSED [ 24%] 2023-03-31T05:57:08.1537664Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Scalar] PASSED [ 24%] 2023-03-31T05:57:08.1538221Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Scalar_out] PASSED [ 24%] 2023-03-31T05:57:08.1538742Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Tensor] PASSED [ 24%] 2023-03-31T05:57:08.1539252Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less.Tensor_out] PASSED [ 24%] 2023-03-31T05:57:08.1539778Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_.Scalar] PASSED [ 24%] 2023-03-31T05:57:08.1540298Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_.Tensor] PASSED [ 24%] 2023-03-31T05:57:08.1540863Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Scalar] PASSED [ 24%] 2023-03-31T05:57:08.1541385Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Scalar_out] PASSED [ 24%] 2023-03-31T05:57:08.1541915Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Tensor] PASSED [ 24%] 2023-03-31T05:57:08.1542448Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal.Tensor_out] PASSED [ 24%] 2023-03-31T05:57:08.1543138Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal_.Scalar] PASSED [ 25%] 2023-03-31T05:57:08.1543689Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::less_equal_.Tensor] PASSED [ 25%] 2023-03-31T05:57:08.1544205Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cholesky.out] PASSED [ 25%] 2023-03-31T05:57:08.1544741Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cholesky] PASSED [ 25%] 2023-03-31T05:57:08.1545299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.out] PASSED [ 25%] 2023-03-31T05:57:08.1545869Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.p_str] PASSED [ 25%] 2023-03-31T05:57:08.1546393Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond.p_str_out] PASSED [ 25%] 2023-03-31T05:57:08.1546928Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_cond] PASSED [ 25%] 2023-03-31T05:57:08.1547449Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_det.out] PASSED [ 25%] 2023-03-31T05:57:08.1547973Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_det] PASSED [ 25%] 2023-03-31T05:57:08.1548496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_diagonal] PASSED [ 25%] 2023-03-31T05:57:08.1549016Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigh.eigvals] PASSED [ 25%] 2023-03-31T05:57:08.1549580Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigh] PASSED [ 25%] 2023-03-31T05:57:08.1550111Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvals.out] PASSED [ 25%] 2023-03-31T05:57:08.1550643Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvals] PASSED [ 25%] 2023-03-31T05:57:08.1551164Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvalsh.out] PASSED [ 25%] 2023-03-31T05:57:08.1551697Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_eigvalsh] PASSED [ 25%] 2023-03-31T05:57:08.1552247Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_inv.out] PASSED [ 25%] 2023-03-31T05:57:08.1552770Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_inv] PASSED [ 26%] 2023-03-31T05:57:08.1553286Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_ldl_factor.out] PASSED [ 26%] 2023-03-31T05:57:08.1553822Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_ldl_factor] PASSED [ 26%] 2023-03-31T05:57:08.1554354Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_lu_factor.out] PASSED [ 26%] 2023-03-31T05:57:08.1554879Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_lu_factor] XFAIL [ 26%] 2023-03-31T05:57:08.1555409Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matmul.out] PASSED [ 26%] 2023-03-31T05:57:08.1555925Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matmul] PASSED [ 26%] 2023-03-31T05:57:08.1556458Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.out] PASSED [ 26%] 2023-03-31T05:57:08.1557027Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.str_ord] PASSED [ 26%] 2023-03-31T05:57:08.1557602Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm.str_ord_out] PASSED [ 26%] 2023-03-31T05:57:08.1558129Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_norm] PASSED [ 26%] 2023-03-31T05:57:08.1558665Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_power.out] PASSED [ 26%] 2023-03-31T05:57:08.1559202Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_power] PASSED [ 26%] 2023-03-31T05:57:08.1559749Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_float] PASSED [ 26%] 2023-03-31T05:57:08.1560315Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_float_out] PASSED [ 26%] 2023-03-31T05:57:08.1560911Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_tensor] PASSED [ 26%] 2023-03-31T05:57:08.1561483Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.atol_rtol_tensor_out] PASSED [ 26%] 2023-03-31T05:57:08.1562036Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.out] PASSED [ 27%] 2023-03-31T05:57:08.1562587Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.out_tol_tensor] PASSED [ 27%] 2023-03-31T05:57:08.1563263Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank.tol_tensor] PASSED [ 27%] 2023-03-31T05:57:08.1563845Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_matrix_rank] PASSED [ 27%] 2023-03-31T05:57:08.1564378Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_multi_dot.out] PASSED [ 27%] 2023-03-31T05:57:08.1564909Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_multi_dot] PASSED [ 27%] 2023-03-31T05:57:08.1565439Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.ord_str] PASSED [ 27%] 2023-03-31T05:57:08.1565963Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.ord_str_out] PASSED [ 27%] 2023-03-31T05:57:08.1566497Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm.out] PASSED [ 27%] 2023-03-31T05:57:08.1567020Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_norm] PASSED [ 27%] 2023-03-31T05:57:08.1567551Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.atol_rtol_float] XFAIL [ 27%] 2023-03-31T05:57:08.1568136Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.atol_rtol_float_out] PASSED [ 27%] 2023-03-31T05:57:08.1568698Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.out] PASSED [ 27%] 2023-03-31T05:57:08.1569243Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.out_rcond_tensor] PASSED [ 27%] 2023-03-31T05:57:08.1569788Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv.rcond_tensor] PASSED [ 27%] 2023-03-31T05:57:08.1570315Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_pinv] XFAIL [ 27%] 2023-03-31T05:57:08.1570828Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_slogdet.out] PASSED [ 27%] 2023-03-31T05:57:08.1571354Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_slogdet] XFAIL [ 28%] 2023-03-31T05:57:08.1571879Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve.out] PASSED [ 28%] 2023-03-31T05:57:08.1572438Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve] PASSED [ 28%] 2023-03-31T05:57:08.1572960Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve_ex.out] PASSED [ 28%] 2023-03-31T05:57:08.1573480Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_solve_ex] PASSED [ 28%] 2023-03-31T05:57:08.1574000Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svd.U] PASSED [ 28%] 2023-03-31T05:57:08.1574514Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svd] PASSED [ 28%] 2023-03-31T05:57:08.1575069Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svdvals.out] PASSED [ 28%] 2023-03-31T05:57:08.1575582Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_svdvals] PASSED [ 28%] 2023-03-31T05:57:08.1576109Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorinv.out] PASSED [ 28%] 2023-03-31T05:57:08.1576639Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorinv] PASSED [ 28%] 2023-03-31T05:57:08.1577176Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorsolve.out] PASSED [ 28%] 2023-03-31T05:57:08.1577717Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_tensorsolve] PASSED [ 28%] 2023-03-31T05:57:08.1578229Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vander] PASSED [ 28%] 2023-03-31T05:57:08.1578754Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vecdot.out] PASSED [ 28%] 2023-03-31T05:57:08.1579304Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linalg_vecdot] PASSED [ 28%] 2023-03-31T05:57:08.1579894Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::linear] XFAIL [ 28%] 2023-03-31T05:57:08.1580402Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_sigmoid.out] PASSED [ 29%] 2023-03-31T05:57:08.1580923Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_sigmoid] XFAIL [ 29%] 2023-03-31T05:57:08.1581454Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_softmax.Dimname] PASSED [ 29%] 2023-03-31T05:57:08.1581989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::log_softmax.int] XFAIL [ 29%] 2023-03-31T05:57:08.1582517Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logcumsumexp.dimname] PASSED [ 29%] 2023-03-31T05:57:08.1583153Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logcumsumexp.dimname_out] PASSED [ 29%] 2023-03-31T05:57:08.1583725Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logdet] XFAIL [ 29%] 2023-03-31T05:57:08.1584249Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logsumexp.names] PASSED [ 29%] 2023-03-31T05:57:08.1584788Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::logsumexp.names_out] PASSED [ 29%] 2023-03-31T05:57:08.1585304Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm.data] PASSED [ 29%] 2023-03-31T05:57:08.1585818Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm.input] PASSED [ 29%] 2023-03-31T05:57:08.1586372Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lstm_cell] PASSED [ 29%] 2023-03-31T05:57:08.1586888Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lu_solve.out] PASSED [ 29%] 2023-03-31T05:57:08.1587387Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::lu_solve] PASSED [ 29%] 2023-03-31T05:57:08.1587896Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mH] PASSED [ 29%] 2023-03-31T05:57:08.1588394Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mT] PASSED [ 29%] 2023-03-31T05:57:08.1588905Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::margin_ranking_loss] PASSED [ 29%] 2023-03-31T05:57:08.1589448Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::masked_select_backward] XFAIL [ 30%] 2023-03-31T05:57:08.1589965Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matmul.out] PASSED [ 30%] 2023-03-31T05:57:08.1590472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matmul] PASSED [ 30%] 2023-03-31T05:57:08.1591011Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_H] PASSED [ 30%] 2023-03-31T05:57:08.1591550Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_exp] PASSED [ 30%] 2023-03-31T05:57:08.1592068Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_exp_backward] PASSED [ 30%] 2023-03-31T05:57:08.1592601Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_power.out] PASSED [ 30%] 2023-03-31T05:57:08.1593123Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::matrix_power] PASSED [ 30%] 2023-03-31T05:57:08.1593643Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.names_dim] PASSED [ 30%] 2023-03-31T05:57:08.1594159Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.names_dim_max] PASSED [ 30%] 2023-03-31T05:57:08.1594677Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.other] PASSED [ 30%] 2023-03-31T05:57:08.1595231Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max.out] PASSED [ 30%] 2023-03-31T05:57:08.1595741Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool1d] PASSED [ 30%] 2023-03-31T05:57:08.1596272Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool1d_with_indices] PASSED [ 30%] 2023-03-31T05:57:08.1596789Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool2d] PASSED [ 30%] 2023-03-31T05:57:08.1597303Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::max_pool3d] PASSED [ 30%] 2023-03-31T05:57:08.1597855Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mean.names_dim] PASSED [ 30%] 2023-03-31T05:57:08.1598372Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mean.names_out] PASSED [ 30%] 2023-03-31T05:57:08.1598889Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::median.names_dim] PASSED [ 31%] 2023-03-31T05:57:08.1599430Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::median.names_dim_values] PASSED [ 31%] 2023-03-31T05:57:08.1599971Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::meshgrid.indexing] PASSED [ 31%] 2023-03-31T05:57:08.1600498Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::meshgrid] PASSED [ 31%] 2023-03-31T05:57:08.1601012Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.names_dim] PASSED [ 31%] 2023-03-31T05:57:08.1601524Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.names_dim_min] PASSED [ 31%] 2023-03-31T05:57:08.1602078Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.other] PASSED [ 31%] 2023-03-31T05:57:08.1602615Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::min.out] PASSED [ 31%] 2023-03-31T05:57:08.1603254Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mish_backward] PASSED [ 31%] 2023-03-31T05:57:08.1603771Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mode.dimname] PASSED [ 31%] 2023-03-31T05:57:08.1604294Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::mode.dimname_out] PASSED [ 31%] 2023-03-31T05:57:08.1604818Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::moveaxis.int] PASSED [ 31%] 2023-03-31T05:57:08.1605350Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::moveaxis.intlist] PASSED [ 31%] 2023-03-31T05:57:08.1605869Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::movedim.int] PASSED [ 31%] 2023-03-31T05:57:08.1606442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::movedim.intlist] XFAIL [ 31%] 2023-03-31T05:57:08.1606961Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::msort.out] PASSED [ 31%] 2023-03-31T05:57:08.1607472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::msort] PASSED [ 31%] 2023-03-31T05:57:08.1608004Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multilabel_margin_loss.out] PASSED [ 32%] 2023-03-31T05:57:08.1608537Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multilabel_margin_loss] PASSED [ 32%] 2023-03-31T05:57:08.1609114Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.Scalar] PASSED [ 32%] 2023-03-31T05:57:08.1609644Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.Tensor] PASSED [ 32%] 2023-03-31T05:57:08.1610170Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply.out] PASSED [ 32%] 2023-03-31T05:57:08.1610684Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply_.Scalar] PASSED [ 32%] 2023-03-31T05:57:08.1611210Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::multiply_.Tensor] PASSED [ 32%] 2023-03-31T05:57:08.1611727Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmean.out] PASSED [ 32%] 2023-03-31T05:57:08.1612241Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmean] PASSED [ 32%] 2023-03-31T05:57:08.1612766Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmedian.names_dim] PASSED [ 32%] 2023-03-31T05:57:08.1613299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanmedian.names_dim_values] PASSED [ 32%] 2023-03-31T05:57:08.1613871Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.out] PASSED [ 32%] 2023-03-31T05:57:08.1614444Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.scalar] PASSED [ 32%] 2023-03-31T05:57:08.1614989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile.scalar_out] PASSED [ 32%] 2023-03-31T05:57:08.1615510Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nanquantile] PASSED [ 32%] 2023-03-31T05:57:08.1616035Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::narrow.Tensor] PASSED [ 32%] 2023-03-31T05:57:08.1616555Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::narrow] PASSED [ 32%] 2023-03-31T05:57:08.1617087Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::native_channel_shuffle] PASSED [ 33%] 2023-03-31T05:57:08.1617621Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative.out] PASSED [ 33%] 2023-03-31T05:57:08.1618156Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative] PASSED [ 33%] 2023-03-31T05:57:08.1618678Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::negative_] PASSED [ 33%] 2023-03-31T05:57:08.1619211Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nested_to_padded_tensor] PASSED [ 33%] 2023-03-31T05:57:08.1619745Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss.out] PASSED [ 33%] 2023-03-31T05:57:08.1620250Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss2d.out] PASSED [ 33%] 2023-03-31T05:57:08.1620802Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss2d] PASSED [ 33%] 2023-03-31T05:57:08.1621318Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss] PASSED [ 33%] 2023-03-31T05:57:08.1621831Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nll_loss_nd] PASSED [ 33%] 2023-03-31T05:57:08.1622348Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nonzero_numpy] PASSED [ 33%] 2023-03-31T05:57:08.1622890Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_ScalarOpt_dim] PASSED [ 33%] 2023-03-31T05:57:08.1623531Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_ScalarOpt_dim_dtype] PASSED [ 33%] 2023-03-31T05:57:08.1624081Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_dtype_out] PASSED [ 33%] 2023-03-31T05:57:08.1624615Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm.names_out] PASSED [ 33%] 2023-03-31T05:57:08.1625171Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::norm_except_dim] PASSED [ 33%] 2023-03-31T05:57:08.1625733Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Scalar] PASSED [ 33%] 2023-03-31T05:57:08.1626275Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Scalar_out] PASSED [ 34%] 2023-03-31T05:57:08.1626811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Tensor] PASSED [ 34%] 2023-03-31T05:57:08.1627330Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal.Tensor_out] PASSED [ 34%] 2023-03-31T05:57:08.1627865Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal_.Scalar] PASSED [ 34%] 2023-03-31T05:57:08.1628393Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::not_equal_.Tensor] PASSED [ 34%] 2023-03-31T05:57:08.1628923Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.dim] PASSED [ 34%] 2023-03-31T05:57:08.1629486Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.dim_out] PASSED [ 34%] 2023-03-31T05:57:08.1630002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm.out] PASSED [ 34%] 2023-03-31T05:57:08.1630521Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::nuclear_norm] PASSED [ 34%] 2023-03-31T05:57:08.1631037Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::numpy_T] PASSED [ 34%] 2023-03-31T05:57:08.1631541Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::one_hot] XFAIL [ 34%] 2023-03-31T05:57:08.1632072Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::orgqr.out] PASSED [ 34%] 2023-03-31T05:57:08.1632581Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::orgqr] PASSED [ 34%] 2023-03-31T05:57:08.1633089Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::outer.out] PASSED [ 34%] 2023-03-31T05:57:08.1633599Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::outer] PASSED [ 34%] 2023-03-31T05:57:08.1634101Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::output_nr] PASSED [ 34%] 2023-03-31T05:57:08.1634612Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pad] PASSED [ 34%] 2023-03-31T05:57:08.1635134Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pad_sequence] PASSED [ 35%] 2023-03-31T05:57:08.1635665Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pairwise_distance] PASSED [ 35%] 2023-03-31T05:57:08.1636187Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pdist] PASSED [ 35%] 2023-03-31T05:57:08.1636717Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pin_memory] PASSED [ 35%] 2023-03-31T05:57:08.1637261Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::pinverse] PASSED [ 35%] 2023-03-31T05:57:08.1637786Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::poisson_nll_loss] PASSED [ 35%] 2023-03-31T05:57:08.1638305Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::positive] PASSED [ 35%] 2023-03-31T05:57:08.1638803Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prelu] PASSED [ 35%] 2023-03-31T05:57:08.1639324Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prod.Dimname_out] PASSED [ 35%] 2023-03-31T05:57:08.1639849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::prod.dim_Dimname] PASSED [ 35%] 2023-03-31T05:57:08.1640372Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::promote_types] PASSED [ 35%] 2023-03-31T05:57:08.1640916Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::qr.Q] PASSED [ 35%] 2023-03-31T05:57:08.1641418Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::qr] PASSED [ 35%] 2023-03-31T05:57:08.1641927Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.out] PASSED [ 35%] 2023-03-31T05:57:08.1642455Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.scalar] PASSED [ 35%] 2023-03-31T05:57:08.1642988Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile.scalar_out] PASSED [ 35%] 2023-03-31T05:57:08.1643654Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantile] PASSED [ 35%] 2023-03-31T05:57:08.1644179Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_gru_cell] PASSED [ 35%] 2023-03-31T05:57:08.1644716Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_lstm_cell] PASSED [ 36%] 2023-03-31T05:57:08.1645257Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_rnn_relu_cell] PASSED [ 36%] 2023-03-31T05:57:08.1645790Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::quantized_rnn_tanh_cell] PASSED [ 36%] 2023-03-31T05:57:08.1646328Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rand.generator_out] PASSED [ 36%] 2023-03-31T05:57:08.1646862Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::randn.generator_out] PASSED [ 36%] 2023-03-31T05:57:08.1647388Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::randn.out] PASSED [ 36%] 2023-03-31T05:57:08.1647943Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::ravel] PASSED [ 36%] 2023-03-31T05:57:08.1648471Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::real] XFAIL [ 36%] 2023-03-31T05:57:08.1648981Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::refine_names] PASSED [ 36%] 2023-03-31T05:57:08.1649497Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::relu6] PASSED [ 36%] 2023-03-31T05:57:08.1650000Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::relu6_] PASSED [ 36%] 2023-03-31T05:57:08.1650493Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rename] PASSED [ 36%] 2023-03-31T05:57:08.1651002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rename_] PASSED [ 36%] 2023-03-31T05:57:08.1651540Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::repeat_interleave.self_Tensor] PASSED [ 36%] 2023-03-31T05:57:08.1652136Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::repeat_interleave.self_int] PASSED [ 36%] 2023-03-31T05:57:08.1652665Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::requires_grad_] PASSED [ 36%] 2023-03-31T05:57:08.1653181Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::reshape] PASSED [ 36%] 2023-03-31T05:57:08.1653697Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::reshape_as] PASSED [ 37%] 2023-03-31T05:57:08.1654216Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::resolve_conj] PASSED [ 37%] 2023-03-31T05:57:08.1654767Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::resolve_neg] PASSED [ 37%] 2023-03-31T05:57:08.1655285Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar] PASSED [ 37%] 2023-03-31T05:57:08.1655827Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar_Scalar] PASSED [ 37%] 2023-03-31T05:57:08.1656382Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Scalar_Tensor] PASSED [ 37%] 2023-03-31T05:57:08.1656918Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::result_type.Tensor] PASSED [ 37%] 2023-03-31T05:57:08.1657427Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::retain_grad] PASSED [ 37%] 2023-03-31T05:57:08.1657948Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::retains_grad] PASSED [ 37%] 2023-03-31T05:57:08.1658465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu.data] PASSED [ 37%] 2023-03-31T05:57:08.1658984Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu.input] PASSED [ 37%] 2023-03-31T05:57:08.1659533Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_relu_cell] PASSED [ 37%] 2023-03-31T05:57:08.1660065Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh.data] PASSED [ 37%] 2023-03-31T05:57:08.1660580Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh.input] PASSED [ 37%] 2023-03-31T05:57:08.1661094Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rnn_tanh_cell] PASSED [ 37%] 2023-03-31T05:57:08.1661604Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::row_stack.out] PASSED [ 37%] 2023-03-31T05:57:08.1662110Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::row_stack] PASSED [ 37%] 2023-03-31T05:57:08.1662618Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rrelu] PASSED [ 38%] 2023-03-31T05:57:08.1663195Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::rrelu_] PASSED [ 38%] 2023-03-31T05:57:08.1663770Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scaled_dot_product_attention] PASSED [ 38%] 2023-03-31T05:57:08.1664317Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter.dimname_src] PASSED [ 38%] 2023-03-31T05:57:08.1664849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter.dimname_value] PASSED [ 38%] 2023-03-31T05:57:08.1665395Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::scatter_add.dimname] PASSED [ 38%] 2023-03-31T05:57:08.1665924Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::select.Dimname] PASSED [ 38%] 2023-03-31T05:57:08.1666472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::selu] PASSED [ 38%] 2023-03-31T05:57:08.1666965Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::selu_] PASSED [ 38%] 2023-03-31T05:57:08.1667497Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::set_.source_Tensor_storage_offset] PASSED [ 38%] 2023-03-31T05:57:08.1668033Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::set_data] PASSED [ 38%] 2023-03-31T05:57:08.1668553Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::silu_backward] XFAIL [ 38%] 2023-03-31T05:57:08.1669060Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::size.Dimname] PASSED [ 38%] 2023-03-31T05:57:08.1669574Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::size.int] PASSED [ 38%] 2023-03-31T05:57:08.1670083Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slogdet.out] PASSED [ 38%] 2023-03-31T05:57:08.1670661Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slogdet] PASSED [ 38%] 2023-03-31T05:57:08.1671208Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slow_conv3d.out] PASSED [ 38%] 2023-03-31T05:57:08.1671711Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::slow_conv3d] PASSED [ 39%] 2023-03-31T05:57:08.1672227Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::smm] PASSED [ 39%] 2023-03-31T05:57:08.1672746Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::softmax.Dimname] PASSED [ 39%] 2023-03-31T05:57:08.1673270Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::softmax.int] PASSED [ 39%] 2023-03-31T05:57:08.1673777Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname] PASSED [ 39%] 2023-03-31T05:57:08.1674306Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_stable] PASSED [ 39%] 2023-03-31T05:57:08.1674876Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_values] PASSED [ 39%] 2023-03-31T05:57:08.1675418Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sort.dimname_values_stable] PASSED [ 39%] 2023-03-31T05:57:08.1675976Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsc_tensor.ccol_row_value] PASSED [ 39%] 2023-03-31T05:57:08.1676539Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsc_tensor.ccol_row_value_size] PASSED [ 39%] 2023-03-31T05:57:08.1677108Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsr_tensor.crow_col_value] PASSED [ 39%] 2023-03-31T05:57:08.1677716Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_bsr_tensor.crow_col_value_size] PASSED [ 39%] 2023-03-31T05:57:08.1678291Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_compressed_tensor.comp_plain_value] PASSED [ 39%] 2023-03-31T05:57:08.1678869Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_compressed_tensor.comp_plain_value_size] PASSED [ 39%] 2023-03-31T05:57:08.1679440Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_coo_tensor.indices] PASSED [ 39%] 2023-03-31T05:57:08.1679994Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_coo_tensor.indices_size] PASSED [ 39%] 2023-03-31T05:57:08.1680556Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csc_tensor.ccol_row_value] PASSED [ 39%] 2023-03-31T05:57:08.1681123Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csc_tensor.ccol_row_value_size] PASSED [ 40%] 2023-03-31T05:57:08.1681715Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csr_tensor.crow_col_value] PASSED [ 40%] 2023-03-31T05:57:08.1682308Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sparse_csr_tensor.crow_col_value_size] PASSED [ 40%] 2023-03-31T05:57:08.1682877Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.n_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1683583Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.x_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1684162Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_t.x_scalar_out] PASSED [ 40%] 2023-03-31T05:57:08.1684732Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.n_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1685305Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.x_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1685928Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_u.x_scalar_out] PASSED [ 40%] 2023-03-31T05:57:08.1686505Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.n_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1687079Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.x_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1687639Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_v.x_scalar_out] PASSED [ 40%] 2023-03-31T05:57:08.1688218Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.n_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1688819Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.x_scalar] PASSED [ 40%] 2023-03-31T05:57:08.1689395Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_chebyshev_polynomial_w.x_scalar_out] PASSED [ 40%] 2023-03-31T05:57:08.1689937Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_digamma.out] PASSED [ 40%] 2023-03-31T05:57:08.1690470Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_digamma] PASSED [ 40%] 2023-03-31T05:57:08.1690991Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erf.out] PASSED [ 40%] 2023-03-31T05:57:08.1691518Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erf] PASSED [ 41%] 2023-03-31T05:57:08.1692040Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfc.out] PASSED [ 41%] 2023-03-31T05:57:08.1692552Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfc] PASSED [ 41%] 2023-03-31T05:57:08.1693106Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfinv.out] PASSED [ 41%] 2023-03-31T05:57:08.1693664Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_erfinv] PASSED [ 41%] 2023-03-31T05:57:08.1694196Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_exp2.out] PASSED [ 41%] 2023-03-31T05:57:08.1694713Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_exp2] PASSED [ 41%] 2023-03-31T05:57:08.1695238Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expit.out] PASSED [ 41%] 2023-03-31T05:57:08.1695771Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expit] PASSED [ 41%] 2023-03-31T05:57:08.1696307Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expm1.out] PASSED [ 41%] 2023-03-31T05:57:08.1696838Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_expm1] PASSED [ 41%] 2023-03-31T05:57:08.1697392Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammainc.out] PASSED [ 41%] 2023-03-31T05:57:08.1697929Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammainc] PASSED [ 41%] 2023-03-31T05:57:08.1698464Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaincc.out] PASSED [ 41%] 2023-03-31T05:57:08.1699007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaincc] PASSED [ 41%] 2023-03-31T05:57:08.1699535Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaln.out] PASSED [ 41%] 2023-03-31T05:57:08.1700094Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_gammaln] PASSED [ 41%] 2023-03-31T05:57:08.1700646Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.n_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1701213Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.x_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1701794Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_h.x_scalar_out] PASSED [ 42%] 2023-03-31T05:57:08.1702360Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.n_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1702934Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.x_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1703590Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_hermite_polynomial_he.x_scalar_out] PASSED [ 42%] 2023-03-31T05:57:08.1704146Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_i0.out] PASSED [ 42%] 2023-03-31T05:57:08.1704704Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_i0] PASSED [ 42%] 2023-03-31T05:57:08.1705272Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.n_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1705847Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.x_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1706430Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_laguerre_polynomial_l.x_scalar_out] PASSED [ 42%] 2023-03-31T05:57:08.1707007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.n_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1707568Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.x_scalar] PASSED [ 42%] 2023-03-31T05:57:08.1708144Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_legendre_polynomial_p.x_scalar_out] PASSED [ 42%] 2023-03-31T05:57:08.1708739Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log1p.out] PASSED [ 42%] 2023-03-31T05:57:08.1709265Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log1p] PASSED [ 42%] 2023-03-31T05:57:08.1709798Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_log_softmax] PASSED [ 42%] 2023-03-31T05:57:08.1710323Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logit.out] PASSED [ 43%] 2023-03-31T05:57:08.1710846Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logit] PASSED [ 43%] 2023-03-31T05:57:08.1711411Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logsumexp.out] PASSED [ 43%] 2023-03-31T05:57:08.1711956Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_logsumexp] PASSED [ 43%] 2023-03-31T05:57:08.1712489Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_multigammaln.out] PASSED [ 43%] 2023-03-31T05:57:08.1713041Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_multigammaln] PASSED [ 43%] 2023-03-31T05:57:08.1713573Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_ndtr.out] PASSED [ 43%] 2023-03-31T05:57:08.1714097Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_ndtr] PASSED [ 43%] 2023-03-31T05:57:08.1714628Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_polygamma.out] PASSED [ 43%] 2023-03-31T05:57:08.1715152Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_polygamma] PASSED [ 43%] 2023-03-31T05:57:08.1715717Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_psi.out] PASSED [ 43%] 2023-03-31T05:57:08.1716271Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_psi] PASSED [ 43%] 2023-03-31T05:57:08.1716799Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_round.out] PASSED [ 43%] 2023-03-31T05:57:08.1717312Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_round] PASSED [ 43%] 2023-03-31T05:57:08.1717876Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.n_scalar] PASSED [ 43%] 2023-03-31T05:57:08.1718472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.x_scalar] PASSED [ 43%] 2023-03-31T05:57:08.1719068Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_t.x_scalar_out] PASSED [ 43%] 2023-03-31T05:57:08.1719708Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.n_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1720289Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.x_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1720882Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_u.x_scalar_out] PASSED [ 44%] 2023-03-31T05:57:08.1721472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.n_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1722057Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.x_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1722678Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_v.x_scalar_out] PASSED [ 44%] 2023-03-31T05:57:08.1723367Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.n_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1723953Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.x_scalar] PASSED [ 44%] 2023-03-31T05:57:08.1724546Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_shifted_chebyshev_polynomial_w.x_scalar_out] PASSED [ 44%] 2023-03-31T05:57:08.1725112Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_sinc.out] PASSED [ 44%] 2023-03-31T05:57:08.1725637Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_sinc] PASSED [ 44%] 2023-03-31T05:57:08.1726151Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_softmax] PASSED [ 44%] 2023-03-31T05:57:08.1726734Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.other_scalar] XFAIL [ 44%] 2023-03-31T05:57:08.1727321Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.other_scalar_out] PASSED [ 44%] 2023-03-31T05:57:08.1727867Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.out] PASSED [ 44%] 2023-03-31T05:57:08.1728395Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.self_scalar] XFAIL [ 44%] 2023-03-31T05:57:08.1728953Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy.self_scalar_out] PASSED [ 44%] 2023-03-31T05:57:08.1729495Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::special_xlogy] XFAIL [ 45%] 2023-03-31T05:57:08.1730018Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::split.sizes] PASSED [ 45%] 2023-03-31T05:57:08.1730535Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square.out] PASSED [ 45%] 2023-03-31T05:57:08.1731077Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square] PASSED [ 45%] 2023-03-31T05:57:08.1731587Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::square_] PASSED [ 45%] 2023-03-31T05:57:08.1732109Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::squeeze.dimname] PASSED [ 45%] 2023-03-31T05:57:08.1732638Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::squeeze_.dimname] PASSED [ 45%] 2023-03-31T05:57:08.1733151Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sspaddmm] PASSED [ 45%] 2023-03-31T05:57:08.1733714Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.correction_names] PASSED [ 45%] 2023-03-31T05:57:08.1734257Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.correction_names_out] PASSED [ 45%] 2023-03-31T05:57:08.1734786Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.dim] PASSED [ 45%] 2023-03-31T05:57:08.1735302Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.names_dim] PASSED [ 45%] 2023-03-31T05:57:08.1735811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.names_out] PASSED [ 45%] 2023-03-31T05:57:08.1736323Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std.out] PASSED [ 45%] 2023-03-31T05:57:08.1736831Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std] PASSED [ 45%] 2023-03-31T05:57:08.1737362Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.correction_names] PASSED [ 45%] 2023-03-31T05:57:08.1737884Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.dim] PASSED [ 45%] 2023-03-31T05:57:08.1738442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean.names_dim] PASSED [ 45%] 2023-03-31T05:57:08.1738991Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::std_mean] PASSED [ 46%] 2023-03-31T05:57:08.1739513Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stft.center] PASSED [ 46%] 2023-03-31T05:57:08.1740015Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stft] PASSED [ 46%] 2023-03-31T05:57:08.1740531Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stride.Dimname] PASSED [ 46%] 2023-03-31T05:57:08.1741055Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::stride.int] PASSED [ 46%] 2023-03-31T05:57:08.1741579Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.Scalar] PASSED [ 46%] 2023-03-31T05:57:08.1742112Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.Tensor] PASSED [ 46%] 2023-03-31T05:57:08.1742660Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract.out] PASSED [ 46%] 2023-03-31T05:57:08.1743276Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract_.Scalar] PASSED [ 46%] 2023-03-31T05:57:08.1743809Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::subtract_.Tensor] PASSED [ 46%] 2023-03-31T05:57:08.1744344Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum.DimnameList_out] PASSED [ 46%] 2023-03-31T05:57:08.1744868Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum.dim_DimnameList] PASSED [ 46%] 2023-03-31T05:57:08.1745430Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::sum_to_size] PASSED [ 46%] 2023-03-31T05:57:08.1745943Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::svd.U] PASSED [ 46%] 2023-03-31T05:57:08.1746452Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::svd] PASSED [ 46%] 2023-03-31T05:57:08.1746959Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapaxes] PASSED [ 46%] 2023-03-31T05:57:08.1747462Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapaxes_] PASSED [ 46%] 2023-03-31T05:57:08.1747981Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapdims] PASSED [ 47%] 2023-03-31T05:57:08.1748493Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::swapdims_] PASSED [ 47%] 2023-03-31T05:57:08.1749013Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::take_along_dim.out] PASSED [ 47%] 2023-03-31T05:57:08.1749557Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::take_along_dim] PASSED [ 47%] 2023-03-31T05:57:08.1750115Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.indices] XFAIL [ 47%] 2023-03-31T05:57:08.1750650Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.sections] XFAIL [ 47%] 2023-03-31T05:57:08.1751214Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensor_split.tensor_indices_or_sections] PASSED [ 47%] 2023-03-31T05:57:08.1751748Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tensordot] PASSED [ 47%] 2023-03-31T05:57:08.1752266Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::thnn_conv2d.out] PASSED [ 47%] 2023-03-31T05:57:08.1752786Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::thnn_conv2d] PASSED [ 47%] 2023-03-31T05:57:08.1753296Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::tile] PASSED [ 47%] 2023-03-31T05:57:08.1753829Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.device] XFAIL [ 47%] 2023-03-31T05:57:08.1754328Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.dtype] XFAIL [ 47%] 2023-03-31T05:57:08.1754843Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.dtype_layout] XFAIL [ 47%] 2023-03-31T05:57:08.1755356Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to.other] XFAIL [ 47%] 2023-03-31T05:57:08.1755867Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_dense] PASSED [ 47%] 2023-03-31T05:57:08.1756442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_dense_backward] PASSED [ 47%] 2023-03-31T05:57:08.1756972Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::to_mkldnn_backward] PASSED [ 48%] 2023-03-31T05:57:08.1757497Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trace_backward] PASSED [ 48%] 2023-03-31T05:57:08.1758034Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::transpose.Dimname] PASSED [ 48%] 2023-03-31T05:57:08.1758567Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapezoid.dx] PASSED [ 48%] 2023-03-31T05:57:08.1759075Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapezoid.x] PASSED [ 48%] 2023-03-31T05:57:08.1759594Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapz.dx] PASSED [ 48%] 2023-03-31T05:57:08.1760106Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::trapz.x] PASSED [ 48%] 2023-03-31T05:57:08.1760628Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::triplet_margin_loss] PASSED [ 48%] 2023-03-31T05:57:08.1761187Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.Scalar] PASSED [ 48%] 2023-03-31T05:57:08.1761744Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.Tensor] PASSED [ 48%] 2023-03-31T05:57:08.1762276Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide.out] PASSED [ 48%] 2023-03-31T05:57:08.1762805Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide_.Scalar] PASSED [ 48%] 2023-03-31T05:57:08.1763457Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::true_divide_.Tensor] PASSED [ 48%] 2023-03-31T05:57:08.1763976Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::type_as] PASSED [ 48%] 2023-03-31T05:57:08.1764496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unbind.Dimname] PASSED [ 48%] 2023-03-31T05:57:08.1765029Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten.Dimname] PASSED [ 48%] 2023-03-31T05:57:08.1765602Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten.int] PASSED [ 48%] 2023-03-31T05:57:08.1766121Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unflatten_dense_tensors] PASSED [ 49%] 2023-03-31T05:57:08.1766655Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::unsafe_chunk] PASSED [ 49%] 2023-03-31T05:57:08.1767186Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_bicubic2d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1767725Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_bilinear2d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1768299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_linear1d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1768836Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest1d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1769386Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest2d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1769924Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_nearest3d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1770468Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::upsample_trilinear3d.vec] XFAIL [ 49%] 2023-03-31T05:57:08.1771018Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::value_selecting_reduction_backward] PASSED [ 49%] 2023-03-31T05:57:08.1771556Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vander] PASSED [ 49%] 2023-03-31T05:57:08.1772081Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.correction_names] PASSED [ 49%] 2023-03-31T05:57:08.1772655Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.correction_names_out] PASSED [ 49%] 2023-03-31T05:57:08.1773206Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.dim] PASSED [ 49%] 2023-03-31T05:57:08.1773728Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.names_dim] PASSED [ 49%] 2023-03-31T05:57:08.1774255Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.names_out] PASSED [ 49%] 2023-03-31T05:57:08.1774771Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var.out] PASSED [ 49%] 2023-03-31T05:57:08.1775277Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var] PASSED [ 50%] 2023-03-31T05:57:08.1775802Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.correction_names] PASSED [ 50%] 2023-03-31T05:57:08.1776341Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.dim] PASSED [ 50%] 2023-03-31T05:57:08.1776900Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean.names_dim] PASSED [ 50%] 2023-03-31T05:57:08.1777509Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::var_mean] PASSED [ 50%] 2023-03-31T05:57:08.1778039Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::view_as] PASSED [ 50%] 2023-03-31T05:57:08.1778629Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vsplit.array] PASSED [ 50%] 2023-03-31T05:57:08.1779218Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vsplit.int] PASSED [ 50%] 2023-03-31T05:57:08.1779965Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vstack.out] PASSED [ 50%] 2023-03-31T05:57:08.1780572Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::vstack] PASSED [ 50%] 2023-03-31T05:57:08.1781126Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.ScalarOther] PASSED [ 50%] 2023-03-31T05:57:08.1781761Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.ScalarSelf] PASSED [ 50%] 2023-03-31T05:57:08.1782356Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where.Scalar] PASSED [ 50%] 2023-03-31T05:57:08.1782963Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[aten::where] XFAIL [ 50%] 2023-03-31T05:57:08.1783640Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[nvprims::var_mean] PASSED [ 50%] 2023-03-31T05:57:08.1784178Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[nvprims::view.shape] PASSED [ 50%] 2023-03-31T05:57:08.1784831Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[prepacked::unpack_prepacked_sizes_conv2d] PASSED [ 50%] 2023-03-31T05:57:08.1785529Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[prepacked::unpack_prepacked_sizes_linear] PASSED [ 50%] 2023-03-31T05:57:08.1786144Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_enter] PASSED [ 51%] 2023-03-31T05:57:08.1786747Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_enter_new] PASSED [ 51%] 2023-03-31T05:57:08.1787378Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_exit._RecordFunction] PASSED [ 51%] 2023-03-31T05:57:08.1788016Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[profiler::_record_function_exit] PASSED [ 51%] 2023-03-31T05:57:08.1788637Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv1d_unpack] PASSED [ 51%] 2023-03-31T05:57:08.1789242Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_dilation] PASSED [ 51%] 2023-03-31T05:57:08.1789827Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_groups] PASSED [ 51%] 2023-03-31T05:57:08.1790438Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_output_padding] PASSED [ 51%] 2023-03-31T05:57:08.1800273Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_padding] PASSED [ 51%] 2023-03-31T05:57:08.1800855Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_stride] PASSED [ 51%] 2023-03-31T05:57:08.1801392Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_transpose] PASSED [ 51%] 2023-03-31T05:57:08.1802054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_unpack] PASSED [ 51%] 2023-03-31T05:57:08.1802600Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv2d_unpack_sizes] PASSED [ 51%] 2023-03-31T05:57:08.1803467Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_dilation] PASSED [ 51%] 2023-03-31T05:57:08.1804007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_groups] PASSED [ 51%] 2023-03-31T05:57:08.1804566Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_output_padding] PASSED [ 51%] 2023-03-31T05:57:08.1805119Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_padding] PASSED [ 51%] 2023-03-31T05:57:08.1805654Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_stride] PASSED [ 52%] 2023-03-31T05:57:08.1806199Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_transpose] PASSED [ 52%] 2023-03-31T05:57:08.1806807Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv3d_unpack] PASSED [ 52%] 2023-03-31T05:57:08.1807408Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose1d_unpack] PASSED [ 52%] 2023-03-31T05:57:08.1807983Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_dilation] PASSED [ 52%] 2023-03-31T05:57:08.1808551Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_groups] PASSED [ 52%] 2023-03-31T05:57:08.1809106Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_output_padding] PASSED [ 52%] 2023-03-31T05:57:08.1809683Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_padding] PASSED [ 52%] 2023-03-31T05:57:08.1810245Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_stride] PASSED [ 52%] 2023-03-31T05:57:08.1810863Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_transpose] PASSED [ 52%] 2023-03-31T05:57:08.1811426Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose2d_unpack] PASSED [ 52%] 2023-03-31T05:57:08.1811973Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_dilation] PASSED [ 52%] 2023-03-31T05:57:08.1812538Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_groups] PASSED [ 52%] 2023-03-31T05:57:08.1813110Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_output_padding] PASSED [ 52%] 2023-03-31T05:57:08.1813717Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_padding] PASSED [ 52%] 2023-03-31T05:57:08.1814265Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_stride] PASSED [ 52%] 2023-03-31T05:57:08.1814822Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_transpose] PASSED [ 52%] 2023-03-31T05:57:08.1815384Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_transpose3d_unpack] PASSED [ 53%] 2023-03-31T05:57:08.1815927Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::conv_unpack] PASSED [ 53%] 2023-03-31T05:57:08.1816474Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::embedding_bag_unpack] PASSED [ 53%] 2023-03-31T05:57:08.1817009Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::linear_unpack] PASSED [ 53%] 2023-03-31T05:57:08.1817549Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::linear_unpack_fp16] PASSED [ 53%] 2023-03-31T05:57:08.1818187Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[quantized::make_quantized_cell_params_fp16] PASSED [ 53%] 2023-03-31T05:57:08.1818766Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_a_batching_rule_for_composite_implicit_autograd_[sparse::qlinear_unpack] PASSED [ 53%] 2023-03-31T05:57:08.1819290Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__and__.Scalar] PASSED [ 53%] 2023-03-31T05:57:08.1819775Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__and__.Tensor] PASSED [ 53%] 2023-03-31T05:57:08.1820273Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__iand__.Scalar] PASSED [ 53%] 2023-03-31T05:57:08.1820773Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__iand__.Tensor] PASSED [ 53%] 2023-03-31T05:57:08.1821271Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ior__.Scalar] PASSED [ 53%] 2023-03-31T05:57:08.1821748Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ior__.Tensor] PASSED [ 53%] 2023-03-31T05:57:08.1822271Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ixor__.Scalar] PASSED [ 53%] 2023-03-31T05:57:08.1822768Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__ixor__.Tensor] PASSED [ 53%] 2023-03-31T05:57:08.1823357Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__or__.Scalar] PASSED [ 53%] 2023-03-31T05:57:08.1823865Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__or__.Tensor] PASSED [ 53%] 2023-03-31T05:57:08.1824359Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__xor__.Scalar] PASSED [ 54%] 2023-03-31T05:57:08.1824849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::__xor__.Tensor] PASSED [ 54%] 2023-03-31T05:57:08.1825394Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_batch_norm_impl_index] PASSED [ 54%] 2023-03-31T05:57:08.1825906Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_convolution_double_backward] PASSED [ 54%] 2023-03-31T05:57:08.1826422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_convolution_mode] PASSED [ 54%] 2023-03-31T05:57:08.1826950Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_has_compatible_shallow_copy_type] PASSED [ 54%] 2023-03-31T05:57:08.1827470Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_lu_with_info] PASSED [ 54%] 2023-03-31T05:57:08.1827953Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_pad_circular] PASSED [ 54%] 2023-03-31T05:57:08.1828451Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::_test_check_tensor] PASSED [ 54%] 2023-03-31T05:57:08.1828945Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::absolute] PASSED [ 54%] 2023-03-31T05:57:08.1829448Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool1d] PASSED [ 54%] 2023-03-31T05:57:08.1829992Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool2d] PASSED [ 54%] 2023-03-31T05:57:08.1830515Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_avg_pool3d] PASSED [ 54%] 2023-03-31T05:57:08.1831028Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adaptive_max_pool1d] PASSED [ 54%] 2023-03-31T05:57:08.1831522Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::adjoint] PASSED [ 54%] 2023-03-31T05:57:08.1832007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arccos] PASSED [ 54%] 2023-03-31T05:57:08.1832477Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arccosh] PASSED [ 54%] 2023-03-31T05:57:08.1832958Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arcsin] PASSED [ 55%] 2023-03-31T05:57:08.1833436Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arcsinh] PASSED [ 55%] 2023-03-31T05:57:08.1833925Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctan2] PASSED [ 55%] 2023-03-31T05:57:08.1834430Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctan] PASSED [ 55%] 2023-03-31T05:57:08.1834905Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::arctanh] PASSED [ 55%] 2023-03-31T05:57:08.1835382Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::argsort] PASSED [ 55%] 2023-03-31T05:57:08.1835871Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_1d.Sequence] PASSED [ 55%] 2023-03-31T05:57:08.1836356Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_1d] PASSED [ 55%] 2023-03-31T05:57:08.1836891Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_2d.Sequence] PASSED [ 55%] 2023-03-31T05:57:08.1837389Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_2d] PASSED [ 55%] 2023-03-31T05:57:08.1837887Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_3d.Sequence] PASSED [ 55%] 2023-03-31T05:57:08.1838376Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::atleast_3d] PASSED [ 55%] 2023-03-31T05:57:08.1838864Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::avg_pool1d] PASSED [ 55%] 2023-03-31T05:57:08.1839351Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::batch_norm] PASSED [ 55%] 2023-03-31T05:57:08.1839851Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_and_.Scalar] PASSED [ 55%] 2023-03-31T05:57:08.1840354Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_or.Scalar] PASSED [ 55%] 2023-03-31T05:57:08.1840847Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_or_.Scalar] PASSED [ 55%] 2023-03-31T05:57:08.1841374Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_xor.Scalar] PASSED [ 55%] 2023-03-31T05:57:08.1841900Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::bitwise_xor_.Scalar] PASSED [ 56%] 2023-03-31T05:57:08.1842402Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::broadcast_tensors] PASSED [ 56%] 2023-03-31T05:57:08.1842893Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::broadcast_to] PASSED [ 56%] 2023-03-31T05:57:08.1843547Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cartesian_prod] PASSED [ 56%] 2023-03-31T05:57:08.1844038Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cdist] PASSED [ 56%] 2023-03-31T05:57:08.1844521Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::chunk] PASSED [ 56%] 2023-03-31T05:57:08.1845004Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::clip.Tensor] PASSED [ 56%] 2023-03-31T05:57:08.1845498Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::clip] PASSED [ 56%] 2023-03-31T05:57:08.1846038Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::combinations] PASSED [ 56%] 2023-03-31T05:57:08.1846525Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::concat] PASSED [ 56%] 2023-03-31T05:57:08.1847004Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::concatenate] PASSED [ 56%] 2023-03-31T05:57:08.1847504Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conj_physical] PASSED [ 56%] 2023-03-31T05:57:08.1848002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::contiguous] PASSED [ 56%] 2023-03-31T05:57:08.1848496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv1d.padding] PASSED [ 56%] 2023-03-31T05:57:08.1849010Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv1d] PASSED [ 56%] 2023-03-31T05:57:08.1849496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv2d.padding] PASSED [ 56%] 2023-03-31T05:57:08.1849975Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv2d] PASSED [ 56%] 2023-03-31T05:57:08.1850465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv3d.padding] PASSED [ 57%] 2023-03-31T05:57:08.1850938Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv3d] PASSED [ 57%] 2023-03-31T05:57:08.1851425Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose1d] PASSED [ 57%] 2023-03-31T05:57:08.1851945Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose2d.input] PASSED [ 57%] 2023-03-31T05:57:08.1852459Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::conv_transpose3d.input] PASSED [ 57%] 2023-03-31T05:57:08.1852964Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::corrcoef] PASSED [ 57%] 2023-03-31T05:57:08.1853494Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cosine_embedding_loss] PASSED [ 57%] 2023-03-31T05:57:08.1854038Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cosine_similarity] PASSED [ 57%] 2023-03-31T05:57:08.1854534Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cov] PASSED [ 57%] 2023-03-31T05:57:08.1855018Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cross] PASSED [ 57%] 2023-03-31T05:57:08.1855495Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cross_entropy_loss] PASSED [ 57%] 2023-03-31T05:57:08.1855999Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumprod_backward] PASSED [ 57%] 2023-03-31T05:57:08.1856509Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumulative_trapezoid.dx] PASSED [ 57%] 2023-03-31T05:57:08.1857034Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::cumulative_trapezoid.x] PASSED [ 57%] 2023-03-31T05:57:08.1857520Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::det] PASSED [ 57%] 2023-03-31T05:57:08.1858024Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diag] PASSED [ 57%] 2023-03-31T05:57:08.1858511Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diagonal_copy] XFAIL [ 57%] 2023-03-31T05:57:08.1858995Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::diff] PASSED [ 58%] 2023-03-31T05:57:08.1859471Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Scalar] PASSED [ 58%] 2023-03-31T05:57:08.1859976Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Scalar_mode] PASSED [ 58%] 2023-03-31T05:57:08.1860516Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Tensor] PASSED [ 58%] 2023-03-31T05:57:08.1861020Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide.Tensor_mode] PASSED [ 58%] 2023-03-31T05:57:08.1861516Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Scalar_mode] PASSED [ 58%] 2023-03-31T05:57:08.1862014Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Tensor] PASSED [ 58%] 2023-03-31T05:57:08.1862515Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::divide_.Tensor_mode] PASSED [ 58%] 2023-03-31T05:57:08.1863104Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dsplit.array] PASSED [ 58%] 2023-03-31T05:57:08.1863609Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dsplit.int] PASSED [ 58%] 2023-03-31T05:57:08.1864085Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::dstack] PASSED [ 58%] 2023-03-31T05:57:08.1864564Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::einsum] PASSED [ 58%] 2023-03-31T05:57:08.1865058Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::embedding_backward] PASSED [ 58%] 2023-03-31T05:57:08.1865603Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::expand_as] PASSED [ 58%] 2023-03-31T05:57:08.1866107Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fft2] PASSED [ 58%] 2023-03-31T05:57:08.1866586Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fft] PASSED [ 58%] 2023-03-31T05:57:08.1867069Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fftn] PASSED [ 58%] 2023-03-31T05:57:08.1867316Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_fftshift] PASSED [ 59%] 2023-03-31T05:57:08.1867553Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfft2] PASSED [ 59%] 2023-03-31T05:57:08.1867792Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfft] PASSED [ 59%] 2023-03-31T05:57:08.1868018Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_hfftn] PASSED [ 59%] 2023-03-31T05:57:08.1868256Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifft2] PASSED [ 59%] 2023-03-31T05:57:08.1868520Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifft] PASSED [ 59%] 2023-03-31T05:57:08.1868759Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifftn] PASSED [ 59%] 2023-03-31T05:57:08.1869005Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ifftshift] PASSED [ 59%] 2023-03-31T05:57:08.1869244Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_ihfft] PASSED [ 59%] 2023-03-31T05:57:08.1869484Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfft2] PASSED [ 59%] 2023-03-31T05:57:08.1869753Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfft] PASSED [ 59%] 2023-03-31T05:57:08.1869994Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_irfftn] PASSED [ 59%] 2023-03-31T05:57:08.1870229Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfft2] PASSED [ 59%] 2023-03-31T05:57:08.1870465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfft] PASSED [ 59%] 2023-03-31T05:57:08.1870691Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fft_rfftn] PASSED [ 59%] 2023-03-31T05:57:08.1870928Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fix] PASSED [ 59%] 2023-03-31T05:57:08.1871167Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::fliplr] PASSED [ 59%] 2023-03-31T05:57:08.1871405Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::flipud] PASSED [ 60%] 2023-03-31T05:57:08.1871655Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Scalar] PASSED [ 60%] 2023-03-31T05:57:08.1871916Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Tensor_Scalar] PASSED [ 60%] 2023-03-31T05:57:08.1872248Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::float_power.Tensor_Tensor] PASSED [ 60%] 2023-03-31T05:57:08.1872530Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::floor_divide.Scalar] PASSED [ 60%] 2023-03-31T05:57:08.1872780Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::frobenius_norm.dim] PASSED [ 60%] 2023-03-31T05:57:08.1873029Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gather_backward] PASSED [ 60%] 2023-03-31T05:57:08.1873260Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ger] PASSED [ 60%] 2023-03-31T05:57:08.1873494Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.array] PASSED [ 60%] 2023-03-31T05:57:08.1873753Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalararray] PASSED [ 60%] 2023-03-31T05:57:08.1874003Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarint] PASSED [ 60%] 2023-03-31T05:57:08.1874298Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarrayarray] PASSED [ 60%] 2023-03-31T05:57:08.1874560Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.scalarrayint] PASSED [ 60%] 2023-03-31T05:57:08.1874817Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.tensorarray] PASSED [ 60%] 2023-03-31T05:57:08.1875084Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::gradient.tensorarrayint] PASSED [ 60%] 2023-03-31T05:57:08.1875333Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater.Scalar] PASSED [ 60%] 2023-03-31T05:57:08.1875609Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater.Tensor] PASSED [ 60%] 2023-03-31T05:57:08.1875863Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater_equal.Scalar] PASSED [ 60%] 2023-03-31T05:57:08.1876116Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::greater_equal.Tensor] PASSED [ 61%] 2023-03-31T05:57:08.1876344Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::grid_sampler] PASSED [ 61%] 2023-03-31T05:57:08.1876587Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::group_norm] PASSED [ 61%] 2023-03-31T05:57:08.1876840Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hinge_embedding_loss] PASSED [ 61%] 2023-03-31T05:57:08.1877085Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hsplit.array] PASSED [ 61%] 2023-03-31T05:57:08.1877327Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hsplit.int] PASSED [ 61%] 2023-03-31T05:57:08.1877563Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::hstack] PASSED [ 61%] 2023-03-31T05:57:08.1877817Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::index_select_backward] PASSED [ 61%] 2023-03-31T05:57:08.1878086Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::inner] PASSED [ 61%] 2023-03-31T05:57:08.1878356Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::instance_norm] PASSED [ 61%] 2023-03-31T05:57:08.1878596Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::inverse] PASSED [ 61%] 2023-03-31T05:57:08.1878837Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::is_complex] PASSED [ 61%] 2023-03-31T05:57:08.1879063Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::is_same_size] XFAIL [ 61%] 2023-03-31T05:57:08.1879300Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::kron] PASSED [ 61%] 2023-03-31T05:57:08.1879539Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::l1_loss] PASSED [ 61%] 2023-03-31T05:57:08.1879780Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::layer_norm] PASSED [ 61%] 2023-03-31T05:57:08.1880027Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ldexp.Tensor] PASSED [ 61%] 2023-03-31T05:57:08.1880306Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less.Scalar] PASSED [ 62%] 2023-03-31T05:57:08.1880550Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less.Tensor] PASSED [ 62%] 2023-03-31T05:57:08.1880800Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less_equal.Scalar] PASSED [ 62%] 2023-03-31T05:57:08.1881044Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::less_equal.Tensor] PASSED [ 62%] 2023-03-31T05:57:08.1881290Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_cholesky] PASSED [ 62%] 2023-03-31T05:57:08.1881559Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_cond] PASSED [ 62%] 2023-03-31T05:57:08.1881784Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_det] PASSED [ 62%] 2023-03-31T05:57:08.1882026Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_diagonal] PASSED [ 62%] 2023-03-31T05:57:08.1882268Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_eigvals] PASSED [ 62%] 2023-03-31T05:57:08.1882514Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_eigvalsh] PASSED [ 62%] 2023-03-31T05:57:08.1882747Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_inv] PASSED [ 62%] 2023-03-31T05:57:08.1882996Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_ldl_factor] PASSED [ 62%] 2023-03-31T05:57:08.1883495Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matmul] PASSED [ 62%] 2023-03-31T05:57:08.1883762Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_norm.str_ord] PASSED [ 62%] 2023-03-31T05:57:08.1884065Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_norm] PASSED [ 62%] 2023-03-31T05:57:08.1884348Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_power] PASSED [ 62%] 2023-03-31T05:57:08.1884612Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_rank.atol_rtol_float] PASSED [ 62%] 2023-03-31T05:57:08.1884869Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_matrix_rank.atol_rtol_tensor] PASSED [ 63%] 2023-03-31T05:57:08.1885118Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_multi_dot] PASSED [ 63%] 2023-03-31T05:57:08.1885365Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_norm.ord_str] PASSED [ 63%] 2023-03-31T05:57:08.1885610Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_norm] PASSED [ 63%] 2023-03-31T05:57:08.1885852Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_solve] PASSED [ 63%] 2023-03-31T05:57:08.1886096Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_solve_ex] PASSED [ 63%] 2023-03-31T05:57:08.1886374Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_svd] PASSED [ 63%] 2023-03-31T05:57:08.1886616Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_svdvals] PASSED [ 63%] 2023-03-31T05:57:08.1886861Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_tensorinv] PASSED [ 63%] 2023-03-31T05:57:08.1887381Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_vander] PASSED [ 63%] 2023-03-31T05:57:08.1887875Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::linalg_vecdot] PASSED [ 63%] 2023-03-31T05:57:08.1888442Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::mH] PASSED [ 63%] 2023-03-31T05:57:08.1888922Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::mT] PASSED [ 63%] 2023-03-31T05:57:08.1889398Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matmul] PASSED [ 63%] 2023-03-31T05:57:08.1890085Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matrix_H] PASSED [ 63%] 2023-03-31T05:57:08.1890656Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::matrix_power] PASSED [ 63%] 2023-03-31T05:57:08.1891221Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max.other] PASSED [ 63%] 2023-03-31T05:57:08.1891731Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max_pool1d_with_indices] PASSED [ 64%] 2023-03-31T05:57:08.1892222Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::max_pool2d] PASSED [ 64%] 2023-03-31T05:57:08.1892728Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::meshgrid.indexing] PASSED [ 64%] 2023-03-31T05:57:08.1893224Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::meshgrid] PASSED [ 64%] 2023-03-31T05:57:08.1893742Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::min.other] PASSED [ 64%] 2023-03-31T05:57:08.1894261Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::moveaxis.intlist] PASSED [ 64%] 2023-03-31T05:57:08.1894771Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::movedim.int] PASSED [ 64%] 2023-03-31T05:57:08.1895272Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::msort] PASSED [ 64%] 2023-03-31T05:57:08.1895772Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply.Scalar] PASSED [ 64%] 2023-03-31T05:57:08.1896280Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply.Tensor] PASSED [ 64%] 2023-03-31T05:57:08.1896811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply_.Scalar] PASSED [ 64%] 2023-03-31T05:57:08.1897313Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::multiply_.Tensor] PASSED [ 64%] 2023-03-31T05:57:08.1897862Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nanmean] PASSED [ 64%] 2023-03-31T05:57:08.1898423Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::narrow] PASSED [ 64%] 2023-03-31T05:57:08.1898989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::negative] PASSED [ 64%] 2023-03-31T05:57:08.1899474Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss2d] PASSED [ 64%] 2023-03-31T05:57:08.1899968Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss] PASSED [ 64%] 2023-03-31T05:57:08.1900466Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nll_loss_nd] PASSED [ 65%] 2023-03-31T05:57:08.1901016Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::not_equal.Scalar] PASSED [ 65%] 2023-03-31T05:57:08.1901505Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::not_equal.Tensor] PASSED [ 65%] 2023-03-31T05:57:08.1902009Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nuclear_norm.dim] PASSED [ 65%] 2023-03-31T05:57:08.1902515Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::nuclear_norm] PASSED [ 65%] 2023-03-31T05:57:08.1903098Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::numpy_T] PASSED [ 65%] 2023-03-31T05:57:08.1903580Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::orgqr] PASSED [ 65%] 2023-03-31T05:57:08.1904068Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::outer] PASSED [ 65%] 2023-03-31T05:57:08.1904550Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pad] PASSED [ 65%] 2023-03-31T05:57:08.1905047Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pairwise_distance] PASSED [ 65%] 2023-03-31T05:57:08.1905537Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::pinverse] PASSED [ 65%] 2023-03-31T05:57:08.1906093Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::poisson_nll_loss] PASSED [ 65%] 2023-03-31T05:57:08.1906626Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::positive] PASSED [ 65%] 2023-03-31T05:57:08.1907295Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::prelu] PASSED [ 65%] 2023-03-31T05:57:08.1907854Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::qr] PASSED [ 65%] 2023-03-31T05:57:08.1908339Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::ravel] PASSED [ 65%] 2023-03-31T05:57:08.1908823Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::relu6] PASSED [ 65%] 2023-03-31T05:57:08.1909309Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::relu6_] PASSED [ 65%] 2023-03-31T05:57:08.1909811Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::repeat_interleave.self_Tensor] PASSED [ 66%] 2023-03-31T05:57:08.1910399Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::repeat_interleave.self_int] PASSED [ 66%] 2023-03-31T05:57:08.1910913Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::reshape] PASSED [ 66%] 2023-03-31T05:57:08.1911400Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::reshape_as] PASSED [ 66%] 2023-03-31T05:57:08.1911893Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::resolve_conj] PASSED [ 66%] 2023-03-31T05:57:08.1912392Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::resolve_neg] PASSED [ 66%] 2023-03-31T05:57:08.1912893Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar] PASSED [ 66%] 2023-03-31T05:57:08.1913466Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar_Scalar] PASSED [ 66%] 2023-03-31T05:57:08.1913981Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Scalar_Tensor] PASSED [ 66%] 2023-03-31T05:57:08.1914499Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::result_type.Tensor] PASSED [ 66%] 2023-03-31T05:57:08.1915052Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::row_stack] PASSED [ 66%] 2023-03-31T05:57:08.1915613Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::rrelu] PASSED [ 66%] 2023-03-31T05:57:08.1916096Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::rrelu_] PASSED [ 66%] 2023-03-31T05:57:08.1916599Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::scaled_dot_product_attention] PASSED [ 66%] 2023-03-31T05:57:08.1917102Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::selu] PASSED [ 66%] 2023-03-31T05:57:08.1917582Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::selu_] PASSED [ 66%] 2023-03-31T05:57:08.1918097Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::size.int] PASSED [ 66%] 2023-03-31T05:57:08.1918597Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::slogdet] PASSED [ 67%] 2023-03-31T05:57:08.1919090Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::softmax.int] PASSED [ 67%] 2023-03-31T05:57:08.1919596Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_digamma] PASSED [ 67%] 2023-03-31T05:57:08.1920098Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erf] PASSED [ 67%] 2023-03-31T05:57:08.1920584Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erfc] PASSED [ 67%] 2023-03-31T05:57:08.1921089Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_erfinv] PASSED [ 67%] 2023-03-31T05:57:08.1921588Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_exp2] PASSED [ 67%] 2023-03-31T05:57:08.1922089Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_expit] PASSED [ 67%] 2023-03-31T05:57:08.1922625Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_expm1] PASSED [ 67%] 2023-03-31T05:57:08.1923275Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammainc] PASSED [ 67%] 2023-03-31T05:57:08.1923789Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammaincc] PASSED [ 67%] 2023-03-31T05:57:08.1924301Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_gammaln] PASSED [ 67%] 2023-03-31T05:57:08.1924795Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_i0] PASSED [ 67%] 2023-03-31T05:57:08.1925383Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_log1p] PASSED [ 67%] 2023-03-31T05:57:08.1925901Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_log_softmax] PASSED [ 67%] 2023-03-31T05:57:08.1926409Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_logit] PASSED [ 67%] 2023-03-31T05:57:08.1926917Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_logsumexp] PASSED [ 67%] 2023-03-31T05:57:08.1927422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_multigammaln] PASSED [ 68%] 2023-03-31T05:57:08.1927936Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_ndtr] PASSED [ 68%] 2023-03-31T05:57:08.1928445Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_polygamma] PASSED [ 68%] 2023-03-31T05:57:08.1928960Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_psi] PASSED [ 68%] 2023-03-31T05:57:08.1929445Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_round] PASSED [ 68%] 2023-03-31T05:57:08.1929944Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_sinc] PASSED [ 68%] 2023-03-31T05:57:08.1930495Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::special_softmax] PASSED [ 68%] 2023-03-31T05:57:08.1931026Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::split.sizes] PASSED [ 68%] 2023-03-31T05:57:08.1931510Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::square] PASSED [ 68%] 2023-03-31T05:57:08.1932002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std.dim] PASSED [ 68%] 2023-03-31T05:57:08.1932483Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std] PASSED [ 68%] 2023-03-31T05:57:08.1932971Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std_mean.dim] PASSED [ 68%] 2023-03-31T05:57:08.1933456Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::std_mean] PASSED [ 68%] 2023-03-31T05:57:08.1933960Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::subtract.Tensor] PASSED [ 68%] 2023-03-31T05:57:08.1934472Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::sum_to_size] PASSED [ 68%] 2023-03-31T05:57:08.1934999Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::svd] PASSED [ 68%] 2023-03-31T05:57:08.1935470Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapaxes] PASSED [ 68%] 2023-03-31T05:57:08.1935963Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapaxes_] PASSED [ 69%] 2023-03-31T05:57:08.1936459Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapdims] PASSED [ 69%] 2023-03-31T05:57:08.1936954Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::swapdims_] PASSED [ 69%] 2023-03-31T05:57:08.1937474Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::take_along_dim] PASSED [ 69%] 2023-03-31T05:57:08.1937974Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::tensordot] PASSED [ 69%] 2023-03-31T05:57:08.1938461Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::tile] PASSED [ 69%] 2023-03-31T05:57:08.1938948Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapezoid.dx] PASSED [ 69%] 2023-03-31T05:57:08.1939439Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapezoid.x] PASSED [ 69%] 2023-03-31T05:57:08.1939934Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapz.dx] PASSED [ 69%] 2023-03-31T05:57:08.1940427Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::trapz.x] PASSED [ 69%] 2023-03-31T05:57:08.1940932Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide.Scalar] PASSED [ 69%] 2023-03-31T05:57:08.1941443Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide.Tensor] PASSED [ 69%] 2023-03-31T05:57:08.1941941Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide_.Scalar] PASSED [ 69%] 2023-03-31T05:57:08.1942496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::true_divide_.Tensor] PASSED [ 69%] 2023-03-31T05:57:08.1943099Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::type_as] PASSED [ 69%] 2023-03-31T05:57:08.1943608Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::unflatten.int] PASSED [ 69%] 2023-03-31T05:57:08.1944093Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::unfold_copy] XFAIL [ 69%] 2023-03-31T05:57:08.1944612Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::value_selecting_reduction_backward] PASSED [ 70%] 2023-03-31T05:57:08.1945130Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var.dim] PASSED [ 70%] 2023-03-31T05:57:08.1945615Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var] PASSED [ 70%] 2023-03-31T05:57:08.1946089Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var_mean.dim] PASSED [ 70%] 2023-03-31T05:57:08.1946626Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::var_mean] PASSED [ 70%] 2023-03-31T05:57:08.1947114Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::view_as] PASSED [ 70%] 2023-03-31T05:57:08.1947602Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vsplit.array] PASSED [ 70%] 2023-03-31T05:57:08.1948088Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vsplit.int] PASSED [ 70%] 2023-03-31T05:57:08.1948582Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::vstack] PASSED [ 70%] 2023-03-31T05:57:08.1949078Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.ScalarOther] PASSED [ 70%] 2023-03-31T05:57:08.1949635Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.ScalarSelf] PASSED [ 70%] 2023-03-31T05:57:08.1950128Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_register_functorch_batched_decomposition_[aten::where.Scalar] PASSED [ 70%] 2023-03-31T05:57:08.1950624Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::absolute] PASSED [ 70%] 2023-03-31T05:57:08.1951110Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::absolute_] XFAIL [ 70%] 2023-03-31T05:57:08.1951607Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool1d] PASSED [ 70%] 2023-03-31T05:57:08.1952099Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool2d] PASSED [ 70%] 2023-03-31T05:57:08.1952610Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_avg_pool3d] PASSED [ 70%] 2023-03-31T05:57:08.1953107Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adaptive_max_pool1d] PASSED [ 70%] 2023-03-31T05:57:08.1953604Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::adjoint] PASSED [ 71%] 2023-03-31T05:57:08.1954150Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::affine_grid_generator_backward] XFAIL [ 71%] 2023-03-31T05:57:08.1954666Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_as] XFAIL [ 71%] 2023-03-31T05:57:08.1955144Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_tensors] XFAIL [ 71%] 2023-03-31T05:57:08.1955642Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_to.ellipsis_idx] XFAIL [ 71%] 2023-03-31T05:57:08.1956131Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::align_to] XFAIL [ 71%] 2023-03-31T05:57:08.1956599Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::alpha_dropout] XFAIL [ 71%] 2023-03-31T05:57:08.1957091Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::alpha_dropout_] XFAIL [ 71%] 2023-03-31T05:57:08.1957578Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccos] PASSED [ 71%] 2023-03-31T05:57:08.1958044Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccos_] XFAIL [ 71%] 2023-03-31T05:57:08.1958552Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccosh] PASSED [ 71%] 2023-03-31T05:57:08.1959028Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arccosh_] XFAIL [ 71%] 2023-03-31T05:57:08.1959506Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsin] PASSED [ 71%] 2023-03-31T05:57:08.1959977Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsin_] XFAIL [ 71%] 2023-03-31T05:57:08.1960440Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsinh] PASSED [ 71%] 2023-03-31T05:57:08.1960913Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arcsinh_] XFAIL [ 71%] 2023-03-31T05:57:08.1961409Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan2] PASSED [ 71%] 2023-03-31T05:57:08.1961875Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan2_] XFAIL [ 72%] 2023-03-31T05:57:08.1962327Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan] PASSED [ 72%] 2023-03-31T05:57:08.1962804Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctan_] XFAIL [ 72%] 2023-03-31T05:57:08.1963516Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctanh] PASSED [ 72%] 2023-03-31T05:57:08.1963993Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::arctanh_] XFAIL [ 72%] 2023-03-31T05:57:08.1964447Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::argsort] PASSED [ 72%] 2023-03-31T05:57:08.1964929Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::argwhere] XFAIL [ 72%] 2023-03-31T05:57:08.1965416Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_1d.Sequence] PASSED [ 72%] 2023-03-31T05:57:08.1965910Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_1d] PASSED [ 72%] 2023-03-31T05:57:08.1966462Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_2d.Sequence] PASSED [ 72%] 2023-03-31T05:57:08.1966994Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_2d] PASSED [ 72%] 2023-03-31T05:57:08.1967488Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_3d.Sequence] PASSED [ 72%] 2023-03-31T05:57:08.1967989Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::atleast_3d] PASSED [ 72%] 2023-03-31T05:57:08.1968455Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::avg_pool1d] PASSED [ 72%] 2023-03-31T05:57:08.1968932Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::batch_norm] PASSED [ 72%] 2023-03-31T05:57:08.1969413Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bilinear] XFAIL [ 72%] 2023-03-31T05:57:08.1969898Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_and_.Scalar] PASSED [ 72%] 2023-03-31T05:57:08.1970386Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_or.Scalar] PASSED [ 73%] 2023-03-31T05:57:08.1970934Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_or_.Scalar] PASSED [ 73%] 2023-03-31T05:57:08.1971431Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_xor.Scalar] PASSED [ 73%] 2023-03-31T05:57:08.1971929Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::bitwise_xor_.Scalar] PASSED [ 73%] 2023-03-31T05:57:08.1972422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::broadcast_tensors] PASSED [ 73%] 2023-03-31T05:57:08.1972914Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::broadcast_to] PASSED [ 73%] 2023-03-31T05:57:08.1973393Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::can_cast] XFAIL [ 73%] 2023-03-31T05:57:08.1973924Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cartesian_prod] PASSED [ 73%] 2023-03-31T05:57:08.1974415Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cat.names] XFAIL [ 73%] 2023-03-31T05:57:08.1974879Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cdist] PASSED [ 73%] 2023-03-31T05:57:08.1975358Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chain_matmul] XFAIL [ 73%] 2023-03-31T05:57:08.1975839Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chalf] XFAIL [ 73%] 2023-03-31T05:57:08.1976335Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::choose_qparams_optimized] XFAIL [ 73%] 2023-03-31T05:57:08.1976813Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::chunk] PASSED [ 73%] 2023-03-31T05:57:08.1977288Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip.Tensor] PASSED [ 73%] 2023-03-31T05:57:08.1977763Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip] PASSED [ 73%] 2023-03-31T05:57:08.1978271Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip_.Tensor] XFAIL [ 73%] 2023-03-31T05:57:08.1978758Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::clip_] XFAIL [ 74%] 2023-03-31T05:57:08.1979233Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::coalesce] XFAIL [ 74%] 2023-03-31T05:57:08.1979709Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::column_stack] XFAIL [ 74%] 2023-03-31T05:57:08.1980204Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::combinations] PASSED [ 74%] 2023-03-31T05:57:08.1980680Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concat.names] XFAIL [ 74%] 2023-03-31T05:57:08.1981157Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concat] PASSED [ 74%] 2023-03-31T05:57:08.1981646Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concatenate.names] XFAIL [ 74%] 2023-03-31T05:57:08.1982137Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::concatenate] PASSED [ 74%] 2023-03-31T05:57:08.1982602Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conj] XFAIL [ 74%] 2023-03-31T05:57:08.1983294Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conj_physical] PASSED [ 74%] 2023-03-31T05:57:08.1983788Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::contiguous] PASSED [ 74%] 2023-03-31T05:57:08.1984277Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv1d.padding] PASSED [ 74%] 2023-03-31T05:57:08.1984756Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv1d] PASSED [ 74%] 2023-03-31T05:57:08.1985246Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv2d.padding] PASSED [ 74%] 2023-03-31T05:57:08.1985768Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv2d] PASSED [ 74%] 2023-03-31T05:57:08.1986252Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv3d.padding] PASSED [ 74%] 2023-03-31T05:57:08.1986726Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv3d] PASSED [ 74%] 2023-03-31T05:57:08.1987208Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_tbc_backward] XFAIL [ 75%] 2023-03-31T05:57:08.1987703Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose1d] PASSED [ 75%] 2023-03-31T05:57:08.1988207Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose2d.input] PASSED [ 75%] 2023-03-31T05:57:08.1988720Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::conv_transpose3d.input] PASSED [ 75%] 2023-03-31T05:57:08.1989226Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::corrcoef] PASSED [ 75%] 2023-03-31T05:57:08.1989724Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cosine_embedding_loss] PASSED [ 75%] 2023-03-31T05:57:08.1990228Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cosine_similarity] PASSED [ 75%] 2023-03-31T05:57:08.1990729Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cov] PASSED [ 75%] 2023-03-31T05:57:08.1991223Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cross] PASSED [ 75%] 2023-03-31T05:57:08.1991710Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cross_entropy_loss] PASSED [ 75%] 2023-03-31T05:57:08.1992214Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ctc_loss.IntList] XFAIL [ 75%] 2023-03-31T05:57:08.1992691Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ctc_loss.Tensor] XFAIL [ 75%] 2023-03-31T05:57:08.1993184Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cudnn_is_acceptable] XFAIL [ 75%] 2023-03-31T05:57:08.1993684Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cummaxmin_backward] XFAIL [ 75%] 2023-03-31T05:57:08.1994188Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumprod_backward] PASSED [ 75%] 2023-03-31T05:57:08.1994699Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumulative_trapezoid.dx] PASSED [ 75%] 2023-03-31T05:57:08.1995235Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::cumulative_trapezoid.x] PASSED [ 75%] 2023-03-31T05:57:08.1995727Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::data] XFAIL [ 75%] 2023-03-31T05:57:08.1996195Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::det] PASSED [ 76%] 2023-03-31T05:57:08.1996670Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diag] PASSED [ 76%] 2023-03-31T05:57:08.1997133Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diagflat] XFAIL [ 76%] 2023-03-31T05:57:08.1997605Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::diff] PASSED [ 76%] 2023-03-31T05:57:08.1998124Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Scalar] PASSED [ 76%] 2023-03-31T05:57:08.1998627Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Scalar_mode] PASSED [ 76%] 2023-03-31T05:57:08.1999111Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Tensor] PASSED [ 76%] 2023-03-31T05:57:08.1999614Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.Tensor_mode] PASSED [ 76%] 2023-03-31T05:57:08.2000112Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide.out_mode] XFAIL [ 76%] 2023-03-31T05:57:08.2000602Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Scalar] XFAIL [ 76%] 2023-03-31T05:57:08.2001085Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Scalar_mode] PASSED [ 76%] 2023-03-31T05:57:08.2001584Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Tensor] PASSED [ 76%] 2023-03-31T05:57:08.2002082Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::divide_.Tensor_mode] PASSED [ 76%] 2023-03-31T05:57:08.2002607Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dropout] XFAIL [ 76%] 2023-03-31T05:57:08.2003240Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dropout_] XFAIL [ 76%] 2023-03-31T05:57:08.2003721Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dsplit.array] PASSED [ 76%] 2023-03-31T05:57:08.2004221Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dsplit.int] PASSED [ 76%] 2023-03-31T05:57:08.2004702Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::dstack] PASSED [ 77%] 2023-03-31T05:57:08.2005164Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::einsum] PASSED [ 77%] 2023-03-31T05:57:08.2005656Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_backward] PASSED [ 77%] 2023-03-31T05:57:08.2006172Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_bag.padding_idx] XFAIL [ 77%] 2023-03-31T05:57:08.2006676Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::embedding_bag] XFAIL [ 77%] 2023-03-31T05:57:08.2007208Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::expand_as] PASSED [ 77%] 2023-03-31T05:57:08.2007704Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_alpha_dropout] XFAIL [ 77%] 2023-03-31T05:57:08.2008207Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_alpha_dropout_] XFAIL [ 77%] 2023-03-31T05:57:08.2008707Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_dropout] XFAIL [ 77%] 2023-03-31T05:57:08.2009187Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::feature_dropout_] XFAIL [ 77%] 2023-03-31T05:57:08.2009669Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fft2] PASSED [ 77%] 2023-03-31T05:57:08.2010197Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fft] PASSED [ 77%] 2023-03-31T05:57:08.2010669Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fftn] PASSED [ 77%] 2023-03-31T05:57:08.2011138Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_fftshift] PASSED [ 77%] 2023-03-31T05:57:08.2011626Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfft2] PASSED [ 77%] 2023-03-31T05:57:08.2012098Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfft] PASSED [ 77%] 2023-03-31T05:57:08.2012573Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_hfftn] PASSED [ 77%] 2023-03-31T05:57:08.2013060Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifft2] PASSED [ 78%] 2023-03-31T05:57:08.2013524Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifft] PASSED [ 78%] 2023-03-31T05:57:08.2014003Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifftn] PASSED [ 78%] 2023-03-31T05:57:08.2014486Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ifftshift] PASSED [ 78%] 2023-03-31T05:57:08.2015003Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfft2] XFAIL [ 78%] 2023-03-31T05:57:08.2015498Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfft] PASSED [ 78%] 2023-03-31T05:57:08.2015978Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_ihfftn] XFAIL [ 78%] 2023-03-31T05:57:08.2016455Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfft2] PASSED [ 78%] 2023-03-31T05:57:08.2016937Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfft] PASSED [ 78%] 2023-03-31T05:57:08.2017402Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_irfftn] PASSED [ 78%] 2023-03-31T05:57:08.2017886Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfft2] PASSED [ 78%] 2023-03-31T05:57:08.2018364Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfft] PASSED [ 78%] 2023-03-31T05:57:08.2018851Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fft_rfftn] PASSED [ 78%] 2023-03-31T05:57:08.2019359Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fill_diagonal_] XFAIL [ 78%] 2023-03-31T05:57:08.2019837Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fix] PASSED [ 78%] 2023-03-31T05:57:08.2020304Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fix_] XFAIL [ 78%] 2023-03-31T05:57:08.2020787Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.named_out_dim] XFAIL [ 78%] 2023-03-31T05:57:08.2021275Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.using_ints] XFAIL [ 79%] 2023-03-31T05:57:08.2021770Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten.using_names] XFAIL [ 79%] 2023-03-31T05:57:08.2022306Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flatten_dense_tensors] XFAIL [ 79%] 2023-03-31T05:57:08.2022793Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fliplr] PASSED [ 79%] 2023-03-31T05:57:08.2023340Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::flipud] PASSED [ 79%] 2023-03-31T05:57:08.2023831Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Scalar] PASSED [ 79%] 2023-03-31T05:57:08.2024342Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Tensor_Scalar] PASSED [ 79%] 2023-03-31T05:57:08.2024860Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power.Tensor_Tensor] PASSED [ 79%] 2023-03-31T05:57:08.2025354Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power_.Scalar] XFAIL [ 79%] 2023-03-31T05:57:08.2025849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::float_power_.Tensor] XFAIL [ 79%] 2023-03-31T05:57:08.2026347Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::floor_divide.Scalar] PASSED [ 79%] 2023-03-31T05:57:08.2026895Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::floor_divide_.Scalar] XFAIL [ 79%] 2023-03-31T05:57:08.2027404Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::frobenius_norm.dim] PASSED [ 79%] 2023-03-31T05:57:08.2027917Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::fused_moving_avg_obs_fake_quant] XFAIL [ 79%] 2023-03-31T05:57:08.2028432Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gather_backward] PASSED [ 79%] 2023-03-31T05:57:08.2028915Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ger] PASSED [ 79%] 2023-03-31T05:57:08.2029390Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::get_gradients] XFAIL [ 79%] 2023-03-31T05:57:08.2029872Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.array] PASSED [ 80%] 2023-03-31T05:57:08.2030381Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalararray] PASSED [ 80%] 2023-03-31T05:57:08.2030891Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarint] PASSED [ 80%] 2023-03-31T05:57:08.2031451Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarrayarray] PASSED [ 80%] 2023-03-31T05:57:08.2031960Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.scalarrayint] PASSED [ 80%] 2023-03-31T05:57:08.2032479Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.tensorarray] PASSED [ 80%] 2023-03-31T05:57:08.2032999Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gradient.tensorarrayint] PASSED [ 80%] 2023-03-31T05:57:08.2033516Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater.Scalar] PASSED [ 80%] 2023-03-31T05:57:08.2034033Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater.Tensor] PASSED [ 80%] 2023-03-31T05:57:08.2034521Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_.Scalar] XFAIL [ 80%] 2023-03-31T05:57:08.2035002Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_.Tensor] XFAIL [ 80%] 2023-03-31T05:57:08.2035491Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal.Scalar] PASSED [ 80%] 2023-03-31T05:57:08.2035978Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal.Tensor] PASSED [ 80%] 2023-03-31T05:57:08.2036483Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal_.Scalar] XFAIL [ 80%] 2023-03-31T05:57:08.2036976Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::greater_equal_.Tensor] XFAIL [ 80%] 2023-03-31T05:57:08.2037465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::grid_sampler] PASSED [ 80%] 2023-03-31T05:57:08.2037937Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::group_norm] PASSED [ 80%] 2023-03-31T05:57:08.2038413Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru.data] XFAIL [ 80%] 2023-03-31T05:57:08.2038920Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru.input] XFAIL [ 81%] 2023-03-31T05:57:08.2039421Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::gru_cell] XFAIL [ 81%] 2023-03-31T05:57:08.2039901Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hinge_embedding_loss] PASSED [ 81%] 2023-03-31T05:57:08.2040421Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd.TensorList_bins] XFAIL [ 81%] 2023-03-31T05:57:08.2040935Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd.int_bins] XFAIL [ 81%] 2023-03-31T05:57:08.2041423Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::histogramdd] XFAIL [ 81%] 2023-03-31T05:57:08.2041902Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hsplit.array] PASSED [ 81%] 2023-03-31T05:57:08.2042393Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hsplit.int] PASSED [ 81%] 2023-03-31T05:57:08.2042876Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::hstack] PASSED [ 81%] 2023-03-31T05:57:08.2043639Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::imag] XFAIL [ 81%] 2023-03-31T05:57:08.2044130Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::index_select_backward] PASSED [ 81%] 2023-03-31T05:57:08.2044646Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::infinitely_differentiable_gelu_backward] XFAIL [ 81%] 2023-03-31T05:57:08.2045155Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::inner] PASSED [ 81%] 2023-03-31T05:57:08.2045638Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::instance_norm] PASSED [ 81%] 2023-03-31T05:57:08.2046158Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::inverse] PASSED [ 81%] 2023-03-31T05:57:08.2046621Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isclose] XFAIL [ 81%] 2023-03-31T05:57:08.2047092Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isfinite] XFAIL [ 81%] 2023-03-31T05:57:08.2047564Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::isreal] XFAIL [ 82%] 2023-03-31T05:57:08.2048035Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::istft] XFAIL [ 82%] 2023-03-31T05:57:08.2048487Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::item] XFAIL [ 82%] 2023-03-31T05:57:08.2048958Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::kl_div] XFAIL [ 82%] 2023-03-31T05:57:08.2049431Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::kron] PASSED [ 82%] 2023-03-31T05:57:08.2049901Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::l1_loss] PASSED [ 82%] 2023-03-31T05:57:08.2050364Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::layer_norm] PASSED [ 82%] 2023-03-31T05:57:08.2050897Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ldexp.Tensor] PASSED [ 82%] 2023-03-31T05:57:08.2051410Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ldexp_] XFAIL [ 82%] 2023-03-31T05:57:08.2051891Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less.Scalar] PASSED [ 82%] 2023-03-31T05:57:08.2052369Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less.Tensor] PASSED [ 82%] 2023-03-31T05:57:08.2052847Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_.Scalar] XFAIL [ 82%] 2023-03-31T05:57:08.2053322Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_.Tensor] XFAIL [ 82%] 2023-03-31T05:57:08.2053816Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal.Scalar] PASSED [ 82%] 2023-03-31T05:57:08.2054299Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal.Tensor] PASSED [ 82%] 2023-03-31T05:57:08.2054787Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal_.Scalar] XFAIL [ 82%] 2023-03-31T05:57:08.2055309Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::less_equal_.Tensor] XFAIL [ 82%] 2023-03-31T05:57:08.2055804Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cholesky] PASSED [ 83%] 2023-03-31T05:57:08.2056285Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cond.p_str] XFAIL [ 83%] 2023-03-31T05:57:08.2056778Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_cond] PASSED [ 83%] 2023-03-31T05:57:08.2057260Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_det] PASSED [ 83%] 2023-03-31T05:57:08.2057750Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_diagonal] PASSED [ 83%] 2023-03-31T05:57:08.2058271Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigh.eigvals] XFAIL [ 83%] 2023-03-31T05:57:08.2058762Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigh] XFAIL [ 83%] 2023-03-31T05:57:08.2059251Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigvals] PASSED [ 83%] 2023-03-31T05:57:08.2059744Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_eigvalsh] PASSED [ 83%] 2023-03-31T05:57:08.2060218Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_inv] PASSED [ 83%] 2023-03-31T05:57:08.2060709Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_ldl_factor] PASSED [ 83%] 2023-03-31T05:57:08.2061202Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_lu_factor] XFAIL [ 83%] 2023-03-31T05:57:08.2061688Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matmul] PASSED [ 83%] 2023-03-31T05:57:08.2066883Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_norm.str_ord] PASSED [ 83%] 2023-03-31T05:57:08.2070859Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_norm] PASSED [ 83%] 2023-03-31T05:57:08.2071364Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_power] PASSED [ 83%] 2023-03-31T05:57:08.2071874Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.atol_rtol_float] PASSED [ 83%] 2023-03-31T05:57:08.2072418Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.atol_rtol_tensor] PASSED [ 84%] 2023-03-31T05:57:08.2072952Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.out_tol_tensor] XFAIL [ 84%] 2023-03-31T05:57:08.2073498Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank.tol_tensor] XFAIL [ 84%] 2023-03-31T05:57:08.2074031Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_matrix_rank] XFAIL [ 84%] 2023-03-31T05:57:08.2074530Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_multi_dot] PASSED [ 84%] 2023-03-31T05:57:08.2075033Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_norm.ord_str] PASSED [ 84%] 2023-03-31T05:57:08.2075573Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_norm] PASSED [ 84%] 2023-03-31T05:57:08.2076080Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.atol_rtol_float] XFAIL [ 84%] 2023-03-31T05:57:08.2076599Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.out_rcond_tensor] XFAIL [ 84%] 2023-03-31T05:57:08.2077121Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv.rcond_tensor] XFAIL [ 84%] 2023-03-31T05:57:08.2077604Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_pinv] XFAIL [ 84%] 2023-03-31T05:57:08.2078133Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_slogdet] XFAIL [ 84%] 2023-03-31T05:57:08.2078623Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_solve] PASSED [ 84%] 2023-03-31T05:57:08.2079114Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_solve_ex] PASSED [ 84%] 2023-03-31T05:57:08.2079592Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svd.U] XFAIL [ 84%] 2023-03-31T05:57:08.2080082Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svd] PASSED [ 84%] 2023-03-31T05:57:08.2080570Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_svdvals] PASSED [ 84%] 2023-03-31T05:57:08.2081072Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_tensorinv] PASSED [ 85%] 2023-03-31T05:57:08.2081572Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_tensorsolve] XFAIL [ 85%] 2023-03-31T05:57:08.2082050Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_vander] PASSED [ 85%] 2023-03-31T05:57:08.2082588Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linalg_vecdot] PASSED [ 85%] 2023-03-31T05:57:08.2083338Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::linear] XFAIL [ 85%] 2023-03-31T05:57:08.2083838Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::log_sigmoid] XFAIL [ 85%] 2023-03-31T05:57:08.2084314Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::log_softmax.int] XFAIL [ 85%] 2023-03-31T05:57:08.2084795Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::logdet] XFAIL [ 85%] 2023-03-31T05:57:08.2085279Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::logsumexp.names] XFAIL [ 85%] 2023-03-31T05:57:08.2085768Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm.data] XFAIL [ 85%] 2023-03-31T05:57:08.2086235Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm.input] XFAIL [ 85%] 2023-03-31T05:57:08.2086715Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lstm_cell] XFAIL [ 85%] 2023-03-31T05:57:08.2087192Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::lu_solve] XFAIL [ 85%] 2023-03-31T05:57:08.2087719Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mH] PASSED [ 85%] 2023-03-31T05:57:08.2088174Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mT] PASSED [ 85%] 2023-03-31T05:57:08.2088655Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::margin_ranking_loss] XFAIL [ 85%] 2023-03-31T05:57:08.2089162Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::masked_select_backward] XFAIL [ 85%] 2023-03-31T05:57:08.2089648Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matmul] PASSED [ 85%] 2023-03-31T05:57:08.2090110Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_H] PASSED [ 86%] 2023-03-31T05:57:08.2090640Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_exp] XFAIL [ 86%] 2023-03-31T05:57:08.2091127Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_exp_backward] XFAIL [ 86%] 2023-03-31T05:57:08.2091624Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::matrix_power] PASSED [ 86%] 2023-03-31T05:57:08.2092102Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.names_dim] XFAIL [ 86%] 2023-03-31T05:57:08.2092586Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.names_dim_max] XFAIL [ 86%] 2023-03-31T05:57:08.2093069Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max.other] PASSED [ 86%] 2023-03-31T05:57:08.2093552Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool1d] XFAIL [ 86%] 2023-03-31T05:57:08.2094031Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool1d_with_indices] PASSED [ 86%] 2023-03-31T05:57:08.2094525Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool2d] PASSED [ 86%] 2023-03-31T05:57:08.2095054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::max_pool3d] XFAIL [ 86%] 2023-03-31T05:57:08.2095565Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mean.names_dim] XFAIL [ 86%] 2023-03-31T05:57:08.2096039Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::median.names_dim] XFAIL [ 86%] 2023-03-31T05:57:08.2096541Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::median.names_dim_values] XFAIL [ 86%] 2023-03-31T05:57:08.2097048Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::meshgrid.indexing] PASSED [ 86%] 2023-03-31T05:57:08.2097540Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::meshgrid] PASSED [ 86%] 2023-03-31T05:57:08.2098010Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.names_dim] XFAIL [ 86%] 2023-03-31T05:57:08.2098499Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.names_dim_min] XFAIL [ 87%] 2023-03-31T05:57:08.2098981Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::min.other] PASSED [ 87%] 2023-03-31T05:57:08.2099496Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::mish_backward] XFAIL [ 87%] 2023-03-31T05:57:08.2099969Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::moveaxis.int] XFAIL [ 87%] 2023-03-31T05:57:08.2100454Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::moveaxis.intlist] PASSED [ 87%] 2023-03-31T05:57:08.2100947Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::movedim.int] PASSED [ 87%] 2023-03-31T05:57:08.2101441Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::movedim.intlist] XFAIL [ 87%] 2023-03-31T05:57:08.2101922Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::msort] PASSED [ 87%] 2023-03-31T05:57:08.2102426Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multilabel_margin_loss] XFAIL [ 87%] 2023-03-31T05:57:08.2102925Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply.Scalar] PASSED [ 87%] 2023-03-31T05:57:08.2103512Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply.Tensor] PASSED [ 87%] 2023-03-31T05:57:08.2104011Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply_.Scalar] PASSED [ 87%] 2023-03-31T05:57:08.2104499Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::multiply_.Tensor] PASSED [ 87%] 2023-03-31T05:57:08.2104981Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmean] PASSED [ 87%] 2023-03-31T05:57:08.2105478Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmedian.names_dim] XFAIL [ 87%] 2023-03-31T05:57:08.2105985Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanmedian.names_dim_values] XFAIL [ 87%] 2023-03-31T05:57:08.2106482Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanquantile.scalar] XFAIL [ 87%] 2023-03-31T05:57:08.2107009Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nanquantile] XFAIL [ 88%] 2023-03-31T05:57:08.2107523Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::narrow.Tensor] XFAIL [ 88%] 2023-03-31T05:57:08.2108007Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::narrow] PASSED [ 88%] 2023-03-31T05:57:08.2108488Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::native_channel_shuffle] XFAIL [ 88%] 2023-03-31T05:57:08.2108976Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::negative] PASSED [ 88%] 2023-03-31T05:57:08.2109448Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::negative_] XFAIL [ 88%] 2023-03-31T05:57:08.2109941Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nested_to_padded_tensor] XFAIL [ 88%] 2023-03-31T05:57:08.2110422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss2d] PASSED [ 88%] 2023-03-31T05:57:08.2110894Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss] PASSED [ 88%] 2023-03-31T05:57:08.2111401Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nll_loss_nd] PASSED [ 88%] 2023-03-31T05:57:08.2111884Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nonzero_numpy] XFAIL [ 88%] 2023-03-31T05:57:08.2112370Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm.names_ScalarOpt_dim] XFAIL [ 88%] 2023-03-31T05:57:08.2112888Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm.names_ScalarOpt_dim_dtype] XFAIL [ 88%] 2023-03-31T05:57:08.2113394Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::norm_except_dim] XFAIL [ 88%] 2023-03-31T05:57:08.2113889Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal.Scalar] PASSED [ 88%] 2023-03-31T05:57:08.2114422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal.Tensor] PASSED [ 88%] 2023-03-31T05:57:08.2114912Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal_.Scalar] XFAIL [ 88%] 2023-03-31T05:57:08.2115392Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::not_equal_.Tensor] XFAIL [ 89%] 2023-03-31T05:57:08.2115883Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nuclear_norm.dim] PASSED [ 89%] 2023-03-31T05:57:08.2116371Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::nuclear_norm] PASSED [ 89%] 2023-03-31T05:57:08.2116836Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::numpy_T] PASSED [ 89%] 2023-03-31T05:57:08.2117313Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::one_hot] XFAIL [ 89%] 2023-03-31T05:57:08.2117783Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::orgqr] PASSED [ 89%] 2023-03-31T05:57:08.2118247Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::outer] PASSED [ 89%] 2023-03-31T05:57:08.2118734Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::output_nr] XFAIL [ 89%] 2023-03-31T05:57:08.2119242Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pad] PASSED [ 89%] 2023-03-31T05:57:08.2119716Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pad_sequence] XFAIL [ 89%] 2023-03-31T05:57:08.2120208Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pairwise_distance] PASSED [ 89%] 2023-03-31T05:57:08.2120678Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pdist] XFAIL [ 89%] 2023-03-31T05:57:08.2121155Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pin_memory] XFAIL [ 89%] 2023-03-31T05:57:08.2121634Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::pinverse] PASSED [ 89%] 2023-03-31T05:57:08.2122123Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::poisson_nll_loss] PASSED [ 89%] 2023-03-31T05:57:08.2122590Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::positive] PASSED [ 89%] 2023-03-31T05:57:08.2123200Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::prelu] PASSED [ 89%] 2023-03-31T05:57:08.2123735Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::promote_types] XFAIL [ 90%] 2023-03-31T05:57:08.2124212Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::qr.Q] XFAIL [ 90%] 2023-03-31T05:57:08.2124670Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::qr] PASSED [ 90%] 2023-03-31T05:57:08.2125151Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::quantile.scalar] XFAIL [ 90%] 2023-03-31T05:57:08.2125637Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::quantile] XFAIL [ 90%] 2023-03-31T05:57:08.2126108Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::ravel] PASSED [ 90%] 2023-03-31T05:57:08.2126605Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::real] XFAIL [ 90%] 2023-03-31T05:57:08.2127080Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::refine_names] XFAIL [ 90%] 2023-03-31T05:57:08.2127549Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::relu6] PASSED [ 90%] 2023-03-31T05:57:08.2128019Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::relu6_] PASSED [ 90%] 2023-03-31T05:57:08.2128477Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rename] XFAIL [ 90%] 2023-03-31T05:57:08.2128943Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rename_] XFAIL [ 90%] 2023-03-31T05:57:08.2129450Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::repeat_interleave.self_Tensor] PASSED [ 90%] 2023-03-31T05:57:08.2129978Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::repeat_interleave.self_int] PASSED [ 90%] 2023-03-31T05:57:08.2130467Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::requires_grad_] XFAIL [ 90%] 2023-03-31T05:57:08.2130986Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::reshape] PASSED [ 90%] 2023-03-31T05:57:08.2131511Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::reshape_as] PASSED [ 90%] 2023-03-31T05:57:08.2131993Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::resolve_conj] PASSED [ 90%] 2023-03-31T05:57:08.2132470Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::resolve_neg] PASSED [ 91%] 2023-03-31T05:57:08.2132962Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar] PASSED [ 91%] 2023-03-31T05:57:08.2133477Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar_Scalar] PASSED [ 91%] 2023-03-31T05:57:08.2133994Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Scalar_Tensor] PASSED [ 91%] 2023-03-31T05:57:08.2134492Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::result_type.Tensor] PASSED [ 91%] 2023-03-31T05:57:08.2134982Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::retain_grad] XFAIL [ 91%] 2023-03-31T05:57:08.2135492Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::retains_grad] XFAIL [ 91%] 2023-03-31T05:57:08.2135975Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu.data] XFAIL [ 91%] 2023-03-31T05:57:08.2136448Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu.input] XFAIL [ 91%] 2023-03-31T05:57:08.2136933Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_relu_cell] XFAIL [ 91%] 2023-03-31T05:57:08.2137418Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh.data] XFAIL [ 91%] 2023-03-31T05:57:08.2137897Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh.input] XFAIL [ 91%] 2023-03-31T05:57:08.2138414Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rnn_tanh_cell] XFAIL [ 91%] 2023-03-31T05:57:08.2138877Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::row_stack] PASSED [ 91%] 2023-03-31T05:57:08.2139349Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rrelu] PASSED [ 91%] 2023-03-31T05:57:08.2139826Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::rrelu_] PASSED [ 91%] 2023-03-31T05:57:08.2140329Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::scaled_dot_product_attention] PASSED [ 91%] 2023-03-31T05:57:08.2140815Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::selu] PASSED [ 92%] 2023-03-31T05:57:08.2141288Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::selu_] PASSED [ 92%] 2023-03-31T05:57:08.2141784Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::set_.source_Tensor_storage_offset] XFAIL [ 92%] 2023-03-31T05:57:08.2142281Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::set_data] XFAIL [ 92%] 2023-03-31T05:57:08.2142781Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::silu_backward] XFAIL [ 92%] 2023-03-31T05:57:08.2143381Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::size.int] PASSED [ 92%] 2023-03-31T05:57:08.2143863Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::slogdet] PASSED [ 92%] 2023-03-31T05:57:08.2144346Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::slow_conv3d] XFAIL [ 92%] 2023-03-31T05:57:08.2144808Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::smm] XFAIL [ 92%] 2023-03-31T05:57:08.2145290Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::softmax.int] PASSED [ 92%] 2023-03-31T05:57:08.2145806Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_t.n_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2146069Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_t.x_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2146337Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_u.n_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2146638Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_u.x_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2146897Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_v.n_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2147151Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_v.x_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2147413Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_w.n_scalar] XFAIL [ 92%] 2023-03-31T05:57:08.2147671Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_chebyshev_polynomial_w.x_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2147947Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_digamma] PASSED [ 93%] 2023-03-31T05:57:08.2148188Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erf] PASSED [ 93%] 2023-03-31T05:57:08.2148426Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erfc] PASSED [ 93%] 2023-03-31T05:57:08.2148670Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_erfinv] PASSED [ 93%] 2023-03-31T05:57:08.2148912Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_exp2] PASSED [ 93%] 2023-03-31T05:57:08.2149157Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_expit] PASSED [ 93%] 2023-03-31T05:57:08.2149399Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_expm1] PASSED [ 93%] 2023-03-31T05:57:08.2149645Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammainc] PASSED [ 93%] 2023-03-31T05:57:08.2149877Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammaincc] PASSED [ 93%] 2023-03-31T05:57:08.2150149Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_gammaln] PASSED [ 93%] 2023-03-31T05:57:08.2150453Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_h.n_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2150712Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_h.x_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2150982Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_he.n_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2151244Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_hermite_polynomial_he.x_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2151483Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_i0] PASSED [ 93%] 2023-03-31T05:57:08.2151744Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_laguerre_polynomial_l.n_scalar] XFAIL [ 93%] 2023-03-31T05:57:08.2152000Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_laguerre_polynomial_l.x_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2152257Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_legendre_polynomial_p.n_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2152558Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_legendre_polynomial_p.x_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2152799Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_log1p] PASSED [ 94%] 2023-03-31T05:57:08.2153033Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_log_softmax] PASSED [ 94%] 2023-03-31T05:57:08.2153273Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_logit] PASSED [ 94%] 2023-03-31T05:57:08.2153516Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_logsumexp] PASSED [ 94%] 2023-03-31T05:57:08.2153801Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_multigammaln] PASSED [ 94%] 2023-03-31T05:57:08.2154042Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_ndtr] PASSED [ 94%] 2023-03-31T05:57:08.2154284Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_polygamma] PASSED [ 94%] 2023-03-31T05:57:08.2154527Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_psi] PASSED [ 94%] 2023-03-31T05:57:08.2154768Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_round] PASSED [ 94%] 2023-03-31T05:57:08.2155047Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_t.n_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2155327Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_t.x_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2155606Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_u.n_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2155894Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_u.x_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2156197Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_v.n_scalar] XFAIL [ 94%] 2023-03-31T05:57:08.2156465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_v.x_scalar] XFAIL [ 95%] 2023-03-31T05:57:08.2156741Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_w.n_scalar] XFAIL [ 95%] 2023-03-31T05:57:08.2157014Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_shifted_chebyshev_polynomial_w.x_scalar] XFAIL [ 95%] 2023-03-31T05:57:08.2157255Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_sinc] PASSED [ 95%] 2023-03-31T05:57:08.2157500Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_softmax] PASSED [ 95%] 2023-03-31T05:57:08.2157753Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy.other_scalar] XFAIL [ 95%] 2023-03-31T05:57:08.2158003Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy.self_scalar] XFAIL [ 95%] 2023-03-31T05:57:08.2158277Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::special_xlogy] XFAIL [ 95%] 2023-03-31T05:57:08.2158518Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::split.sizes] PASSED [ 95%] 2023-03-31T05:57:08.2158737Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::square] PASSED [ 95%] 2023-03-31T05:57:08.2158964Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::square_] XFAIL [ 95%] 2023-03-31T05:57:08.2159194Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::sspaddmm] XFAIL [ 95%] 2023-03-31T05:57:08.2159440Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.correction_names] XFAIL [ 95%] 2023-03-31T05:57:08.2159708Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.dim] PASSED [ 95%] 2023-03-31T05:57:08.2159943Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std.names_dim] XFAIL [ 95%] 2023-03-31T05:57:08.2160168Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std] PASSED [ 95%] 2023-03-31T05:57:08.2160422Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.correction_names] XFAIL [ 95%] 2023-03-31T05:57:08.2160661Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.dim] PASSED [ 95%] 2023-03-31T05:57:08.2160906Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean.names_dim] XFAIL [ 96%] 2023-03-31T05:57:08.2161142Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::std_mean] PASSED [ 96%] 2023-03-31T05:57:08.2161365Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stft.center] XFAIL [ 96%] 2023-03-31T05:57:08.2161591Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stft] XFAIL [ 96%] 2023-03-31T05:57:08.2161853Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::stride.int] XFAIL [ 96%] 2023-03-31T05:57:08.2162123Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract.Scalar] XFAIL [ 96%] 2023-03-31T05:57:08.2162365Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract.Tensor] PASSED [ 96%] 2023-03-31T05:57:08.2162607Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract_.Scalar] XFAIL [ 96%] 2023-03-31T05:57:08.2162843Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::subtract_.Tensor] XFAIL [ 96%] 2023-03-31T05:57:08.2163364Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::sum_to_size] PASSED [ 96%] 2023-03-31T05:57:08.2163629Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::svd.U] XFAIL [ 96%] 2023-03-31T05:57:08.2163859Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::svd] PASSED [ 96%] 2023-03-31T05:57:08.2164096Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapaxes] PASSED [ 96%] 2023-03-31T05:57:08.2164322Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapaxes_] PASSED [ 96%] 2023-03-31T05:57:08.2164623Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapdims] PASSED [ 96%] 2023-03-31T05:57:08.2164855Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::swapdims_] PASSED [ 96%] 2023-03-31T05:57:08.2165095Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::take_along_dim] PASSED [ 96%] 2023-03-31T05:57:08.2165344Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.indices] XFAIL [ 97%] 2023-03-31T05:57:08.2165593Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.sections] XFAIL [ 97%] 2023-03-31T05:57:08.2165899Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensor_split.tensor_indices_or_sections] XFAIL [ 97%] 2023-03-31T05:57:08.2166136Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tensordot] PASSED [ 97%] 2023-03-31T05:57:08.2166370Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::thnn_conv2d] XFAIL [ 97%] 2023-03-31T05:57:08.2166601Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::tile] PASSED [ 97%] 2023-03-31T05:57:08.2166822Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.device] XFAIL [ 97%] 2023-03-31T05:57:08.2167054Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.dtype] XFAIL [ 97%] 2023-03-31T05:57:08.2167294Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.dtype_layout] XFAIL [ 97%] 2023-03-31T05:57:08.2167527Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to.other] XFAIL [ 97%] 2023-03-31T05:57:08.2167756Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_dense] XFAIL [ 97%] 2023-03-31T05:57:08.2168033Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_dense_backward] XFAIL [ 97%] 2023-03-31T05:57:08.2168308Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::to_mkldnn_backward] XFAIL [ 97%] 2023-03-31T05:57:08.2168545Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trace_backward] XFAIL [ 97%] 2023-03-31T05:57:08.2168785Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapezoid.dx] PASSED [ 97%] 2023-03-31T05:57:08.2169025Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapezoid.x] PASSED [ 97%] 2023-03-31T05:57:08.2169257Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapz.dx] PASSED [ 97%] 2023-03-31T05:57:08.2169479Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::trapz.x] PASSED [ 98%] 2023-03-31T05:57:08.2169725Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::triplet_margin_loss] XFAIL [ 98%] 2023-03-31T05:57:08.2169967Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide.Scalar] PASSED [ 98%] 2023-03-31T05:57:08.2170210Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide.Tensor] PASSED [ 98%] 2023-03-31T05:57:08.2170493Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide_.Scalar] PASSED [ 98%] 2023-03-31T05:57:08.2170737Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::true_divide_.Tensor] PASSED [ 98%] 2023-03-31T05:57:08.2170969Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::type_as] PASSED [ 98%] 2023-03-31T05:57:08.2171215Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unflatten.int] PASSED [ 98%] 2023-03-31T05:57:08.2171465Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unflatten_dense_tensors] XFAIL [ 98%] 2023-03-31T05:57:08.2171736Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::unsafe_chunk] XFAIL [ 98%] 2023-03-31T05:57:08.2171986Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_bicubic2d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2172223Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_bilinear2d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2172469Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_linear1d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2172716Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest1d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2172957Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest2d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2173195Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_nearest3d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2173444Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::upsample_trilinear3d.vec] XFAIL [ 98%] 2023-03-31T05:57:08.2173706Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::value_selecting_reduction_backward] PASSED [ 99%] 2023-03-31T05:57:08.2173964Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vander] XFAIL [ 99%] 2023-03-31T05:57:08.2174234Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.correction_names] XFAIL [ 99%] 2023-03-31T05:57:08.2174466Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.dim] PASSED [ 99%] 2023-03-31T05:57:08.2174701Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var.names_dim] XFAIL [ 99%] 2023-03-31T05:57:08.2174917Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var] PASSED [ 99%] 2023-03-31T05:57:08.2175167Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.correction_names] XFAIL [ 99%] 2023-03-31T05:57:08.2175410Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.dim] PASSED [ 99%] 2023-03-31T05:57:08.2175652Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean.names_dim] XFAIL [ 99%] 2023-03-31T05:57:08.2175881Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::var_mean] PASSED [ 99%] 2023-03-31T05:57:08.2176110Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::view_as] PASSED [ 99%] 2023-03-31T05:57:08.2176377Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vsplit.array] PASSED [ 99%] 2023-03-31T05:57:08.2176615Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vsplit.int] PASSED [ 99%] 2023-03-31T05:57:08.2176849Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::vstack] PASSED [ 99%] 2023-03-31T05:57:08.2177099Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.ScalarOther] PASSED [ 99%] 2023-03-31T05:57:08.2177331Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.ScalarSelf] PASSED [ 99%] 2023-03-31T05:57:08.2177608Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where.Scalar] PASSED [ 99%] 2023-03-31T05:57:08.2177836Z functorch/test_vmap_registrations.py::TestFunctorchDispatcher::test_unimplemented_batched_registrations_[aten::where] XFAIL [100%] 2023-03-31T05:57:08.2177847Z 2023-03-31T05:57:08.2178408Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap_registrations/functorch.test_vmap_registrations-1e483d8efa4950e7.xml - 2023-03-31T05:57:08.2178535Z ====================== 1450 passed, 270 xfailed in 2.83s ======================= 2023-03-31T05:57:08.2178720Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:08.2178726Z 2023-03-31T05:57:08.2179088Z ##[endgroup] 2023-03-31T05:57:08.2179469Z FINISHED PRINTING LOG FILE of functorch/test_vmap_registrations (/var/lib/jenkins/workspace/test/test-reports/functorch-test_vmap_registrations_w8ccd5em.log) 2023-03-31T05:57:08.2179478Z 2023-03-31T05:57:10.1131268Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:10.1159057Z Ignoring disabled issues: [] 2023-03-31T05:57:10.1331492Z Running test_module_init ... [2023-03-31 05:57:10.132745] 2023-03-31T05:57:10.1333191Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_module_init.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:10.133074] 2023-03-31T05:57:15.0156883Z 2023-03-31T05:57:15.0157730Z Expand the folded group to see the log file of test_module_init 2023-03-31T05:57:15.0159090Z ##[group]PRINTING LOG FILE of test_module_init (/var/lib/jenkins/workspace/test/test-reports/test_module_init_lmqnl5dh.log) 2023-03-31T05:57:15.0160216Z Test results will be stored in test-reports/python-pytest/test_module_init/test_module_init-855433dfa1eb4a15.xml 2023-03-31T05:57:15.0160790Z ============================= test session starts ============================== 2023-03-31T05:57:15.0161447Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:15.0161905Z cachedir: .pytest_cache 2023-03-31T05:57:15.0162621Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:15.0163431Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:15.0164181Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:15.0164736Z collecting ... collected 378 items 2023-03-31T05:57:15.0202661Z Running 378 items in this shard: test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveLogSoftmaxWithLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveLogSoftmaxWithLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AlphaDropout_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AlphaDropout_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_AvgPool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_BCELoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_BCELoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_BCEWithLogitsLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_BCEWithLogitsLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Bilinear_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Bilinear_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CELU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CELU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CTCLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CTCLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ChannelShuffle_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ChannelShuffle_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Conv3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CosineEmbeddingLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CosineEmbeddingLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CosineSimilarity_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CosineSimilarity_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CrossEntropyLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CrossEntropyLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_CrossMapLRN2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_CrossMapLRN2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Dropout_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ELU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ELU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_EmbeddingBag_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_EmbeddingBag_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Embedding_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Embedding_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_FeatureAlphaDropout_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_FeatureAlphaDropout_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Flatten_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Flatten_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Fold_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Fold_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GELU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GELU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GRUCell_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GRUCell_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GRU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GRU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GaussianNLLLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GaussianNLLLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_GroupNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_GroupNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardshrink_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardshrink_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardsigmoid_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardsigmoid_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardswish_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardswish_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardtanh_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Hardtanh_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_HingeEmbeddingLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_HingeEmbeddingLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_HuberLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_HuberLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Identity_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Identity_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_KLDivLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_KLDivLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_L1Loss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_L1Loss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LPPool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LPPool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LPPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LPPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LSTMCell_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LSTMCell_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LSTM_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LSTM_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LayerNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LayerNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConv3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyLinear_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LazyLinear_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LeakyReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LeakyReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Linear_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Linear_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LocalResponseNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LocalResponseNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LogSigmoid_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LogSigmoid_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_LogSoftmax_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_LogSoftmax_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MSELoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MSELoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MarginRankingLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MarginRankingLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxPool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Mish_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Mish_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ModuleDict_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ModuleDict_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ModuleList_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ModuleList_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelMarginLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelMarginLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelSoftMarginLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelSoftMarginLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiMarginLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiMarginLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiheadAttention_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_MultiheadAttention_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_NLLLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_NLLLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_PReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_PReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_PairwiseDistance_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_PairwiseDistance_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ParameterDict_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ParameterDict_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ParameterList_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ParameterList_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_PixelShuffle_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_PixelShuffle_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_PixelUnshuffle_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_PixelUnshuffle_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_PoissonNLLLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_PoissonNLLLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNBase_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNBase_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNCellBase_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNCellBase_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNCell_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_RNNCell_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_RNN_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_RNN_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_RReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_RReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReLU6_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReLU6_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_SELU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_SELU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Sequential_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Sequential_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_SiLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_SiLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Sigmoid_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Sigmoid_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_SmoothL1Loss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_SmoothL1Loss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_SoftMarginLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_SoftMarginLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmax2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmax2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmax_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmax_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmin_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softmin_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softplus_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softplus_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softshrink_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softshrink_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Softsign_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Softsign_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_SyncBatchNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_SyncBatchNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Tanh_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Tanh_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Tanhshrink_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Tanhshrink_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Threshold_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Threshold_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoderLayer_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoderLayer_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoder_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoder_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoderLayer_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoderLayer_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoder_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoder_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Transformer_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Transformer_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginWithDistanceLoss_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginWithDistanceLoss_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Unflatten_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Unflatten_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Unfold_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Unfold_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_Upsample_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_Upsample_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingBilinear2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingBilinear2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingNearest2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingNearest2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_Conv3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_EmbeddingBag_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_EmbeddingBag_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_Embedding_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_Embedding_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_qat_Linear_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_qat_Linear_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantizable_LSTMCell_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantizable_LSTMCell_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantizable_LSTM_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantizable_LSTM_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantizable_MultiheadAttention_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantizable_MultiheadAttention_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Conv3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_DeQuantize_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_DeQuantize_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Dropout_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Dropout_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_ELU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_ELU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_FXFloatFunctional_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_FXFloatFunctional_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_FloatFunctional_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_FloatFunctional_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_GroupNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_GroupNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Hardswish_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Hardswish_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm1d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm1d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm2d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm2d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm3d_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm3d_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_LayerNorm_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_LayerNorm_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_LeakyReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_LeakyReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Linear_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Linear_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_PReLU_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_PReLU_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_QFunctional_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_QFunctional_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Quantize_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Quantize_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_ReLU6_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_ReLU6_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Sigmoid_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Sigmoid_cpu_float64, test/test_module_init.py::TestModuleInitCPU::test_quantized_Softmax_cpu_float32, test/test_module_init.py::TestModuleInitCPU::test_quantized_Softmax_cpu_float64 2023-03-31T05:57:15.0245486Z 2023-03-31T05:57:15.0245703Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool1d_cpu_float32 PASSED [ 0%] 2023-03-31T05:57:15.0246101Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool1d_cpu_float64 PASSED [ 0%] 2023-03-31T05:57:15.0246484Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool2d_cpu_float32 PASSED [ 0%] 2023-03-31T05:57:15.0246853Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool2d_cpu_float64 PASSED [ 1%] 2023-03-31T05:57:15.0247210Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool3d_cpu_float32 PASSED [ 1%] 2023-03-31T05:57:15.0247585Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveAvgPool3d_cpu_float64 PASSED [ 1%] 2023-03-31T05:57:15.0247978Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveLogSoftmaxWithLoss_cpu_float32 PASSED [ 1%] 2023-03-31T05:57:15.0248389Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveLogSoftmaxWithLoss_cpu_float64 PASSED [ 2%] 2023-03-31T05:57:15.0248776Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool1d_cpu_float32 PASSED [ 2%] 2023-03-31T05:57:15.0249146Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool1d_cpu_float64 PASSED [ 2%] 2023-03-31T05:57:15.0249513Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool2d_cpu_float32 PASSED [ 2%] 2023-03-31T05:57:15.0249862Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool2d_cpu_float64 PASSED [ 3%] 2023-03-31T05:57:15.0250320Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool3d_cpu_float32 PASSED [ 3%] 2023-03-31T05:57:15.0250731Z test_module_init.py::TestModuleInitCPU::test_nn_AdaptiveMaxPool3d_cpu_float64 PASSED [ 3%] 2023-03-31T05:57:15.0251096Z test_module_init.py::TestModuleInitCPU::test_nn_AlphaDropout_cpu_float32 PASSED [ 3%] 2023-03-31T05:57:15.0251439Z test_module_init.py::TestModuleInitCPU::test_nn_AlphaDropout_cpu_float64 PASSED [ 4%] 2023-03-31T05:57:15.0251796Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool1d_cpu_float32 PASSED [ 4%] 2023-03-31T05:57:15.0252138Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool1d_cpu_float64 PASSED [ 4%] 2023-03-31T05:57:15.0252469Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool2d_cpu_float32 PASSED [ 5%] 2023-03-31T05:57:15.0252800Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool2d_cpu_float64 PASSED [ 5%] 2023-03-31T05:57:15.0253137Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool3d_cpu_float32 PASSED [ 5%] 2023-03-31T05:57:15.0253484Z test_module_init.py::TestModuleInitCPU::test_nn_AvgPool3d_cpu_float64 PASSED [ 5%] 2023-03-31T05:57:15.0253815Z test_module_init.py::TestModuleInitCPU::test_nn_BCELoss_cpu_float32 PASSED [ 6%] 2023-03-31T05:57:15.0254156Z test_module_init.py::TestModuleInitCPU::test_nn_BCELoss_cpu_float64 PASSED [ 6%] 2023-03-31T05:57:15.0254510Z test_module_init.py::TestModuleInitCPU::test_nn_BCEWithLogitsLoss_cpu_float32 PASSED [ 6%] 2023-03-31T05:57:15.0254869Z test_module_init.py::TestModuleInitCPU::test_nn_BCEWithLogitsLoss_cpu_float64 PASSED [ 6%] 2023-03-31T05:57:15.0255282Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm1d_cpu_float32 PASSED [ 7%] 2023-03-31T05:57:15.0255633Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm1d_cpu_float64 PASSED [ 7%] 2023-03-31T05:57:15.0255973Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm2d_cpu_float32 PASSED [ 7%] 2023-03-31T05:57:15.0256297Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm2d_cpu_float64 PASSED [ 7%] 2023-03-31T05:57:15.0256649Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm3d_cpu_float32 PASSED [ 8%] 2023-03-31T05:57:15.0256992Z test_module_init.py::TestModuleInitCPU::test_nn_BatchNorm3d_cpu_float64 PASSED [ 8%] 2023-03-31T05:57:15.0257331Z test_module_init.py::TestModuleInitCPU::test_nn_Bilinear_cpu_float32 PASSED [ 8%] 2023-03-31T05:57:15.0257659Z test_module_init.py::TestModuleInitCPU::test_nn_Bilinear_cpu_float64 PASSED [ 8%] 2023-03-31T05:57:15.0258057Z test_module_init.py::TestModuleInitCPU::test_nn_CELU_cpu_float32 PASSED [ 9%] 2023-03-31T05:57:15.0258393Z test_module_init.py::TestModuleInitCPU::test_nn_CELU_cpu_float64 PASSED [ 9%] 2023-03-31T05:57:15.0258719Z test_module_init.py::TestModuleInitCPU::test_nn_CTCLoss_cpu_float32 PASSED [ 9%] 2023-03-31T05:57:15.0259058Z test_module_init.py::TestModuleInitCPU::test_nn_CTCLoss_cpu_float64 PASSED [ 10%] 2023-03-31T05:57:15.0259408Z test_module_init.py::TestModuleInitCPU::test_nn_ChannelShuffle_cpu_float32 PASSED [ 10%] 2023-03-31T05:57:15.0259774Z test_module_init.py::TestModuleInitCPU::test_nn_ChannelShuffle_cpu_float64 PASSED [ 10%] 2023-03-31T05:57:15.0260123Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad1d_cpu_float32 PASSED [ 10%] 2023-03-31T05:57:15.0260480Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad1d_cpu_float64 PASSED [ 11%] 2023-03-31T05:57:15.0260832Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad2d_cpu_float32 PASSED [ 11%] 2023-03-31T05:57:15.0261177Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad2d_cpu_float64 PASSED [ 11%] 2023-03-31T05:57:15.0261532Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad3d_cpu_float32 PASSED [ 11%] 2023-03-31T05:57:15.0261883Z test_module_init.py::TestModuleInitCPU::test_nn_ConstantPad3d_cpu_float64 PASSED [ 12%] 2023-03-31T05:57:15.0262227Z test_module_init.py::TestModuleInitCPU::test_nn_Conv1d_cpu_float32 PASSED [ 12%] 2023-03-31T05:57:15.0262550Z test_module_init.py::TestModuleInitCPU::test_nn_Conv1d_cpu_float64 PASSED [ 12%] 2023-03-31T05:57:15.0262992Z test_module_init.py::TestModuleInitCPU::test_nn_Conv2d_cpu_float32 PASSED [ 12%] 2023-03-31T05:57:15.0263360Z test_module_init.py::TestModuleInitCPU::test_nn_Conv2d_cpu_float64 PASSED [ 13%] 2023-03-31T05:57:15.0263682Z test_module_init.py::TestModuleInitCPU::test_nn_Conv3d_cpu_float32 PASSED [ 13%] 2023-03-31T05:57:15.0264011Z test_module_init.py::TestModuleInitCPU::test_nn_Conv3d_cpu_float64 PASSED [ 13%] 2023-03-31T05:57:15.0264365Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose1d_cpu_float32 PASSED [ 14%] 2023-03-31T05:57:15.0264726Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose1d_cpu_float64 PASSED [ 14%] 2023-03-31T05:57:15.0265077Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose2d_cpu_float32 PASSED [ 14%] 2023-03-31T05:57:15.0265436Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose2d_cpu_float64 PASSED [ 14%] 2023-03-31T05:57:15.0265798Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose3d_cpu_float32 PASSED [ 15%] 2023-03-31T05:57:15.0266144Z test_module_init.py::TestModuleInitCPU::test_nn_ConvTranspose3d_cpu_float64 PASSED [ 15%] 2023-03-31T05:57:15.0266515Z test_module_init.py::TestModuleInitCPU::test_nn_CosineEmbeddingLoss_cpu_float32 PASSED [ 15%] 2023-03-31T05:57:15.0266894Z test_module_init.py::TestModuleInitCPU::test_nn_CosineEmbeddingLoss_cpu_float64 PASSED [ 15%] 2023-03-31T05:57:15.0267265Z test_module_init.py::TestModuleInitCPU::test_nn_CosineSimilarity_cpu_float32 PASSED [ 16%] 2023-03-31T05:57:15.0267618Z test_module_init.py::TestModuleInitCPU::test_nn_CosineSimilarity_cpu_float64 PASSED [ 16%] 2023-03-31T05:57:15.0268018Z test_module_init.py::TestModuleInitCPU::test_nn_CrossEntropyLoss_cpu_float32 PASSED [ 16%] 2023-03-31T05:57:15.0268384Z test_module_init.py::TestModuleInitCPU::test_nn_CrossEntropyLoss_cpu_float64 PASSED [ 16%] 2023-03-31T05:57:15.0268728Z test_module_init.py::TestModuleInitCPU::test_nn_CrossMapLRN2d_cpu_float32 PASSED [ 17%] 2023-03-31T05:57:15.0269089Z test_module_init.py::TestModuleInitCPU::test_nn_CrossMapLRN2d_cpu_float64 PASSED [ 17%] 2023-03-31T05:57:15.0269447Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout1d_cpu_float32 PASSED [ 17%] 2023-03-31T05:57:15.0269790Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout1d_cpu_float64 PASSED [ 17%] 2023-03-31T05:57:15.0270122Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout2d_cpu_float32 PASSED [ 18%] 2023-03-31T05:57:15.0270461Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout2d_cpu_float64 PASSED [ 18%] 2023-03-31T05:57:15.0270835Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout3d_cpu_float32 PASSED [ 18%] 2023-03-31T05:57:15.0271175Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout3d_cpu_float64 PASSED [ 19%] 2023-03-31T05:57:15.0271501Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout_cpu_float32 PASSED [ 19%] 2023-03-31T05:57:15.0271843Z test_module_init.py::TestModuleInitCPU::test_nn_Dropout_cpu_float64 PASSED [ 19%] 2023-03-31T05:57:15.0272183Z test_module_init.py::TestModuleInitCPU::test_nn_ELU_cpu_float32 PASSED [ 19%] 2023-03-31T05:57:15.0272502Z test_module_init.py::TestModuleInitCPU::test_nn_ELU_cpu_float64 PASSED [ 20%] 2023-03-31T05:57:15.0272846Z test_module_init.py::TestModuleInitCPU::test_nn_EmbeddingBag_cpu_float32 PASSED [ 20%] 2023-03-31T05:57:15.0273206Z test_module_init.py::TestModuleInitCPU::test_nn_EmbeddingBag_cpu_float64 PASSED [ 20%] 2023-03-31T05:57:15.0273555Z test_module_init.py::TestModuleInitCPU::test_nn_Embedding_cpu_float32 PASSED [ 20%] 2023-03-31T05:57:15.0273888Z test_module_init.py::TestModuleInitCPU::test_nn_Embedding_cpu_float64 PASSED [ 21%] 2023-03-31T05:57:15.0274253Z test_module_init.py::TestModuleInitCPU::test_nn_FeatureAlphaDropout_cpu_float32 PASSED [ 21%] 2023-03-31T05:57:15.0274634Z test_module_init.py::TestModuleInitCPU::test_nn_FeatureAlphaDropout_cpu_float64 PASSED [ 21%] 2023-03-31T05:57:15.0274978Z test_module_init.py::TestModuleInitCPU::test_nn_Flatten_cpu_float32 PASSED [ 21%] 2023-03-31T05:57:15.0275351Z test_module_init.py::TestModuleInitCPU::test_nn_Flatten_cpu_float64 PASSED [ 22%] 2023-03-31T05:57:15.0275720Z test_module_init.py::TestModuleInitCPU::test_nn_Fold_cpu_float32 PASSED [ 22%] 2023-03-31T05:57:15.0276052Z test_module_init.py::TestModuleInitCPU::test_nn_Fold_cpu_float64 PASSED [ 22%] 2023-03-31T05:57:15.0276396Z test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool2d_cpu_float32 PASSED [ 23%] 2023-03-31T05:57:15.0276772Z test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool2d_cpu_float64 PASSED [ 23%] 2023-03-31T05:57:15.0277148Z test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool3d_cpu_float32 PASSED [ 23%] 2023-03-31T05:57:15.0277508Z test_module_init.py::TestModuleInitCPU::test_nn_FractionalMaxPool3d_cpu_float64 PASSED [ 23%] 2023-03-31T05:57:15.0277858Z test_module_init.py::TestModuleInitCPU::test_nn_GELU_cpu_float32 PASSED [ 24%] 2023-03-31T05:57:15.0278190Z test_module_init.py::TestModuleInitCPU::test_nn_GELU_cpu_float64 PASSED [ 24%] 2023-03-31T05:57:15.0278523Z test_module_init.py::TestModuleInitCPU::test_nn_GLU_cpu_float32 PASSED [ 24%] 2023-03-31T05:57:15.0278843Z test_module_init.py::TestModuleInitCPU::test_nn_GLU_cpu_float64 PASSED [ 24%] 2023-03-31T05:57:15.0279177Z test_module_init.py::TestModuleInitCPU::test_nn_GRUCell_cpu_float32 PASSED [ 25%] 2023-03-31T05:57:15.0279516Z test_module_init.py::TestModuleInitCPU::test_nn_GRUCell_cpu_float64 PASSED [ 25%] 2023-03-31T05:57:15.0279838Z test_module_init.py::TestModuleInitCPU::test_nn_GRU_cpu_float32 PASSED [ 25%] 2023-03-31T05:57:15.0280169Z test_module_init.py::TestModuleInitCPU::test_nn_GRU_cpu_float64 PASSED [ 25%] 2023-03-31T05:57:15.0280544Z test_module_init.py::TestModuleInitCPU::test_nn_GaussianNLLLoss_cpu_float32 PASSED [ 26%] 2023-03-31T05:57:15.0280913Z test_module_init.py::TestModuleInitCPU::test_nn_GaussianNLLLoss_cpu_float64 PASSED [ 26%] 2023-03-31T05:57:15.0281255Z test_module_init.py::TestModuleInitCPU::test_nn_GroupNorm_cpu_float32 PASSED [ 26%] 2023-03-31T05:57:15.0281601Z test_module_init.py::TestModuleInitCPU::test_nn_GroupNorm_cpu_float64 PASSED [ 26%] 2023-03-31T05:57:15.0281951Z test_module_init.py::TestModuleInitCPU::test_nn_Hardshrink_cpu_float32 PASSED [ 27%] 2023-03-31T05:57:15.0282285Z test_module_init.py::TestModuleInitCPU::test_nn_Hardshrink_cpu_float64 PASSED [ 27%] 2023-03-31T05:57:15.0282633Z test_module_init.py::TestModuleInitCPU::test_nn_Hardsigmoid_cpu_float32 PASSED [ 27%] 2023-03-31T05:57:15.0282980Z test_module_init.py::TestModuleInitCPU::test_nn_Hardsigmoid_cpu_float64 PASSED [ 28%] 2023-03-31T05:57:15.0283581Z test_module_init.py::TestModuleInitCPU::test_nn_Hardswish_cpu_float32 PASSED [ 28%] 2023-03-31T05:57:15.0283914Z test_module_init.py::TestModuleInitCPU::test_nn_Hardswish_cpu_float64 PASSED [ 28%] 2023-03-31T05:57:15.0284258Z test_module_init.py::TestModuleInitCPU::test_nn_Hardtanh_cpu_float32 PASSED [ 28%] 2023-03-31T05:57:15.0284604Z test_module_init.py::TestModuleInitCPU::test_nn_Hardtanh_cpu_float64 PASSED [ 29%] 2023-03-31T05:57:15.0284955Z test_module_init.py::TestModuleInitCPU::test_nn_HingeEmbeddingLoss_cpu_float32 PASSED [ 29%] 2023-03-31T05:57:15.0285336Z test_module_init.py::TestModuleInitCPU::test_nn_HingeEmbeddingLoss_cpu_float64 PASSED [ 29%] 2023-03-31T05:57:15.0285699Z test_module_init.py::TestModuleInitCPU::test_nn_HuberLoss_cpu_float32 PASSED [ 29%] 2023-03-31T05:57:15.0286041Z test_module_init.py::TestModuleInitCPU::test_nn_HuberLoss_cpu_float64 PASSED [ 30%] 2023-03-31T05:57:15.0286374Z test_module_init.py::TestModuleInitCPU::test_nn_Identity_cpu_float32 PASSED [ 30%] 2023-03-31T05:57:15.0286715Z test_module_init.py::TestModuleInitCPU::test_nn_Identity_cpu_float64 PASSED [ 30%] 2023-03-31T05:57:15.0287071Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm1d_cpu_float32 PASSED [ 30%] 2023-03-31T05:57:15.0287417Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm1d_cpu_float64 PASSED [ 31%] 2023-03-31T05:57:15.0287773Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm2d_cpu_float32 PASSED [ 31%] 2023-03-31T05:57:15.0288180Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm2d_cpu_float64 PASSED [ 31%] 2023-03-31T05:57:15.0288573Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm3d_cpu_float32 PASSED [ 32%] 2023-03-31T05:57:15.0288914Z test_module_init.py::TestModuleInitCPU::test_nn_InstanceNorm3d_cpu_float64 PASSED [ 32%] 2023-03-31T05:57:15.0289268Z test_module_init.py::TestModuleInitCPU::test_nn_KLDivLoss_cpu_float32 PASSED [ 32%] 2023-03-31T05:57:15.0289616Z test_module_init.py::TestModuleInitCPU::test_nn_KLDivLoss_cpu_float64 PASSED [ 32%] 2023-03-31T05:57:15.0289958Z test_module_init.py::TestModuleInitCPU::test_nn_L1Loss_cpu_float32 PASSED [ 33%] 2023-03-31T05:57:15.0290283Z test_module_init.py::TestModuleInitCPU::test_nn_L1Loss_cpu_float64 PASSED [ 33%] 2023-03-31T05:57:15.0290621Z test_module_init.py::TestModuleInitCPU::test_nn_LPPool1d_cpu_float32 PASSED [ 33%] 2023-03-31T05:57:15.0290964Z test_module_init.py::TestModuleInitCPU::test_nn_LPPool1d_cpu_float64 PASSED [ 33%] 2023-03-31T05:57:15.0291295Z test_module_init.py::TestModuleInitCPU::test_nn_LPPool2d_cpu_float32 PASSED [ 34%] 2023-03-31T05:57:15.0291633Z test_module_init.py::TestModuleInitCPU::test_nn_LPPool2d_cpu_float64 PASSED [ 34%] 2023-03-31T05:57:15.0291976Z test_module_init.py::TestModuleInitCPU::test_nn_LSTMCell_cpu_float32 PASSED [ 34%] 2023-03-31T05:57:15.0292309Z test_module_init.py::TestModuleInitCPU::test_nn_LSTMCell_cpu_float64 PASSED [ 34%] 2023-03-31T05:57:15.0292632Z test_module_init.py::TestModuleInitCPU::test_nn_LSTM_cpu_float32 PASSED [ 35%] 2023-03-31T05:57:15.0292972Z test_module_init.py::TestModuleInitCPU::test_nn_LSTM_cpu_float64 PASSED [ 35%] 2023-03-31T05:57:15.0293360Z test_module_init.py::TestModuleInitCPU::test_nn_LayerNorm_cpu_float32 PASSED [ 35%] 2023-03-31T05:57:15.0293695Z test_module_init.py::TestModuleInitCPU::test_nn_LayerNorm_cpu_float64 PASSED [ 35%] 2023-03-31T05:57:15.0294048Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm1d_cpu_float32 PASSED [ 36%] 2023-03-31T05:57:15.0294414Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm1d_cpu_float64 PASSED [ 36%] 2023-03-31T05:57:15.0294775Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm2d_cpu_float32 PASSED [ 36%] 2023-03-31T05:57:15.0295122Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm2d_cpu_float64 PASSED [ 37%] 2023-03-31T05:57:15.0295476Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm3d_cpu_float32 PASSED [ 37%] 2023-03-31T05:57:15.0295893Z test_module_init.py::TestModuleInitCPU::test_nn_LazyBatchNorm3d_cpu_float64 PASSED [ 37%] 2023-03-31T05:57:15.0296234Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv1d_cpu_float32 PASSED [ 37%] 2023-03-31T05:57:15.0296582Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv1d_cpu_float64 PASSED [ 38%] 2023-03-31T05:57:15.0296927Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv2d_cpu_float32 PASSED [ 38%] 2023-03-31T05:57:15.0297270Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv2d_cpu_float64 PASSED [ 38%] 2023-03-31T05:57:15.0297597Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv3d_cpu_float32 PASSED [ 38%] 2023-03-31T05:57:15.0297945Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConv3d_cpu_float64 PASSED [ 39%] 2023-03-31T05:57:15.0298309Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose1d_cpu_float32 PASSED [ 39%] 2023-03-31T05:57:15.0298675Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose1d_cpu_float64 PASSED [ 39%] 2023-03-31T05:57:15.0299052Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose2d_cpu_float32 PASSED [ 39%] 2023-03-31T05:57:15.0299428Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose2d_cpu_float64 PASSED [ 40%] 2023-03-31T05:57:15.0299804Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose3d_cpu_float32 PASSED [ 40%] 2023-03-31T05:57:15.0300162Z test_module_init.py::TestModuleInitCPU::test_nn_LazyConvTranspose3d_cpu_float64 PASSED [ 40%] 2023-03-31T05:57:15.0300597Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm1d_cpu_float32 PASSED [ 41%] 2023-03-31T05:57:15.0301000Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm1d_cpu_float64 PASSED [ 41%] 2023-03-31T05:57:15.0301367Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm2d_cpu_float32 PASSED [ 41%] 2023-03-31T05:57:15.0301724Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm2d_cpu_float64 PASSED [ 41%] 2023-03-31T05:57:15.0302093Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm3d_cpu_float32 PASSED [ 42%] 2023-03-31T05:57:15.0302461Z test_module_init.py::TestModuleInitCPU::test_nn_LazyInstanceNorm3d_cpu_float64 PASSED [ 42%] 2023-03-31T05:57:15.0302803Z test_module_init.py::TestModuleInitCPU::test_nn_LazyLinear_cpu_float32 PASSED [ 42%] 2023-03-31T05:57:15.0303241Z test_module_init.py::TestModuleInitCPU::test_nn_LazyLinear_cpu_float64 PASSED [ 42%] 2023-03-31T05:57:15.0303591Z test_module_init.py::TestModuleInitCPU::test_nn_LeakyReLU_cpu_float32 PASSED [ 43%] 2023-03-31T05:57:15.0303939Z test_module_init.py::TestModuleInitCPU::test_nn_LeakyReLU_cpu_float64 PASSED [ 43%] 2023-03-31T05:57:15.0304273Z test_module_init.py::TestModuleInitCPU::test_nn_Linear_cpu_float32 PASSED [ 43%] 2023-03-31T05:57:15.0304612Z test_module_init.py::TestModuleInitCPU::test_nn_Linear_cpu_float64 PASSED [ 43%] 2023-03-31T05:57:15.0304967Z test_module_init.py::TestModuleInitCPU::test_nn_LocalResponseNorm_cpu_float32 PASSED [ 44%] 2023-03-31T05:57:15.0305325Z test_module_init.py::TestModuleInitCPU::test_nn_LocalResponseNorm_cpu_float64 PASSED [ 44%] 2023-03-31T05:57:15.0305690Z test_module_init.py::TestModuleInitCPU::test_nn_LogSigmoid_cpu_float32 PASSED [ 44%] 2023-03-31T05:57:15.0306075Z test_module_init.py::TestModuleInitCPU::test_nn_LogSigmoid_cpu_float64 PASSED [ 44%] 2023-03-31T05:57:15.0306418Z test_module_init.py::TestModuleInitCPU::test_nn_LogSoftmax_cpu_float32 PASSED [ 45%] 2023-03-31T05:57:15.0306745Z test_module_init.py::TestModuleInitCPU::test_nn_LogSoftmax_cpu_float64 PASSED [ 45%] 2023-03-31T05:57:15.0307090Z test_module_init.py::TestModuleInitCPU::test_nn_MSELoss_cpu_float32 PASSED [ 45%] 2023-03-31T05:57:15.0307434Z test_module_init.py::TestModuleInitCPU::test_nn_MSELoss_cpu_float64 PASSED [ 46%] 2023-03-31T05:57:15.0307779Z test_module_init.py::TestModuleInitCPU::test_nn_MarginRankingLoss_cpu_float32 PASSED [ 46%] 2023-03-31T05:57:15.0308147Z test_module_init.py::TestModuleInitCPU::test_nn_MarginRankingLoss_cpu_float64 PASSED [ 46%] 2023-03-31T05:57:15.0308542Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool1d_cpu_float32 PASSED [ 46%] 2023-03-31T05:57:15.0308886Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool1d_cpu_float64 PASSED [ 47%] 2023-03-31T05:57:15.0309214Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool2d_cpu_float32 PASSED [ 47%] 2023-03-31T05:57:15.0309556Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool2d_cpu_float64 PASSED [ 47%] 2023-03-31T05:57:15.0309896Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool3d_cpu_float32 PASSED [ 47%] 2023-03-31T05:57:15.0310225Z test_module_init.py::TestModuleInitCPU::test_nn_MaxPool3d_cpu_float64 PASSED [ 48%] 2023-03-31T05:57:15.0310571Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool1d_cpu_float32 PASSED [ 48%] 2023-03-31T05:57:15.0310920Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool1d_cpu_float64 PASSED [ 48%] 2023-03-31T05:57:15.0311263Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool2d_cpu_float32 PASSED [ 48%] 2023-03-31T05:57:15.0311597Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool2d_cpu_float64 PASSED [ 49%] 2023-03-31T05:57:15.0311944Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool3d_cpu_float32 PASSED [ 49%] 2023-03-31T05:57:15.0312286Z test_module_init.py::TestModuleInitCPU::test_nn_MaxUnpool3d_cpu_float64 PASSED [ 49%] 2023-03-31T05:57:15.0312608Z test_module_init.py::TestModuleInitCPU::test_nn_Mish_cpu_float32 PASSED [ 50%] 2023-03-31T05:57:15.0312943Z test_module_init.py::TestModuleInitCPU::test_nn_Mish_cpu_float64 PASSED [ 50%] 2023-03-31T05:57:15.0313317Z test_module_init.py::TestModuleInitCPU::test_nn_ModuleDict_cpu_float32 PASSED [ 50%] 2023-03-31T05:57:15.0313692Z test_module_init.py::TestModuleInitCPU::test_nn_ModuleDict_cpu_float64 PASSED [ 50%] 2023-03-31T05:57:15.0314022Z test_module_init.py::TestModuleInitCPU::test_nn_ModuleList_cpu_float32 PASSED [ 51%] 2023-03-31T05:57:15.0314363Z test_module_init.py::TestModuleInitCPU::test_nn_ModuleList_cpu_float64 PASSED [ 51%] 2023-03-31T05:57:15.0314733Z test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelMarginLoss_cpu_float32 PASSED [ 51%] 2023-03-31T05:57:15.0315108Z test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelMarginLoss_cpu_float64 PASSED [ 51%] 2023-03-31T05:57:15.0315507Z test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelSoftMarginLoss_cpu_float32 PASSED [ 52%] 2023-03-31T05:57:15.0315909Z test_module_init.py::TestModuleInitCPU::test_nn_MultiLabelSoftMarginLoss_cpu_float64 PASSED [ 52%] 2023-03-31T05:57:15.0316291Z test_module_init.py::TestModuleInitCPU::test_nn_MultiMarginLoss_cpu_float32 PASSED [ 52%] 2023-03-31T05:57:15.0316644Z test_module_init.py::TestModuleInitCPU::test_nn_MultiMarginLoss_cpu_float64 PASSED [ 52%] 2023-03-31T05:57:15.0317013Z test_module_init.py::TestModuleInitCPU::test_nn_MultiheadAttention_cpu_float32 PASSED [ 53%] 2023-03-31T05:57:15.0317388Z test_module_init.py::TestModuleInitCPU::test_nn_MultiheadAttention_cpu_float64 PASSED [ 53%] 2023-03-31T05:57:15.0317745Z test_module_init.py::TestModuleInitCPU::test_nn_NLLLoss_cpu_float32 PASSED [ 53%] 2023-03-31T05:57:15.0318075Z test_module_init.py::TestModuleInitCPU::test_nn_NLLLoss_cpu_float64 PASSED [ 53%] 2023-03-31T05:57:15.0318445Z test_module_init.py::TestModuleInitCPU::test_nn_PReLU_cpu_float32 PASSED [ 54%] 2023-03-31T05:57:15.0318782Z test_module_init.py::TestModuleInitCPU::test_nn_PReLU_cpu_float64 PASSED [ 54%] 2023-03-31T05:57:15.0319123Z test_module_init.py::TestModuleInitCPU::test_nn_PairwiseDistance_cpu_float32 PASSED [ 54%] 2023-03-31T05:57:15.0319488Z test_module_init.py::TestModuleInitCPU::test_nn_PairwiseDistance_cpu_float64 PASSED [ 55%] 2023-03-31T05:57:15.0319849Z test_module_init.py::TestModuleInitCPU::test_nn_ParameterDict_cpu_float32 PASSED [ 55%] 2023-03-31T05:57:15.0320209Z test_module_init.py::TestModuleInitCPU::test_nn_ParameterDict_cpu_float64 PASSED [ 55%] 2023-03-31T05:57:15.0320552Z test_module_init.py::TestModuleInitCPU::test_nn_ParameterList_cpu_float32 PASSED [ 55%] 2023-03-31T05:57:15.0320910Z test_module_init.py::TestModuleInitCPU::test_nn_ParameterList_cpu_float64 PASSED [ 56%] 2023-03-31T05:57:15.0321299Z test_module_init.py::TestModuleInitCPU::test_nn_PixelShuffle_cpu_float32 PASSED [ 56%] 2023-03-31T05:57:15.0321642Z test_module_init.py::TestModuleInitCPU::test_nn_PixelShuffle_cpu_float64 PASSED [ 56%] 2023-03-31T05:57:15.0321998Z test_module_init.py::TestModuleInitCPU::test_nn_PixelUnshuffle_cpu_float32 PASSED [ 56%] 2023-03-31T05:57:15.0322359Z test_module_init.py::TestModuleInitCPU::test_nn_PixelUnshuffle_cpu_float64 PASSED [ 57%] 2023-03-31T05:57:15.0322724Z test_module_init.py::TestModuleInitCPU::test_nn_PoissonNLLLoss_cpu_float32 PASSED [ 57%] 2023-03-31T05:57:15.0323246Z test_module_init.py::TestModuleInitCPU::test_nn_PoissonNLLLoss_cpu_float64 PASSED [ 57%] 2023-03-31T05:57:15.0323604Z test_module_init.py::TestModuleInitCPU::test_nn_RNNBase_cpu_float32 PASSED [ 57%] 2023-03-31T05:57:15.0323942Z test_module_init.py::TestModuleInitCPU::test_nn_RNNBase_cpu_float64 PASSED [ 58%] 2023-03-31T05:57:15.0324274Z test_module_init.py::TestModuleInitCPU::test_nn_RNNCellBase_cpu_float32 PASSED [ 58%] 2023-03-31T05:57:15.0324627Z test_module_init.py::TestModuleInitCPU::test_nn_RNNCellBase_cpu_float64 PASSED [ 58%] 2023-03-31T05:57:15.0324972Z test_module_init.py::TestModuleInitCPU::test_nn_RNNCell_cpu_float32 PASSED [ 58%] 2023-03-31T05:57:15.0325313Z test_module_init.py::TestModuleInitCPU::test_nn_RNNCell_cpu_float64 PASSED [ 59%] 2023-03-31T05:57:15.0325636Z test_module_init.py::TestModuleInitCPU::test_nn_RNN_cpu_float32 PASSED [ 59%] 2023-03-31T05:57:15.0326023Z test_module_init.py::TestModuleInitCPU::test_nn_RNN_cpu_float64 PASSED [ 59%] 2023-03-31T05:57:15.0326392Z test_module_init.py::TestModuleInitCPU::test_nn_RReLU_cpu_float32 PASSED [ 60%] 2023-03-31T05:57:15.0326714Z test_module_init.py::TestModuleInitCPU::test_nn_RReLU_cpu_float64 PASSED [ 60%] 2023-03-31T05:57:15.0327049Z test_module_init.py::TestModuleInitCPU::test_nn_ReLU6_cpu_float32 PASSED [ 60%] 2023-03-31T05:57:15.0327377Z test_module_init.py::TestModuleInitCPU::test_nn_ReLU6_cpu_float64 PASSED [ 60%] 2023-03-31T05:57:15.0327713Z test_module_init.py::TestModuleInitCPU::test_nn_ReLU_cpu_float32 PASSED [ 61%] 2023-03-31T05:57:15.0328031Z test_module_init.py::TestModuleInitCPU::test_nn_ReLU_cpu_float64 PASSED [ 61%] 2023-03-31T05:57:15.0328374Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad1d_cpu_float32 PASSED [ 61%] 2023-03-31T05:57:15.0328741Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad1d_cpu_float64 PASSED [ 61%] 2023-03-31T05:57:15.0329089Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad2d_cpu_float32 PASSED [ 62%] 2023-03-31T05:57:15.0329450Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad2d_cpu_float64 PASSED [ 62%] 2023-03-31T05:57:15.0329806Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad3d_cpu_float32 PASSED [ 62%] 2023-03-31T05:57:15.0330165Z test_module_init.py::TestModuleInitCPU::test_nn_ReflectionPad3d_cpu_float64 PASSED [ 62%] 2023-03-31T05:57:15.0330516Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad1d_cpu_float32 PASSED [ 63%] 2023-03-31T05:57:15.0330886Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad1d_cpu_float64 PASSED [ 63%] 2023-03-31T05:57:15.0331294Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad2d_cpu_float32 PASSED [ 63%] 2023-03-31T05:57:15.0331657Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad2d_cpu_float64 PASSED [ 64%] 2023-03-31T05:57:15.0332004Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad3d_cpu_float32 PASSED [ 64%] 2023-03-31T05:57:15.0332367Z test_module_init.py::TestModuleInitCPU::test_nn_ReplicationPad3d_cpu_float64 PASSED [ 64%] 2023-03-31T05:57:15.0332715Z test_module_init.py::TestModuleInitCPU::test_nn_SELU_cpu_float32 PASSED [ 64%] 2023-03-31T05:57:15.0333034Z test_module_init.py::TestModuleInitCPU::test_nn_SELU_cpu_float64 PASSED [ 65%] 2023-03-31T05:57:15.0333372Z test_module_init.py::TestModuleInitCPU::test_nn_Sequential_cpu_float32 PASSED [ 65%] 2023-03-31T05:57:15.0333764Z test_module_init.py::TestModuleInitCPU::test_nn_Sequential_cpu_float64 PASSED [ 65%] 2023-03-31T05:57:15.0334107Z test_module_init.py::TestModuleInitCPU::test_nn_SiLU_cpu_float32 PASSED [ 65%] 2023-03-31T05:57:15.0334427Z test_module_init.py::TestModuleInitCPU::test_nn_SiLU_cpu_float64 PASSED [ 66%] 2023-03-31T05:57:15.0334760Z test_module_init.py::TestModuleInitCPU::test_nn_Sigmoid_cpu_float32 PASSED [ 66%] 2023-03-31T05:57:15.0335095Z test_module_init.py::TestModuleInitCPU::test_nn_Sigmoid_cpu_float64 PASSED [ 66%] 2023-03-31T05:57:15.0335434Z test_module_init.py::TestModuleInitCPU::test_nn_SmoothL1Loss_cpu_float32 PASSED [ 66%] 2023-03-31T05:57:15.0335789Z test_module_init.py::TestModuleInitCPU::test_nn_SmoothL1Loss_cpu_float64 PASSED [ 67%] 2023-03-31T05:57:15.0336143Z test_module_init.py::TestModuleInitCPU::test_nn_SoftMarginLoss_cpu_float32 PASSED [ 67%] 2023-03-31T05:57:15.0336503Z test_module_init.py::TestModuleInitCPU::test_nn_SoftMarginLoss_cpu_float64 PASSED [ 67%] 2023-03-31T05:57:15.0336846Z test_module_init.py::TestModuleInitCPU::test_nn_Softmax2d_cpu_float32 PASSED [ 67%] 2023-03-31T05:57:15.0337189Z test_module_init.py::TestModuleInitCPU::test_nn_Softmax2d_cpu_float64 PASSED [ 68%] 2023-03-31T05:57:15.0337536Z test_module_init.py::TestModuleInitCPU::test_nn_Softmax_cpu_float32 PASSED [ 68%] 2023-03-31T05:57:15.0337859Z test_module_init.py::TestModuleInitCPU::test_nn_Softmax_cpu_float64 PASSED [ 68%] 2023-03-31T05:57:15.0338196Z test_module_init.py::TestModuleInitCPU::test_nn_Softmin_cpu_float32 PASSED [ 69%] 2023-03-31T05:57:15.0338563Z test_module_init.py::TestModuleInitCPU::test_nn_Softmin_cpu_float64 PASSED [ 69%] 2023-03-31T05:57:15.0338947Z test_module_init.py::TestModuleInitCPU::test_nn_Softplus_cpu_float32 PASSED [ 69%] 2023-03-31T05:57:15.0339277Z test_module_init.py::TestModuleInitCPU::test_nn_Softplus_cpu_float64 PASSED [ 69%] 2023-03-31T05:57:15.0339627Z test_module_init.py::TestModuleInitCPU::test_nn_Softshrink_cpu_float32 PASSED [ 70%] 2023-03-31T05:57:15.0339976Z test_module_init.py::TestModuleInitCPU::test_nn_Softshrink_cpu_float64 PASSED [ 70%] 2023-03-31T05:57:15.0340307Z test_module_init.py::TestModuleInitCPU::test_nn_Softsign_cpu_float32 PASSED [ 70%] 2023-03-31T05:57:15.0340653Z test_module_init.py::TestModuleInitCPU::test_nn_Softsign_cpu_float64 PASSED [ 70%] 2023-03-31T05:57:15.0341004Z test_module_init.py::TestModuleInitCPU::test_nn_SyncBatchNorm_cpu_float32 PASSED [ 71%] 2023-03-31T05:57:15.0341366Z test_module_init.py::TestModuleInitCPU::test_nn_SyncBatchNorm_cpu_float64 PASSED [ 71%] 2023-03-31T05:57:15.0341702Z test_module_init.py::TestModuleInitCPU::test_nn_Tanh_cpu_float32 PASSED [ 71%] 2023-03-31T05:57:15.0342036Z test_module_init.py::TestModuleInitCPU::test_nn_Tanh_cpu_float64 PASSED [ 71%] 2023-03-31T05:57:15.0342381Z test_module_init.py::TestModuleInitCPU::test_nn_Tanhshrink_cpu_float32 PASSED [ 72%] 2023-03-31T05:57:15.0342714Z test_module_init.py::TestModuleInitCPU::test_nn_Tanhshrink_cpu_float64 PASSED [ 72%] 2023-03-31T05:57:15.0343141Z test_module_init.py::TestModuleInitCPU::test_nn_Threshold_cpu_float32 PASSED [ 72%] 2023-03-31T05:57:15.0343487Z test_module_init.py::TestModuleInitCPU::test_nn_Threshold_cpu_float64 PASSED [ 73%] 2023-03-31T05:57:15.0343905Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoderLayer_cpu_float32 PASSED [ 73%] 2023-03-31T05:57:15.0344293Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoderLayer_cpu_float64 PASSED [ 73%] 2023-03-31T05:57:15.0344685Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoder_cpu_float32 PASSED [ 73%] 2023-03-31T05:57:15.0345064Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerDecoder_cpu_float64 PASSED [ 74%] 2023-03-31T05:57:15.0345450Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoderLayer_cpu_float32 PASSED [ 74%] 2023-03-31T05:57:15.0345833Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoderLayer_cpu_float64 PASSED [ 74%] 2023-03-31T05:57:15.0346217Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoder_cpu_float32 PASSED [ 74%] 2023-03-31T05:57:15.0346627Z test_module_init.py::TestModuleInitCPU::test_nn_TransformerEncoder_cpu_float64 PASSED [ 75%] 2023-03-31T05:57:15.0346978Z test_module_init.py::TestModuleInitCPU::test_nn_Transformer_cpu_float32 PASSED [ 75%] 2023-03-31T05:57:15.0347329Z test_module_init.py::TestModuleInitCPU::test_nn_Transformer_cpu_float64 PASSED [ 75%] 2023-03-31T05:57:15.0347686Z test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginLoss_cpu_float32 PASSED [ 75%] 2023-03-31T05:57:15.0348060Z test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginLoss_cpu_float64 PASSED [ 76%] 2023-03-31T05:57:15.0348451Z test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginWithDistanceLoss_cpu_float32 PASSED [ 76%] 2023-03-31T05:57:15.0348877Z test_module_init.py::TestModuleInitCPU::test_nn_TripletMarginWithDistanceLoss_cpu_float64 PASSED [ 76%] 2023-03-31T05:57:15.0349259Z test_module_init.py::TestModuleInitCPU::test_nn_Unflatten_cpu_float32 PASSED [ 76%] 2023-03-31T05:57:15.0349590Z test_module_init.py::TestModuleInitCPU::test_nn_Unflatten_cpu_float64 PASSED [ 77%] 2023-03-31T05:57:15.0349938Z test_module_init.py::TestModuleInitCPU::test_nn_Unfold_cpu_float32 PASSED [ 77%] 2023-03-31T05:57:15.0350275Z test_module_init.py::TestModuleInitCPU::test_nn_Unfold_cpu_float64 PASSED [ 77%] 2023-03-31T05:57:15.0350619Z test_module_init.py::TestModuleInitCPU::test_nn_Upsample_cpu_float32 PASSED [ 78%] 2023-03-31T05:57:15.0350951Z test_module_init.py::TestModuleInitCPU::test_nn_Upsample_cpu_float64 PASSED [ 78%] 2023-03-31T05:57:15.0351354Z test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingBilinear2d_cpu_float32 PASSED [ 78%] 2023-03-31T05:57:15.0351765Z test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingBilinear2d_cpu_float64 PASSED [ 78%] 2023-03-31T05:57:15.0352135Z test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingNearest2d_cpu_float32 PASSED [ 79%] 2023-03-31T05:57:15.0352515Z test_module_init.py::TestModuleInitCPU::test_nn_UpsamplingNearest2d_cpu_float64 PASSED [ 79%] 2023-03-31T05:57:15.0352882Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad1d_cpu_float32 PASSED [ 79%] 2023-03-31T05:57:15.0353226Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad1d_cpu_float64 PASSED [ 79%] 2023-03-31T05:57:15.0353553Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad2d_cpu_float32 PASSED [ 80%] 2023-03-31T05:57:15.0353900Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad2d_cpu_float64 PASSED [ 80%] 2023-03-31T05:57:15.0354243Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad3d_cpu_float32 PASSED [ 80%] 2023-03-31T05:57:15.0354581Z test_module_init.py::TestModuleInitCPU::test_nn_ZeroPad3d_cpu_float64 PASSED [ 80%] 2023-03-31T05:57:15.0354914Z test_module_init.py::TestModuleInitCPU::test_qat_Conv1d_cpu_float32 PASSED [ 81%] 2023-03-31T05:57:15.0355249Z test_module_init.py::TestModuleInitCPU::test_qat_Conv1d_cpu_float64 PASSED [ 81%] 2023-03-31T05:57:15.0355586Z test_module_init.py::TestModuleInitCPU::test_qat_Conv2d_cpu_float32 PASSED [ 81%] 2023-03-31T05:57:15.0355914Z test_module_init.py::TestModuleInitCPU::test_qat_Conv2d_cpu_float64 PASSED [ 82%] 2023-03-31T05:57:15.0356280Z test_module_init.py::TestModuleInitCPU::test_qat_Conv3d_cpu_float32 PASSED [ 82%] 2023-03-31T05:57:15.0356615Z test_module_init.py::TestModuleInitCPU::test_qat_Conv3d_cpu_float64 PASSED [ 82%] 2023-03-31T05:57:15.0356963Z test_module_init.py::TestModuleInitCPU::test_qat_EmbeddingBag_cpu_float32 PASSED [ 82%] 2023-03-31T05:57:15.0357308Z test_module_init.py::TestModuleInitCPU::test_qat_EmbeddingBag_cpu_float64 PASSED [ 83%] 2023-03-31T05:57:15.0357661Z test_module_init.py::TestModuleInitCPU::test_qat_Embedding_cpu_float32 PASSED [ 83%] 2023-03-31T05:57:15.0358011Z test_module_init.py::TestModuleInitCPU::test_qat_Embedding_cpu_float64 PASSED [ 83%] 2023-03-31T05:57:15.0358342Z test_module_init.py::TestModuleInitCPU::test_qat_Linear_cpu_float32 PASSED [ 83%] 2023-03-31T05:57:15.0358680Z test_module_init.py::TestModuleInitCPU::test_qat_Linear_cpu_float64 PASSED [ 84%] 2023-03-31T05:57:15.0359060Z test_module_init.py::TestModuleInitCPU::test_quantizable_LSTMCell_cpu_float32 PASSED [ 84%] 2023-03-31T05:57:15.0359431Z test_module_init.py::TestModuleInitCPU::test_quantizable_LSTMCell_cpu_float64 PASSED [ 84%] 2023-03-31T05:57:15.0359778Z test_module_init.py::TestModuleInitCPU::test_quantizable_LSTM_cpu_float32 PASSED [ 84%] 2023-03-31T05:57:15.0360136Z test_module_init.py::TestModuleInitCPU::test_quantizable_LSTM_cpu_float64 PASSED [ 85%] 2023-03-31T05:57:15.0360513Z test_module_init.py::TestModuleInitCPU::test_quantizable_MultiheadAttention_cpu_float32 PASSED [ 85%] 2023-03-31T05:57:15.0360903Z test_module_init.py::TestModuleInitCPU::test_quantizable_MultiheadAttention_cpu_float64 PASSED [ 85%] 2023-03-31T05:57:15.0361287Z test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm2d_cpu_float32 PASSED [ 85%] 2023-03-31T05:57:15.0361654Z test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm2d_cpu_float64 PASSED [ 86%] 2023-03-31T05:57:15.0362015Z test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm3d_cpu_float32 PASSED [ 86%] 2023-03-31T05:57:15.0362362Z test_module_init.py::TestModuleInitCPU::test_quantized_BatchNorm3d_cpu_float64 PASSED [ 86%] 2023-03-31T05:57:15.0362722Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv1d_cpu_float32 PASSED [ 87%] 2023-03-31T05:57:15.0363352Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv1d_cpu_float64 PASSED [ 87%] 2023-03-31T05:57:15.0363709Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv2d_cpu_float32 PASSED [ 87%] 2023-03-31T05:57:15.0364118Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv2d_cpu_float64 PASSED [ 87%] 2023-03-31T05:57:15.0364505Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv3d_cpu_float32 PASSED [ 88%] 2023-03-31T05:57:15.0364852Z test_module_init.py::TestModuleInitCPU::test_quantized_Conv3d_cpu_float64 PASSED [ 88%] 2023-03-31T05:57:15.0365220Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose1d_cpu_float32 PASSED [ 88%] 2023-03-31T05:57:15.0365597Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose1d_cpu_float64 PASSED [ 88%] 2023-03-31T05:57:15.0365980Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose2d_cpu_float32 PASSED [ 89%] 2023-03-31T05:57:15.0366360Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose2d_cpu_float64 PASSED [ 89%] 2023-03-31T05:57:15.0366728Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose3d_cpu_float32 PASSED [ 89%] 2023-03-31T05:57:15.0367108Z test_module_init.py::TestModuleInitCPU::test_quantized_ConvTranspose3d_cpu_float64 PASSED [ 89%] 2023-03-31T05:57:15.0367485Z test_module_init.py::TestModuleInitCPU::test_quantized_DeQuantize_cpu_float32 PASSED [ 90%] 2023-03-31T05:57:15.0367855Z test_module_init.py::TestModuleInitCPU::test_quantized_DeQuantize_cpu_float64 PASSED [ 90%] 2023-03-31T05:57:15.0368204Z test_module_init.py::TestModuleInitCPU::test_quantized_Dropout_cpu_float32 PASSED [ 90%] 2023-03-31T05:57:15.0368562Z test_module_init.py::TestModuleInitCPU::test_quantized_Dropout_cpu_float64 PASSED [ 91%] 2023-03-31T05:57:15.0368921Z test_module_init.py::TestModuleInitCPU::test_quantized_ELU_cpu_float32 PASSED [ 91%] 2023-03-31T05:57:15.0369299Z test_module_init.py::TestModuleInitCPU::test_quantized_ELU_cpu_float64 PASSED [ 91%] 2023-03-31T05:57:15.0369677Z test_module_init.py::TestModuleInitCPU::test_quantized_FXFloatFunctional_cpu_float32 PASSED [ 91%] 2023-03-31T05:57:15.0370062Z test_module_init.py::TestModuleInitCPU::test_quantized_FXFloatFunctional_cpu_float64 PASSED [ 92%] 2023-03-31T05:57:15.0370455Z test_module_init.py::TestModuleInitCPU::test_quantized_FloatFunctional_cpu_float32 PASSED [ 92%] 2023-03-31T05:57:15.0370828Z test_module_init.py::TestModuleInitCPU::test_quantized_FloatFunctional_cpu_float64 PASSED [ 92%] 2023-03-31T05:57:15.0371196Z test_module_init.py::TestModuleInitCPU::test_quantized_GroupNorm_cpu_float32 PASSED [ 92%] 2023-03-31T05:57:15.0371560Z test_module_init.py::TestModuleInitCPU::test_quantized_GroupNorm_cpu_float64 PASSED [ 93%] 2023-03-31T05:57:15.0371964Z test_module_init.py::TestModuleInitCPU::test_quantized_Hardswish_cpu_float32 PASSED [ 93%] 2023-03-31T05:57:15.0372309Z test_module_init.py::TestModuleInitCPU::test_quantized_Hardswish_cpu_float64 PASSED [ 93%] 2023-03-31T05:57:15.0372678Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm1d_cpu_float32 PASSED [ 93%] 2023-03-31T05:57:15.0373054Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm1d_cpu_float64 PASSED [ 94%] 2023-03-31T05:57:15.0373418Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm2d_cpu_float32 PASSED [ 94%] 2023-03-31T05:57:15.0373793Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm2d_cpu_float64 PASSED [ 94%] 2023-03-31T05:57:15.0374164Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm3d_cpu_float32 PASSED [ 94%] 2023-03-31T05:57:15.0374536Z test_module_init.py::TestModuleInitCPU::test_quantized_InstanceNorm3d_cpu_float64 PASSED [ 95%] 2023-03-31T05:57:15.0374891Z test_module_init.py::TestModuleInitCPU::test_quantized_LayerNorm_cpu_float32 PASSED [ 95%] 2023-03-31T05:57:15.0375254Z test_module_init.py::TestModuleInitCPU::test_quantized_LayerNorm_cpu_float64 PASSED [ 95%] 2023-03-31T05:57:15.0375608Z test_module_init.py::TestModuleInitCPU::test_quantized_LeakyReLU_cpu_float32 PASSED [ 96%] 2023-03-31T05:57:15.0375948Z test_module_init.py::TestModuleInitCPU::test_quantized_LeakyReLU_cpu_float64 PASSED [ 96%] 2023-03-31T05:57:15.0376300Z test_module_init.py::TestModuleInitCPU::test_quantized_Linear_cpu_float32 PASSED [ 96%] 2023-03-31T05:57:15.0376685Z test_module_init.py::TestModuleInitCPU::test_quantized_Linear_cpu_float64 PASSED [ 96%] 2023-03-31T05:57:15.0377070Z test_module_init.py::TestModuleInitCPU::test_quantized_PReLU_cpu_float32 PASSED [ 97%] 2023-03-31T05:57:15.0377411Z test_module_init.py::TestModuleInitCPU::test_quantized_PReLU_cpu_float64 PASSED [ 97%] 2023-03-31T05:57:15.0377771Z test_module_init.py::TestModuleInitCPU::test_quantized_QFunctional_cpu_float32 PASSED [ 97%] 2023-03-31T05:57:15.0378146Z test_module_init.py::TestModuleInitCPU::test_quantized_QFunctional_cpu_float64 PASSED [ 97%] 2023-03-31T05:57:15.0378511Z test_module_init.py::TestModuleInitCPU::test_quantized_Quantize_cpu_float32 PASSED [ 98%] 2023-03-31T05:57:15.0378858Z test_module_init.py::TestModuleInitCPU::test_quantized_Quantize_cpu_float64 PASSED [ 98%] 2023-03-31T05:57:15.0379210Z test_module_init.py::TestModuleInitCPU::test_quantized_ReLU6_cpu_float32 PASSED [ 98%] 2023-03-31T05:57:15.0379559Z test_module_init.py::TestModuleInitCPU::test_quantized_ReLU6_cpu_float64 PASSED [ 98%] 2023-03-31T05:57:15.0379905Z test_module_init.py::TestModuleInitCPU::test_quantized_Sigmoid_cpu_float32 PASSED [ 99%] 2023-03-31T05:57:15.0380263Z test_module_init.py::TestModuleInitCPU::test_quantized_Sigmoid_cpu_float64 PASSED [ 99%] 2023-03-31T05:57:15.0380615Z test_module_init.py::TestModuleInitCPU::test_quantized_Softmax_cpu_float32 PASSED [ 99%] 2023-03-31T05:57:15.0380967Z test_module_init.py::TestModuleInitCPU::test_quantized_Softmax_cpu_float64 PASSED [100%] 2023-03-31T05:57:15.0381156Z 2023-03-31T05:57:15.0381649Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_module_init/test_module_init-855433dfa1eb4a15.xml - 2023-03-31T05:57:15.0382021Z ============================= 378 passed in 2.52s ============================== 2023-03-31T05:57:15.0382338Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:15.0382541Z 2023-03-31T05:57:15.0382879Z ##[endgroup] 2023-03-31T05:57:15.0383395Z FINISHED PRINTING LOG FILE of test_module_init (/var/lib/jenkins/workspace/test/test-reports/test_module_init_lmqnl5dh.log) 2023-03-31T05:57:15.0383630Z 2023-03-31T05:57:17.0872609Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:17.0900157Z Ignoring disabled issues: [] 2023-03-31T05:57:17.1073459Z Running dynamo/test_optimizers ... [2023-03-31 05:57:17.106991] 2023-03-31T05:57:17.1075196Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_optimizers.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:17.107280] 2023-03-31T05:57:21.2343696Z 2023-03-31T05:57:21.2344262Z Expand the folded group to see the log file of dynamo/test_optimizers 2023-03-31T05:57:21.2345484Z ##[group]PRINTING LOG FILE of dynamo/test_optimizers (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_optimizers_9a07r5wo.log) 2023-03-31T05:57:21.2346610Z Test results will be stored in test-reports/python-pytest/dynamo.test_optimizers/dynamo.test_optimizers-2c37486a74067b12.xml 2023-03-31T05:57:21.2347205Z ============================= test session starts ============================== 2023-03-31T05:57:21.2348724Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:21.2349194Z cachedir: .pytest_cache 2023-03-31T05:57:21.2349957Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:21.2350584Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:21.2351333Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:21.2351831Z collecting ... collected 12 items 2023-03-31T05:57:21.2353854Z Running 12 items in this shard: test/dynamo/test_optimizers.py::OptimizerTests::test_adadelta, test/dynamo/test_optimizers.py::OptimizerTests::test_adagrad, test/dynamo/test_optimizers.py::OptimizerTests::test_adam, test/dynamo/test_optimizers.py::OptimizerTests::test_adamax, test/dynamo/test_optimizers.py::OptimizerTests::test_adamw, test/dynamo/test_optimizers.py::OptimizerTests::test_asgd, test/dynamo/test_optimizers.py::OptimizerTests::test_nadam, test/dynamo/test_optimizers.py::OptimizerTests::test_radam, test/dynamo/test_optimizers.py::OptimizerTests::test_rmsprop, test/dynamo/test_optimizers.py::OptimizerTests::test_rprop, test/dynamo/test_optimizers.py::OptimizerTests::test_sgd, test/dynamo/test_optimizers.py::End2EndTests::test_optimizing_over_tensor_with_requires_grad 2023-03-31T05:57:21.2355542Z 2023-03-31T05:57:21.2356192Z dynamo/test_optimizers.py::OptimizerTests::test_adadelta [2023-03-31 05:57:18,990] torch._dynamo.variables.torch: [WARNING] Profiler will be ignored 2023-03-31T05:57:21.2356730Z PASSED [ 8%] 2023-03-31T05:57:21.2357163Z dynamo/test_optimizers.py::OptimizerTests::test_adagrad PASSED [ 16%] 2023-03-31T05:57:21.2357702Z dynamo/test_optimizers.py::OptimizerTests::test_adam PASSED [ 25%] 2023-03-31T05:57:21.2358235Z dynamo/test_optimizers.py::OptimizerTests::test_adamax PASSED [ 33%] 2023-03-31T05:57:21.2358767Z dynamo/test_optimizers.py::OptimizerTests::test_adamw PASSED [ 41%] 2023-03-31T05:57:21.2359330Z dynamo/test_optimizers.py::OptimizerTests::test_asgd PASSED [ 50%] 2023-03-31T05:57:21.2359852Z dynamo/test_optimizers.py::OptimizerTests::test_nadam PASSED [ 58%] 2023-03-31T05:57:21.2360947Z dynamo/test_optimizers.py::OptimizerTests::test_radam [2023-03-31 05:57:20,231] torch._dynamo.variables.torch: [WARNING] Profiler will be ignored 2023-03-31T05:57:21.2361512Z PASSED [ 66%] 2023-03-31T05:57:21.2361955Z dynamo/test_optimizers.py::OptimizerTests::test_rmsprop PASSED [ 75%] 2023-03-31T05:57:21.2362522Z dynamo/test_optimizers.py::OptimizerTests::test_rprop PASSED [ 83%] 2023-03-31T05:57:21.2363323Z dynamo/test_optimizers.py::OptimizerTests::test_sgd PASSED [ 91%] 2023-03-31T05:57:21.2363945Z dynamo/test_optimizers.py::End2EndTests::test_optimizing_over_tensor_with_requires_grad PASSED [100%] 2023-03-31T05:57:21.2364274Z 2023-03-31T05:57:21.2364906Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_optimizers/dynamo.test_optimizers-2c37486a74067b12.xml - 2023-03-31T05:57:21.2365667Z ============================== 12 passed in 2.06s ============================== 2023-03-31T05:57:21.2366195Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:21.2366525Z 2023-03-31T05:57:21.2366940Z ##[endgroup] 2023-03-31T05:57:21.2367669Z FINISHED PRINTING LOG FILE of dynamo/test_optimizers (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_optimizers_9a07r5wo.log) 2023-03-31T05:57:21.2368085Z 2023-03-31T05:57:23.2766497Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:23.2794557Z Ignoring disabled issues: [] 2023-03-31T05:57:23.2966094Z Running test_logging ... [2023-03-31 05:57:23.296267] 2023-03-31T05:57:23.2968561Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_logging.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:23.296592] 2023-03-31T05:57:27.2705158Z 2023-03-31T05:57:27.2705731Z Expand the folded group to see the log file of test_logging 2023-03-31T05:57:27.2706629Z ##[group]PRINTING LOG FILE of test_logging (/var/lib/jenkins/workspace/test/test-reports/test_logging_18j0gylj.log) 2023-03-31T05:57:27.2707257Z Test results will be stored in test-reports/python-pytest/test_logging/test_logging-70826ade4cb4ade2.xml 2023-03-31T05:57:27.2707595Z ============================= test session starts ============================== 2023-03-31T05:57:27.2708211Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:27.2708529Z cachedir: .pytest_cache 2023-03-31T05:57:27.2708967Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:27.2709343Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:27.2709792Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:27.2710093Z collecting ... collected 1 item 2023-03-31T05:57:27.2710376Z Running 1 items in this shard: test/test_logging.py::LoggingTest::testApiUsage 2023-03-31T05:57:27.2710567Z 2023-03-31T05:57:27.2710718Z test_logging.py::LoggingTest::testApiUsage PASSED [100%] 2023-03-31T05:57:27.2710894Z 2023-03-31T05:57:27.2711223Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_logging/test_logging-70826ade4cb4ade2.xml - 2023-03-31T05:57:27.2711571Z ============================== 1 passed in 2.07s =============================== 2023-03-31T05:57:27.2711886Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:27.2712090Z 2023-03-31T05:57:27.2712324Z ##[endgroup] 2023-03-31T05:57:27.2712705Z FINISHED PRINTING LOG FILE of test_logging (/var/lib/jenkins/workspace/test/test-reports/test_logging_18j0gylj.log) 2023-03-31T05:57:27.2712928Z 2023-03-31T05:57:29.3385142Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:29.3412404Z Ignoring disabled issues: [] 2023-03-31T05:57:29.3582676Z Running test_scatter_gather_ops ... [2023-03-31 05:57:29.357958] 2023-03-31T05:57:29.3585158Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_scatter_gather_ops.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:29.358268] 2023-03-31T05:57:33.2377928Z 2023-03-31T05:57:33.2378504Z Expand the folded group to see the log file of test_scatter_gather_ops 2023-03-31T05:57:33.2379644Z ##[group]PRINTING LOG FILE of test_scatter_gather_ops (/var/lib/jenkins/workspace/test/test-reports/test_scatter_gather_ops_cs1k_7ry.log) 2023-03-31T05:57:33.2380649Z Test results will be stored in test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-009af74fcb58de5f.xml 2023-03-31T05:57:33.2381007Z ============================= test session starts ============================== 2023-03-31T05:57:33.2381631Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:33.2381906Z cachedir: .pytest_cache 2023-03-31T05:57:33.2382344Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:33.2382716Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:33.2383251Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:33.2383556Z collecting ... collected 80 items 2023-03-31T05:57:33.2392734Z Running 80 items in this shard: test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_False_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_False_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_True_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_True_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_bool_cpu_bool, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add_mult_index_base_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_bool, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_uint8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_bool, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_uint8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_complex128, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_uint8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_bool, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_complex128, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_uint8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_bfloat16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_complex128, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_complex64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int16, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int32, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int64, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int8, test/test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_uint8 2023-03-31T05:57:33.2401376Z 2023-03-31T05:57:33.2401621Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_False_cpu_float32 PASSED [ 1%] 2023-03-31T05:57:33.2402109Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_False_cpu_float64 PASSED [ 2%] 2023-03-31T05:57:33.2402570Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_True_cpu_float32 PASSED [ 3%] 2023-03-31T05:57:33.2403207Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_backward_with_empty_index_tensor_sparse_grad_True_cpu_float64 PASSED [ 5%] 2023-03-31T05:57:33.2403652Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_bool_cpu_bool PASSED [ 6%] 2023-03-31T05:57:33.2404025Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_cpu_complex64 PASSED [ 7%] 2023-03-31T05:57:33.2404383Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_cpu_float32 PASSED [ 8%] 2023-03-31T05:57:33.2404775Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_bfloat16 PASSED [ 10%] 2023-03-31T05:57:33.2405181Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_float32 PASSED [ 11%] 2023-03-31T05:57:33.2405580Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_gather_expanded_index_cpu_float64 PASSED [ 12%] 2023-03-31T05:57:33.2405953Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_complex64 PASSED [ 13%] 2023-03-31T05:57:33.2406389Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_float16 PASSED [ 15%] 2023-03-31T05:57:33.2406788Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__cpu_float32 PASSED [ 16%] 2023-03-31T05:57:33.2407161Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_complex64 PASSED [ 17%] 2023-03-31T05:57:33.2407555Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_float16 PASSED [ 18%] 2023-03-31T05:57:33.2407946Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__reductions_cpu_float32 PASSED [ 20%] 2023-03-31T05:57:33.2408344Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_complex64 PASSED [ 21%] 2023-03-31T05:57:33.2408717Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_float16 PASSED [ 22%] 2023-03-31T05:57:33.2409099Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter__scalar_cpu_float32 PASSED [ 23%] 2023-03-31T05:57:33.2409479Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_complex64 PASSED [ 25%] 2023-03-31T05:57:33.2409861Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_float16 PASSED [ 26%] 2023-03-31T05:57:33.2410225Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add__cpu_float32 PASSED [ 27%] 2023-03-31T05:57:33.2410620Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_add_mult_index_base_cpu_float32 PASSED [ 28%] 2023-03-31T05:57:33.2411032Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_bfloat16 PASSED [ 30%] 2023-03-31T05:57:33.2411468Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_float32 PASSED [ 31%] 2023-03-31T05:57:33.2411866Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_expanded_index_cpu_float64 PASSED [ 32%] 2023-03-31T05:57:33.2412258Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_bfloat16 PASSED [ 33%] 2023-03-31T05:57:33.2412647Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_bool PASSED [ 35%] 2023-03-31T05:57:33.2413026Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float16 PASSED [ 36%] 2023-03-31T05:57:33.2413412Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float32 PASSED [ 37%] 2023-03-31T05:57:33.2413800Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_float64 PASSED [ 38%] 2023-03-31T05:57:33.2414233Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int16 PASSED [ 40%] 2023-03-31T05:57:33.2414606Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int32 PASSED [ 41%] 2023-03-31T05:57:33.2414989Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int64 PASSED [ 42%] 2023-03-31T05:57:33.2415374Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_int8 PASSED [ 43%] 2023-03-31T05:57:33.2415748Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amax_cpu_uint8 PASSED [ 45%] 2023-03-31T05:57:33.2416139Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_bfloat16 PASSED [ 46%] 2023-03-31T05:57:33.2416524Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_bool PASSED [ 47%] 2023-03-31T05:57:33.2416908Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float16 PASSED [ 48%] 2023-03-31T05:57:33.2417284Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float32 PASSED [ 50%] 2023-03-31T05:57:33.2417674Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_float64 PASSED [ 51%] 2023-03-31T05:57:33.2418062Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int16 PASSED [ 52%] 2023-03-31T05:57:33.2418443Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int32 PASSED [ 53%] 2023-03-31T05:57:33.2418858Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int64 PASSED [ 55%] 2023-03-31T05:57:33.2419270Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_int8 PASSED [ 56%] 2023-03-31T05:57:33.2419653Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_amin_cpu_uint8 PASSED [ 57%] 2023-03-31T05:57:33.2420024Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_bfloat16 PASSED [ 58%] 2023-03-31T05:57:33.2420421Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_complex128 PASSED [ 60%] 2023-03-31T05:57:33.2420823Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_complex64 PASSED [ 61%] 2023-03-31T05:57:33.2421216Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float16 PASSED [ 62%] 2023-03-31T05:57:33.2421592Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float32 PASSED [ 63%] 2023-03-31T05:57:33.2421981Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_float64 PASSED [ 65%] 2023-03-31T05:57:33.2422366Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int16 PASSED [ 66%] 2023-03-31T05:57:33.2422749Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int32 PASSED [ 67%] 2023-03-31T05:57:33.2423193Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int64 PASSED [ 68%] 2023-03-31T05:57:33.2423584Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_int8 PASSED [ 70%] 2023-03-31T05:57:33.2424004Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_mean_cpu_uint8 PASSED [ 71%] 2023-03-31T05:57:33.2424379Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_bfloat16 PASSED [ 72%] 2023-03-31T05:57:33.2424766Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_bool PASSED [ 73%] 2023-03-31T05:57:33.2425168Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_complex128 PASSED [ 75%] 2023-03-31T05:57:33.2425573Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_complex64 PASSED [ 76%] 2023-03-31T05:57:33.2425959Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float16 PASSED [ 77%] 2023-03-31T05:57:33.2426350Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float32 PASSED [ 78%] 2023-03-31T05:57:33.2426788Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_float64 PASSED [ 80%] 2023-03-31T05:57:33.2427179Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int16 PASSED [ 81%] 2023-03-31T05:57:33.2427553Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int32 PASSED [ 82%] 2023-03-31T05:57:33.2427940Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int64 PASSED [ 83%] 2023-03-31T05:57:33.2428327Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_int8 PASSED [ 85%] 2023-03-31T05:57:33.2428701Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_prod_cpu_uint8 PASSED [ 86%] 2023-03-31T05:57:33.2429091Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_bfloat16 PASSED [ 87%] 2023-03-31T05:57:33.2429492Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_complex128 PASSED [ 88%] 2023-03-31T05:57:33.2429891Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_complex64 PASSED [ 90%] 2023-03-31T05:57:33.2430270Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float16 PASSED [ 91%] 2023-03-31T05:57:33.2430662Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float32 PASSED [ 92%] 2023-03-31T05:57:33.2431086Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_float64 PASSED [ 93%] 2023-03-31T05:57:33.2431495Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int16 PASSED [ 95%] 2023-03-31T05:57:33.2431859Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int32 PASSED [ 96%] 2023-03-31T05:57:33.2432234Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int64 PASSED [ 97%] 2023-03-31T05:57:33.2432625Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_int8 PASSED [ 98%] 2023-03-31T05:57:33.2432994Z test_scatter_gather_ops.py::TestScatterGatherCPU::test_scatter_reduce_sum_cpu_uint8 PASSED [100%] 2023-03-31T05:57:33.2433202Z 2023-03-31T05:57:33.2433626Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-009af74fcb58de5f.xml - 2023-03-31T05:57:33.2434014Z ============================== 80 passed in 1.74s ============================== 2023-03-31T05:57:33.2434340Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:33.2434547Z 2023-03-31T05:57:33.2434827Z ##[endgroup] 2023-03-31T05:57:33.2435265Z FINISHED PRINTING LOG FILE of test_scatter_gather_ops (/var/lib/jenkins/workspace/test/test-reports/test_scatter_gather_ops_cs1k_7ry.log) 2023-03-31T05:57:33.2435512Z 2023-03-31T05:57:35.2968132Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:35.2994228Z Ignoring disabled issues: [] 2023-03-31T05:57:35.3167300Z Running test_public_bindings ... [2023-03-31 05:57:35.316383] 2023-03-31T05:57:35.3169801Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_public_bindings.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:35.316716] 2023-03-31T05:57:39.2912089Z 2023-03-31T05:57:39.2913100Z Expand the folded group to see the log file of test_public_bindings 2023-03-31T05:57:39.2914426Z ##[group]PRINTING LOG FILE of test_public_bindings (/var/lib/jenkins/workspace/test/test-reports/test_public_bindings_3xe300o5.log) 2023-03-31T05:57:39.2915607Z Test results will be stored in test-reports/python-pytest/test_public_bindings/test_public_bindings-46374e86c8090fb4.xml 2023-03-31T05:57:39.2916138Z ============================= test session starts ============================== 2023-03-31T05:57:39.2916847Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:39.2917640Z cachedir: .pytest_cache 2023-03-31T05:57:39.2918435Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:39.2919018Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:39.2919469Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:39.2919780Z collecting ... collected 2 items 2023-03-31T05:57:39.2920174Z Running 2 items in this shard: test/test_public_bindings.py::TestPublicBindings::test_correct_module_names, test/test_public_bindings.py::TestPublicBindings::test_no_new_bindings 2023-03-31T05:57:39.2920465Z 2023-03-31T05:57:39.2920639Z test_public_bindings.py::TestPublicBindings::test_correct_module_names PASSED [ 50%] 2023-03-31T05:57:39.2920997Z test_public_bindings.py::TestPublicBindings::test_no_new_bindings PASSED [100%] 2023-03-31T05:57:39.2921192Z 2023-03-31T05:57:39.2921552Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_public_bindings/test_public_bindings-46374e86c8090fb4.xml - 2023-03-31T05:57:39.2921928Z ============================== 2 passed in 1.86s =============================== 2023-03-31T05:57:39.2922241Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:39.2922444Z 2023-03-31T05:57:39.2922693Z ##[endgroup] 2023-03-31T05:57:39.2923391Z FINISHED PRINTING LOG FILE of test_public_bindings (/var/lib/jenkins/workspace/test/test-reports/test_public_bindings_3xe300o5.log) 2023-03-31T05:57:39.2923688Z 2023-03-31T05:57:41.3723682Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:41.3752298Z Ignoring disabled issues: [] 2023-03-31T05:57:41.3925504Z Running dynamo/test_unspec ... [2023-03-31 05:57:41.392149] 2023-03-31T05:57:41.3927537Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_unspec.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:41.392528] 2023-03-31T05:57:45.2146890Z 2023-03-31T05:57:45.2147409Z Expand the folded group to see the log file of dynamo/test_unspec 2023-03-31T05:57:45.2148398Z ##[group]PRINTING LOG FILE of dynamo/test_unspec (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_unspec_kba9n2e_.log) 2023-03-31T05:57:45.2149093Z Test results will be stored in test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-e0d23d9b5160040a.xml 2023-03-31T05:57:45.2149456Z ============================= test session starts ============================== 2023-03-31T05:57:45.2149854Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:45.2150133Z cachedir: .pytest_cache 2023-03-31T05:57:45.2150555Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:45.2150930Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:45.2151578Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:45.2151892Z collecting ... collected 12 items 2023-03-31T05:57:45.2153044Z Running 12 items in this shard: test/dynamo/test_unspec.py::UnspecTests::test_builtin_functions_on_cuda, test/dynamo/test_unspec.py::UnspecTests::test_builtin_getitem, test/dynamo/test_unspec.py::UnspecTests::test_builtin_max_min, test/dynamo/test_unspec.py::UnspecTests::test_feed_random_values_into_graph_only, test/dynamo/test_unspec.py::UnspecTests::test_multiple_consecutive_random_calls_before_graph, test/dynamo/test_unspec.py::UnspecTests::test_no_recompilations, test/dynamo/test_unspec.py::UnspecTests::test_numpy_correctness, test/dynamo/test_unspec.py::UnspecTests::test_random_call_with_while_loop, test/dynamo/test_unspec.py::UnspecTests::test_random_values_with_graph_break, test/dynamo/test_unspec.py::UnspecTests::test_shape_graph_break, test/dynamo/test_unspec.py::UnspecTests::test_specializing_numpy_float_in_control_flow, test/dynamo/test_unspec.py::UnspecTests::test_unspec_float_precision 2023-03-31T05:57:45.2154136Z 2023-03-31T05:57:45.2154320Z dynamo/test_unspec.py::UnspecTests::test_builtin_functions_on_cuda SKIPPED (requires cuda) [ 8%] 2023-03-31T05:57:45.2154672Z dynamo/test_unspec.py::UnspecTests::test_builtin_getitem XFAIL [ 16%] 2023-03-31T05:57:45.2154990Z dynamo/test_unspec.py::UnspecTests::test_builtin_max_min PASSED [ 25%] 2023-03-31T05:57:45.2155331Z dynamo/test_unspec.py::UnspecTests::test_feed_random_values_into_graph_only PASSED [ 33%] 2023-03-31T05:57:45.2155709Z dynamo/test_unspec.py::UnspecTests::test_multiple_consecutive_random_calls_before_graph PASSED [ 41%] 2023-03-31T05:57:45.2156060Z dynamo/test_unspec.py::UnspecTests::test_no_recompilations PASSED [ 50%] 2023-03-31T05:57:45.2156394Z dynamo/test_unspec.py::UnspecTests::test_numpy_correctness PASSED [ 58%] 2023-03-31T05:57:45.2156735Z dynamo/test_unspec.py::UnspecTests::test_random_call_with_while_loop PASSED [ 66%] 2023-03-31T05:57:45.2157083Z dynamo/test_unspec.py::UnspecTests::test_random_values_with_graph_break PASSED [ 75%] 2023-03-31T05:57:45.2157408Z dynamo/test_unspec.py::UnspecTests::test_shape_graph_break PASSED [ 83%] 2023-03-31T05:57:45.2157759Z dynamo/test_unspec.py::UnspecTests::test_specializing_numpy_float_in_control_flow PASSED [ 91%] 2023-03-31T05:57:45.2158166Z dynamo/test_unspec.py::UnspecTests::test_unspec_float_precision PASSED [100%] 2023-03-31T05:57:45.2158394Z 2023-03-31T05:57:45.2158741Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-e0d23d9b5160040a.xml - 2023-03-31T05:57:45.2159128Z =================== 10 passed, 1 skipped, 1 xfailed in 1.70s =================== 2023-03-31T05:57:45.2159459Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:45.2159667Z 2023-03-31T05:57:45.2159901Z ##[endgroup] 2023-03-31T05:57:45.2160313Z FINISHED PRINTING LOG FILE of dynamo/test_unspec (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_unspec_kba9n2e_.log) 2023-03-31T05:57:45.2160549Z 2023-03-31T05:57:47.3177147Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:47.3206882Z Ignoring disabled issues: [] 2023-03-31T05:57:47.3379977Z Running dynamo/test_functions ... [2023-03-31 05:57:47.337495] 2023-03-31T05:57:47.3381480Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_functions.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:47.337870] 2023-03-31T05:57:51.1665871Z 2023-03-31T05:57:51.1666731Z Expand the folded group to see the log file of dynamo/test_functions 2023-03-31T05:57:51.1667906Z ##[group]PRINTING LOG FILE of dynamo/test_functions (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_functions_r0p_nj4k.log) 2023-03-31T05:57:51.1669381Z Test results will be stored in test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-85421862794e1580.xml 2023-03-31T05:57:51.1670025Z ============================= test session starts ============================== 2023-03-31T05:57:51.1670439Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:51.1670717Z cachedir: .pytest_cache 2023-03-31T05:57:51.1671146Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:51.1671531Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:51.1671988Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:51.1672302Z collecting ... collected 128 items 2023-03-31T05:57:51.1681721Z Running 128 items in this shard: test/dynamo/test_functions.py::FunctionTests::test_T, test/dynamo/test_functions.py::FunctionTests::test_add, test/dynamo/test_functions.py::FunctionTests::test_add_, test/dynamo/test_functions.py::FunctionTests::test_addcdiv, test/dynamo/test_functions.py::FunctionTests::test_addcdiv_, test/dynamo/test_functions.py::FunctionTests::test_build_list_unpack, test/dynamo/test_functions.py::FunctionTests::test_call_dict1, test/dynamo/test_functions.py::FunctionTests::test_call_dict2, test/dynamo/test_functions.py::FunctionTests::test_call_dict3, test/dynamo/test_functions.py::FunctionTests::test_call_dict4, test/dynamo/test_functions.py::FunctionTests::test_call_dict5, test/dynamo/test_functions.py::FunctionTests::test_chunks1, test/dynamo/test_functions.py::FunctionTests::test_const_tuple_add1, test/dynamo/test_functions.py::FunctionTests::test_const_tuple_add2, test/dynamo/test_functions.py::FunctionTests::test_constant1, test/dynamo/test_functions.py::FunctionTests::test_constant2, test/dynamo/test_functions.py::FunctionTests::test_constant3, test/dynamo/test_functions.py::FunctionTests::test_constant4, test/dynamo/test_functions.py::FunctionTests::test_context_wrapping_nested_functions_no_closure, test/dynamo/test_functions.py::FunctionTests::test_default_dict, test/dynamo/test_functions.py::FunctionTests::test_del, test/dynamo/test_functions.py::FunctionTests::test_device, test/dynamo/test_functions.py::FunctionTests::test_device_constant, test/dynamo/test_functions.py::FunctionTests::test_dict_copy, test/dynamo/test_functions.py::FunctionTests::test_dict_ops, test/dynamo/test_functions.py::FunctionTests::test_dict_param_keys, test/dynamo/test_functions.py::FunctionTests::test_dict_sorted, test/dynamo/test_functions.py::FunctionTests::test_distributed_is_available, test/dynamo/test_functions.py::FunctionTests::test_distributed_is_initialized, test/dynamo/test_functions.py::FunctionTests::test_dtype, test/dynamo/test_functions.py::FunctionTests::test_dtype_compare, test/dynamo/test_functions.py::FunctionTests::test_finfo, test/dynamo/test_functions.py::FunctionTests::test_float, test/dynamo/test_functions.py::FunctionTests::test_fn_with_self_set, test/dynamo/test_functions.py::FunctionTests::test_fstrings1, test/dynamo/test_functions.py::FunctionTests::test_fstrings2, test/dynamo/test_functions.py::FunctionTests::test_fstrings3, test/dynamo/test_functions.py::FunctionTests::test_funcdef_closure, test/dynamo/test_functions.py::FunctionTests::test_get_default_dtype, test/dynamo/test_functions.py::FunctionTests::test_globalfn, test/dynamo/test_functions.py::FunctionTests::test_globalmodule, test/dynamo/test_functions.py::FunctionTests::test_globalvar, test/dynamo/test_functions.py::FunctionTests::test_import1, test/dynamo/test_functions.py::FunctionTests::test_indirect1, test/dynamo/test_functions.py::FunctionTests::test_indirect2, test/dynamo/test_functions.py::FunctionTests::test_indirect3, test/dynamo/test_functions.py::FunctionTests::test_inline_jit_annotations, test/dynamo/test_functions.py::FunctionTests::test_inline_softmax, test/dynamo/test_functions.py::FunctionTests::test_inline_with_default, test/dynamo/test_functions.py::FunctionTests::test_inner_function, test/dynamo/test_functions.py::FunctionTests::test_is_contiguous_memory_format, test/dynamo/test_functions.py::FunctionTests::test_is_fx_tracing, test/dynamo/test_functions.py::FunctionTests::test_is_in_onnx_export, test/dynamo/test_functions.py::FunctionTests::test_is_not_null, test/dynamo/test_functions.py::FunctionTests::test_is_quantized, test/dynamo/test_functions.py::FunctionTests::test_is_sparse, test/dynamo/test_functions.py::FunctionTests::test_islice_chain, test/dynamo/test_functions.py::FunctionTests::test_jit_annotate, test/dynamo/test_functions.py::FunctionTests::test_len_constant_dict, test/dynamo/test_functions.py::FunctionTests::test_len_constant_list, test/dynamo/test_functions.py::FunctionTests::test_len_constant_misc_iterables, test/dynamo/test_functions.py::FunctionTests::test_len_tensor, test/dynamo/test_functions.py::FunctionTests::test_list_add, test/dynamo/test_functions.py::FunctionTests::test_list_clear, test/dynamo/test_functions.py::FunctionTests::test_list_convert, test/dynamo/test_functions.py::FunctionTests::test_list_reversed, test/dynamo/test_functions.py::FunctionTests::test_list_slice_assignment, test/dynamo/test_functions.py::FunctionTests::test_list_sorted1, test/dynamo/test_functions.py::FunctionTests::test_list_sorted2, test/dynamo/test_functions.py::FunctionTests::test_list_truth, test/dynamo/test_functions.py::FunctionTests::test_listarg1, test/dynamo/test_functions.py::FunctionTests::test_listarg2, test/dynamo/test_functions.py::FunctionTests::test_listarg3, test/dynamo/test_functions.py::FunctionTests::test_listarg4, test/dynamo/test_functions.py::FunctionTests::test_listarg5, test/dynamo/test_functions.py::FunctionTests::test_load_global_bool, test/dynamo/test_functions.py::FunctionTests::test_mT, test/dynamo/test_functions.py::FunctionTests::test_map_sum, test/dynamo/test_functions.py::FunctionTests::test_methodcall1, test/dynamo/test_functions.py::FunctionTests::test_methodcall2, test/dynamo/test_functions.py::FunctionTests::test_methodcall3, test/dynamo/test_functions.py::FunctionTests::test_min_max, test/dynamo/test_functions.py::FunctionTests::test_module_constant, test/dynamo/test_functions.py::FunctionTests::test_namedtuple, test/dynamo/test_functions.py::FunctionTests::test_ndim, test/dynamo/test_functions.py::FunctionTests::test_pop, test/dynamo/test_functions.py::FunctionTests::test_range1, test/dynamo/test_functions.py::FunctionTests::test_range2, test/dynamo/test_functions.py::FunctionTests::test_reduce, test/dynamo/test_functions.py::FunctionTests::test_return_dict, test/dynamo/test_functions.py::FunctionTests::test_return_dict2, test/dynamo/test_functions.py::FunctionTests::test_return_tuple1, test/dynamo/test_functions.py::FunctionTests::test_return_tuple2, test/dynamo/test_functions.py::FunctionTests::test_shape1, test/dynamo/test_functions.py::FunctionTests::test_shape2, test/dynamo/test_functions.py::FunctionTests::test_slice1, test/dynamo/test_functions.py::FunctionTests::test_slice2, test/dynamo/test_functions.py::FunctionTests::test_slice3, test/dynamo/test_functions.py::FunctionTests::test_slice4, test/dynamo/test_functions.py::FunctionTests::test_slice5, test/dynamo/test_functions.py::FunctionTests::test_slice6, test/dynamo/test_functions.py::FunctionTests::test_startswith, test/dynamo/test_functions.py::FunctionTests::test_tensor_len, test/dynamo/test_functions.py::FunctionTests::test_tensor_new_with_shape, test/dynamo/test_functions.py::FunctionTests::test_tensor_new_with_size, test/dynamo/test_functions.py::FunctionTests::test_tensor_type, test/dynamo/test_functions.py::FunctionTests::test_tensor_type2, test/dynamo/test_functions.py::FunctionTests::test_tensor_type3, test/dynamo/test_functions.py::FunctionTests::test_tensor_type4, test/dynamo/test_functions.py::FunctionTests::test_tensor_type5, test/dynamo/test_functions.py::FunctionTests::test_torch_distributions_functions, test/dynamo/test_functions.py::FunctionTests::test_transpose_for_scores, test/dynamo/test_functions.py::FunctionTests::test_tuple1, test/dynamo/test_functions.py::FunctionTests::test_tuple2, test/dynamo/test_functions.py::FunctionTests::test_tuple_contains, test/dynamo/test_functions.py::FunctionTests::test_tuple_iadd, test/dynamo/test_functions.py::FunctionTests::test_tuple_sorted, test/dynamo/test_functions.py::FunctionTests::test_unpack1, test/dynamo/test_functions.py::FunctionTests::test_unpack2, test/dynamo/test_functions.py::FunctionTests::test_unpack3, test/dynamo/test_functions.py::FunctionTests::test_unpack_ex1, test/dynamo/test_functions.py::FunctionTests::test_unpack_ex2, test/dynamo/test_functions.py::FunctionTests::test_unpack_ex3, test/dynamo/test_functions.py::FunctionTests::test_viamethod, test/dynamo/test_functions.py::FunctionTests::test_viatorch, test/dynamo/test_functions.py::DefaultsTests::test_func_default_tensor_args, test/dynamo/test_functions.py::DefaultsTests::test_func_default_torch_args, test/dynamo/test_functions.py::DefaultsTests::test_meth_default_tensor_args 2023-03-31T05:57:51.1690723Z 2023-03-31T05:57:51.1690871Z dynamo/test_functions.py::FunctionTests::test_T PASSED [ 0%] 2023-03-31T05:57:51.1691199Z dynamo/test_functions.py::FunctionTests::test_add PASSED [ 1%] 2023-03-31T05:57:51.1691515Z dynamo/test_functions.py::FunctionTests::test_add_ PASSED [ 2%] 2023-03-31T05:57:51.1691834Z dynamo/test_functions.py::FunctionTests::test_addcdiv PASSED [ 3%] 2023-03-31T05:57:51.1692144Z dynamo/test_functions.py::FunctionTests::test_addcdiv_ PASSED [ 3%] 2023-03-31T05:57:51.1692467Z dynamo/test_functions.py::FunctionTests::test_build_list_unpack PASSED [ 4%] 2023-03-31T05:57:51.1692790Z dynamo/test_functions.py::FunctionTests::test_call_dict1 PASSED [ 5%] 2023-03-31T05:57:51.1693092Z dynamo/test_functions.py::FunctionTests::test_call_dict2 PASSED [ 6%] 2023-03-31T05:57:51.1693411Z dynamo/test_functions.py::FunctionTests::test_call_dict3 PASSED [ 7%] 2023-03-31T05:57:51.1693726Z dynamo/test_functions.py::FunctionTests::test_call_dict4 PASSED [ 7%] 2023-03-31T05:57:51.1694042Z dynamo/test_functions.py::FunctionTests::test_call_dict5 PASSED [ 8%] 2023-03-31T05:57:51.1694343Z dynamo/test_functions.py::FunctionTests::test_chunks1 PASSED [ 9%] 2023-03-31T05:57:51.1694656Z dynamo/test_functions.py::FunctionTests::test_const_tuple_add1 PASSED [ 10%] 2023-03-31T05:57:51.1695053Z dynamo/test_functions.py::FunctionTests::test_const_tuple_add2 PASSED [ 10%] 2023-03-31T05:57:51.1695403Z dynamo/test_functions.py::FunctionTests::test_constant1 PASSED [ 11%] 2023-03-31T05:57:51.1695717Z dynamo/test_functions.py::FunctionTests::test_constant2 PASSED [ 12%] 2023-03-31T05:57:51.1696028Z dynamo/test_functions.py::FunctionTests::test_constant3 PASSED [ 13%] 2023-03-31T05:57:51.1696341Z dynamo/test_functions.py::FunctionTests::test_constant4 PASSED [ 14%] 2023-03-31T05:57:51.1696784Z dynamo/test_functions.py::FunctionTests::test_context_wrapping_nested_functions_no_closure PASSED [ 14%] 2023-03-31T05:57:51.1697138Z dynamo/test_functions.py::FunctionTests::test_default_dict PASSED [ 15%] 2023-03-31T05:57:51.1697457Z dynamo/test_functions.py::FunctionTests::test_del PASSED [ 16%] 2023-03-31T05:57:51.1697770Z dynamo/test_functions.py::FunctionTests::test_device PASSED [ 17%] 2023-03-31T05:57:51.1698079Z dynamo/test_functions.py::FunctionTests::test_device_constant PASSED [ 17%] 2023-03-31T05:57:51.1698401Z dynamo/test_functions.py::FunctionTests::test_dict_copy PASSED [ 18%] 2023-03-31T05:57:51.1698714Z dynamo/test_functions.py::FunctionTests::test_dict_ops PASSED [ 19%] 2023-03-31T05:57:51.1699032Z dynamo/test_functions.py::FunctionTests::test_dict_param_keys PASSED [ 20%] 2023-03-31T05:57:51.1699342Z dynamo/test_functions.py::FunctionTests::test_dict_sorted PASSED [ 21%] 2023-03-31T05:57:51.1699687Z dynamo/test_functions.py::FunctionTests::test_distributed_is_available PASSED [ 21%] 2023-03-31T05:57:51.1700089Z dynamo/test_functions.py::FunctionTests::test_distributed_is_initialized PASSED [ 22%] 2023-03-31T05:57:51.1700417Z dynamo/test_functions.py::FunctionTests::test_dtype PASSED [ 23%] 2023-03-31T05:57:51.1700736Z dynamo/test_functions.py::FunctionTests::test_dtype_compare PASSED [ 24%] 2023-03-31T05:57:51.1701055Z dynamo/test_functions.py::FunctionTests::test_finfo PASSED [ 25%] 2023-03-31T05:57:51.1701365Z dynamo/test_functions.py::FunctionTests::test_float PASSED [ 25%] 2023-03-31T05:57:51.1701669Z dynamo/test_functions.py::FunctionTests::test_fn_with_self_set PASSED [ 26%] 2023-03-31T05:57:51.1701986Z dynamo/test_functions.py::FunctionTests::test_fstrings1 PASSED [ 27%] 2023-03-31T05:57:51.1702303Z dynamo/test_functions.py::FunctionTests::test_fstrings2 PASSED [ 28%] 2023-03-31T05:57:51.1702771Z dynamo/test_functions.py::FunctionTests::test_fstrings3 PASSED [ 28%] 2023-03-31T05:57:51.1703092Z dynamo/test_functions.py::FunctionTests::test_funcdef_closure PASSED [ 29%] 2023-03-31T05:57:51.1703422Z dynamo/test_functions.py::FunctionTests::test_get_default_dtype PASSED [ 30%] 2023-03-31T05:57:51.1703746Z dynamo/test_functions.py::FunctionTests::test_globalfn PASSED [ 31%] 2023-03-31T05:57:51.1704052Z dynamo/test_functions.py::FunctionTests::test_globalmodule PASSED [ 32%] 2023-03-31T05:57:51.1704375Z dynamo/test_functions.py::FunctionTests::test_globalvar PASSED [ 32%] 2023-03-31T05:57:51.1704688Z dynamo/test_functions.py::FunctionTests::test_import1 PASSED [ 33%] 2023-03-31T05:57:51.1704985Z dynamo/test_functions.py::FunctionTests::test_indirect1 PASSED [ 34%] 2023-03-31T05:57:51.1705301Z dynamo/test_functions.py::FunctionTests::test_indirect2 PASSED [ 35%] 2023-03-31T05:57:51.1705616Z dynamo/test_functions.py::FunctionTests::test_indirect3 PASSED [ 35%] 2023-03-31T05:57:51.1705950Z dynamo/test_functions.py::FunctionTests::test_inline_jit_annotations PASSED [ 36%] 2023-03-31T05:57:51.1706276Z dynamo/test_functions.py::FunctionTests::test_inline_softmax PASSED [ 37%] 2023-03-31T05:57:51.1706599Z dynamo/test_functions.py::FunctionTests::test_inline_with_default PASSED [ 38%] 2023-03-31T05:57:51.1706922Z dynamo/test_functions.py::FunctionTests::test_inner_function PASSED [ 39%] 2023-03-31T05:57:51.1707293Z dynamo/test_functions.py::FunctionTests::test_is_contiguous_memory_format PASSED [ 39%] 2023-03-31T05:57:51.1707664Z dynamo/test_functions.py::FunctionTests::test_is_fx_tracing PASSED [ 40%] 2023-03-31T05:57:51.1707988Z dynamo/test_functions.py::FunctionTests::test_is_in_onnx_export PASSED [ 41%] 2023-03-31T05:57:51.1708308Z dynamo/test_functions.py::FunctionTests::test_is_not_null PASSED [ 42%] 2023-03-31T05:57:51.1708611Z dynamo/test_functions.py::FunctionTests::test_is_quantized PASSED [ 42%] 2023-03-31T05:57:51.1708931Z dynamo/test_functions.py::FunctionTests::test_is_sparse PASSED [ 43%] 2023-03-31T05:57:51.1709247Z dynamo/test_functions.py::FunctionTests::test_islice_chain PASSED [ 44%] 2023-03-31T05:57:51.1709552Z dynamo/test_functions.py::FunctionTests::test_jit_annotate PASSED [ 45%] 2023-03-31T05:57:51.1709875Z dynamo/test_functions.py::FunctionTests::test_len_constant_dict PASSED [ 46%] 2023-03-31T05:57:51.1710194Z dynamo/test_functions.py::FunctionTests::test_len_constant_list PASSED [ 46%] 2023-03-31T05:57:51.1710536Z dynamo/test_functions.py::FunctionTests::test_len_constant_misc_iterables PASSED [ 47%] 2023-03-31T05:57:51.1710862Z dynamo/test_functions.py::FunctionTests::test_len_tensor PASSED [ 48%] 2023-03-31T05:57:51.1711172Z dynamo/test_functions.py::FunctionTests::test_list_add PASSED [ 49%] 2023-03-31T05:57:51.1711484Z dynamo/test_functions.py::FunctionTests::test_list_clear PASSED [ 50%] 2023-03-31T05:57:51.1711791Z dynamo/test_functions.py::FunctionTests::test_list_convert PASSED [ 50%] 2023-03-31T05:57:51.1712149Z dynamo/test_functions.py::FunctionTests::test_list_reversed PASSED [ 51%] 2023-03-31T05:57:51.1712483Z dynamo/test_functions.py::FunctionTests::test_list_slice_assignment PASSED [ 52%] 2023-03-31T05:57:51.1712811Z dynamo/test_functions.py::FunctionTests::test_list_sorted1 PASSED [ 53%] 2023-03-31T05:57:51.1713114Z dynamo/test_functions.py::FunctionTests::test_list_sorted2 PASSED [ 53%] 2023-03-31T05:57:51.1713432Z dynamo/test_functions.py::FunctionTests::test_list_truth PASSED [ 54%] 2023-03-31T05:57:51.1713745Z dynamo/test_functions.py::FunctionTests::test_listarg1 PASSED [ 55%] 2023-03-31T05:57:51.1714043Z dynamo/test_functions.py::FunctionTests::test_listarg2 PASSED [ 56%] 2023-03-31T05:57:51.1714351Z dynamo/test_functions.py::FunctionTests::test_listarg3 PASSED [ 57%] 2023-03-31T05:57:51.1714662Z dynamo/test_functions.py::FunctionTests::test_listarg4 PASSED [ 57%] 2023-03-31T05:57:51.1715014Z dynamo/test_functions.py::FunctionTests::test_listarg5 PASSED [ 58%] 2023-03-31T05:57:51.1715322Z dynamo/test_functions.py::FunctionTests::test_load_global_bool PASSED [ 59%] 2023-03-31T05:57:51.1715630Z dynamo/test_functions.py::FunctionTests::test_mT PASSED [ 60%] 2023-03-31T05:57:51.1715934Z dynamo/test_functions.py::FunctionTests::test_map_sum PASSED [ 60%] 2023-03-31T05:57:51.1716233Z dynamo/test_functions.py::FunctionTests::test_methodcall1 PASSED [ 61%] 2023-03-31T05:57:51.1716553Z dynamo/test_functions.py::FunctionTests::test_methodcall2 PASSED [ 62%] 2023-03-31T05:57:51.1716872Z dynamo/test_functions.py::FunctionTests::test_methodcall3 PASSED [ 63%] 2023-03-31T05:57:51.1717181Z dynamo/test_functions.py::FunctionTests::test_min_max PASSED [ 64%] 2023-03-31T05:57:51.1717483Z dynamo/test_functions.py::FunctionTests::test_module_constant PASSED [ 64%] 2023-03-31T05:57:51.1717804Z dynamo/test_functions.py::FunctionTests::test_namedtuple PASSED [ 65%] 2023-03-31T05:57:51.1718118Z dynamo/test_functions.py::FunctionTests::test_ndim PASSED [ 66%] 2023-03-31T05:57:51.1718412Z dynamo/test_functions.py::FunctionTests::test_pop PASSED [ 67%] 2023-03-31T05:57:51.1718717Z dynamo/test_functions.py::FunctionTests::test_range1 PASSED [ 67%] 2023-03-31T05:57:51.1719019Z dynamo/test_functions.py::FunctionTests::test_range2 PASSED [ 68%] 2023-03-31T05:57:51.1719362Z dynamo/test_functions.py::FunctionTests::test_reduce PASSED [ 69%] 2023-03-31T05:57:51.1719699Z dynamo/test_functions.py::FunctionTests::test_return_dict PASSED [ 70%] 2023-03-31T05:57:51.1720010Z dynamo/test_functions.py::FunctionTests::test_return_dict2 PASSED [ 71%] 2023-03-31T05:57:51.1720327Z dynamo/test_functions.py::FunctionTests::test_return_tuple1 PASSED [ 71%] 2023-03-31T05:57:51.1720638Z dynamo/test_functions.py::FunctionTests::test_return_tuple2 PASSED [ 72%] 2023-03-31T05:57:51.1720950Z dynamo/test_functions.py::FunctionTests::test_shape1 PASSED [ 73%] 2023-03-31T05:57:51.1721263Z dynamo/test_functions.py::FunctionTests::test_shape2 PASSED [ 74%] 2023-03-31T05:57:51.1721572Z dynamo/test_functions.py::FunctionTests::test_slice1 PASSED [ 75%] 2023-03-31T05:57:51.1721867Z dynamo/test_functions.py::FunctionTests::test_slice2 PASSED [ 75%] 2023-03-31T05:57:51.1722176Z dynamo/test_functions.py::FunctionTests::test_slice3 PASSED [ 76%] 2023-03-31T05:57:51.1722482Z dynamo/test_functions.py::FunctionTests::test_slice4 PASSED [ 77%] 2023-03-31T05:57:51.1722778Z dynamo/test_functions.py::FunctionTests::test_slice5 PASSED [ 78%] 2023-03-31T05:57:51.1723208Z dynamo/test_functions.py::FunctionTests::test_slice6 PASSED [ 78%] 2023-03-31T05:57:51.1723520Z dynamo/test_functions.py::FunctionTests::test_startswith PASSED [ 79%] 2023-03-31T05:57:51.1723839Z dynamo/test_functions.py::FunctionTests::test_tensor_len PASSED [ 80%] 2023-03-31T05:57:51.1724226Z dynamo/test_functions.py::FunctionTests::test_tensor_new_with_shape PASSED [ 81%] 2023-03-31T05:57:51.1724565Z dynamo/test_functions.py::FunctionTests::test_tensor_new_with_size PASSED [ 82%] 2023-03-31T05:57:51.1724891Z dynamo/test_functions.py::FunctionTests::test_tensor_type PASSED [ 82%] 2023-03-31T05:57:51.1725220Z dynamo/test_functions.py::FunctionTests::test_tensor_type2 SKIPPED (requires cuda) [ 83%] 2023-03-31T05:57:51.1725563Z dynamo/test_functions.py::FunctionTests::test_tensor_type3 PASSED [ 84%] 2023-03-31T05:57:51.1725886Z dynamo/test_functions.py::FunctionTests::test_tensor_type4 PASSED [ 85%] 2023-03-31T05:57:51.1726225Z dynamo/test_functions.py::FunctionTests::test_tensor_type5 SKIPPED (requires cuda) [ 85%] 2023-03-31T05:57:51.1726572Z dynamo/test_functions.py::FunctionTests::test_torch_distributions_functions PASSED [ 86%] 2023-03-31T05:57:51.1726963Z dynamo/test_functions.py::FunctionTests::test_transpose_for_scores PASSED [ 87%] 2023-03-31T05:57:51.1727293Z dynamo/test_functions.py::FunctionTests::test_tuple1 PASSED [ 88%] 2023-03-31T05:57:51.1727596Z dynamo/test_functions.py::FunctionTests::test_tuple2 PASSED [ 89%] 2023-03-31T05:57:51.1727908Z dynamo/test_functions.py::FunctionTests::test_tuple_contains PASSED [ 89%] 2023-03-31T05:57:51.1728228Z dynamo/test_functions.py::FunctionTests::test_tuple_iadd PASSED [ 90%] 2023-03-31T05:57:51.1728548Z dynamo/test_functions.py::FunctionTests::test_tuple_sorted PASSED [ 91%] 2023-03-31T05:57:51.1728854Z dynamo/test_functions.py::FunctionTests::test_unpack1 PASSED [ 92%] 2023-03-31T05:57:51.1729161Z dynamo/test_functions.py::FunctionTests::test_unpack2 PASSED [ 92%] 2023-03-31T05:57:51.1729471Z dynamo/test_functions.py::FunctionTests::test_unpack3 PASSED [ 93%] 2023-03-31T05:57:51.1729772Z dynamo/test_functions.py::FunctionTests::test_unpack_ex1 PASSED [ 94%] 2023-03-31T05:57:51.1730083Z dynamo/test_functions.py::FunctionTests::test_unpack_ex2 PASSED [ 95%] 2023-03-31T05:57:51.1730394Z dynamo/test_functions.py::FunctionTests::test_unpack_ex3 PASSED [ 96%] 2023-03-31T05:57:51.1730703Z dynamo/test_functions.py::FunctionTests::test_viamethod PASSED [ 96%] 2023-03-31T05:57:51.1731004Z dynamo/test_functions.py::FunctionTests::test_viatorch PASSED [ 97%] 2023-03-31T05:57:51.1731850Z dynamo/test_functions.py::DefaultsTests::test_func_default_tensor_args [2023-03-31 05:57:50,610] torch._dynamo.symbolic_convert: [WARNING] dynamo/test_functions.py [] {} __init__() got an unexpected keyword argument 'guards' 2023-03-31T05:57:51.1732330Z PASSED [ 98%] 2023-03-31T05:57:51.1732596Z dynamo/test_functions.py::DefaultsTests::test_func_default_torch_args PASSED [ 99%] 2023-03-31T05:57:51.1733416Z dynamo/test_functions.py::DefaultsTests::test_meth_default_tensor_args [2023-03-31 05:57:50,643] torch._dynamo.symbolic_convert: [WARNING] dynamo/test_functions.py [NNModuleVariable()] {} __init__() got an unexpected keyword argument 'guards' 2023-03-31T05:57:51.1733916Z PASSED [100%] 2023-03-31T05:57:51.1734028Z 2023-03-31T05:57:51.1734392Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-85421862794e1580.xml - 2023-03-31T05:57:51.1734775Z ======================== 126 passed, 2 skipped in 1.76s ======================== 2023-03-31T05:57:51.1735090Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:51.1735295Z 2023-03-31T05:57:51.1735603Z ##[endgroup] 2023-03-31T05:57:51.1736033Z FINISHED PRINTING LOG FILE of dynamo/test_functions (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_functions_r0p_nj4k.log) 2023-03-31T05:57:51.1736280Z 2023-03-31T05:57:53.2586617Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:53.2616545Z Ignoring disabled issues: [] 2023-03-31T05:57:53.2789858Z Running test_python_dispatch ... [2023-03-31 05:57:53.278648] 2023-03-31T05:57:53.2791982Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_python_dispatch.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:53.278985] 2023-03-31T05:57:56.9607096Z 2023-03-31T05:57:56.9608415Z Expand the folded group to see the log file of test_python_dispatch 2023-03-31T05:57:56.9609505Z ##[group]PRINTING LOG FILE of test_python_dispatch (/var/lib/jenkins/workspace/test/test-reports/test_python_dispatch_h99_luku.log) 2023-03-31T05:57:56.9610510Z Test results will be stored in test-reports/python-pytest/test_python_dispatch/test_python_dispatch-0ca79c911b8030a4.xml 2023-03-31T05:57:56.9611115Z ============================= test session starts ============================== 2023-03-31T05:57:56.9611672Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:57:56.9612094Z cachedir: .pytest_cache 2023-03-31T05:57:56.9612827Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:57:56.9613478Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:57:56.9614121Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:57:56.9614589Z collecting ... collected 73 items 2023-03-31T05:57:56.9622135Z Running 73 items in this shard: test/test_python_dispatch.py::TestPythonRegistration::test_alias_analysis, test/test_python_dispatch.py::TestPythonRegistration::test_create_new_library, test/test_python_dispatch.py::TestPythonRegistration::test_error_for_unsupported_ns_or_kind, test/test_python_dispatch.py::TestPythonRegistration::test_error_if_fn_not_callable, test/test_python_dispatch.py::TestPythonRegistration::test_extend_library_with_dispatch_key_arg, test/test_python_dispatch.py::TestPythonRegistration::test_override_aten_ops_with_multiple_libraries, test/test_python_dispatch.py::TestPythonRegistration::test_override_cpu_sum, test/test_python_dispatch.py::TestPythonRegistration::test_override_cuda_with_jiterator, test/test_python_dispatch.py::TestPythonRegistration::test_returning_symint, test/test_python_dispatch.py::TestPythonDispatch::test_all_same_mode, test/test_python_dispatch.py::TestPythonDispatch::test_autograd_in_attr, test/test_python_dispatch.py::TestPythonDispatch::test_basic, test/test_python_dispatch.py::TestPythonDispatch::test_capture_logs_with_torch_dispatch_mode, test/test_python_dispatch.py::TestPythonDispatch::test_construct_int_tensor, test/test_python_dispatch.py::TestPythonDispatch::test_custom_autograd, test/test_python_dispatch.py::TestPythonDispatch::test_deepcopy_non_wrapper_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_deepcopy_wrapper_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_deepcopy_wrapper_subclass_with_clone_returning_different_type, test/test_python_dispatch.py::TestPythonDispatch::test_detach_appears_twice_when_called_once, test/test_python_dispatch.py::TestPythonDispatch::test_device_slowpath, test/test_python_dispatch.py::TestPythonDispatch::test_dim_slowpath, test/test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_call, test/test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_call_list_arg, test/test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_dont_autograd, test/test_python_dispatch.py::TestPythonDispatch::test_error_using_class_method_on_mode, test/test_python_dispatch.py::TestPythonDispatch::test_exception_handling, test/test_python_dispatch.py::TestPythonDispatch::test_fancy_strides, test/test_python_dispatch.py::TestPythonDispatch::test_format, test/test_python_dispatch.py::TestPythonDispatch::test_get_cur_mode, test/test_python_dispatch.py::TestPythonDispatch::test_get_mode_stack, test/test_python_dispatch.py::TestPythonDispatch::test_index_put_where_only_index_is_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_invalid_ret, test/test_python_dispatch.py::TestPythonDispatch::test_is_contiguous_slow_path, test/test_python_dispatch.py::TestPythonDispatch::test_kwarg_only, test/test_python_dispatch.py::TestPythonDispatch::test_kwarg_only_and_positional_default, test/test_python_dispatch.py::TestPythonDispatch::test_layout_slow_path, test/test_python_dispatch.py::TestPythonDispatch::test_like, test/test_python_dispatch.py::TestPythonDispatch::test_list_ret, test/test_python_dispatch.py::TestPythonDispatch::test_make_subclass_with_modes, test/test_python_dispatch.py::TestPythonDispatch::test_make_wrapper_subclass_noalloc, test/test_python_dispatch.py::TestPythonDispatch::test_make_wrapper_subclass_propagates_metadata, test/test_python_dispatch.py::TestPythonDispatch::test_maybe_tuple_bug, test/test_python_dispatch.py::TestPythonDispatch::test_mode_with_make_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_multiple_ops_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_nested_push_logging_tensor_mode, test/test_python_dispatch.py::TestPythonDispatch::test_nesting_same_mode, test/test_python_dispatch.py::TestPythonDispatch::test_new_ones, test/test_python_dispatch.py::TestPythonDispatch::test_none_wrapping, test/test_python_dispatch.py::TestPythonDispatch::test_notimplemented_mode, test/test_python_dispatch.py::TestPythonDispatch::test_optional_tensor_list, test/test_python_dispatch.py::TestPythonDispatch::test_out, test/test_python_dispatch.py::TestPythonDispatch::test_produce_real_type, test/test_python_dispatch.py::TestPythonDispatch::test_set_data, test/test_python_dispatch.py::TestPythonDispatch::test_shallow_copy_and_detach, test/test_python_dispatch.py::TestPythonDispatch::test_sizes_slow_path, test/test_python_dispatch.py::TestPythonDispatch::test_standard_is_not_subclass, test/test_python_dispatch.py::TestPythonDispatch::test_storage, test/test_python_dispatch.py::TestPythonDispatch::test_storage_can_be_converted_to_python_object, test/test_python_dispatch.py::TestPythonDispatch::test_strides_slow_path, test/test_python_dispatch.py::TestPythonDispatch::test_subclass_autograd_device_check, test/test_python_dispatch.py::TestPythonDispatch::test_subclass_creation, test/test_python_dispatch.py::TestPythonDispatch::test_subclass_priority, test/test_python_dispatch.py::TestPythonDispatch::test_tolist_numpy_with_torch_dispatch_mode, test/test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_basic, test/test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_respects_no_dispatch, test/test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_subclass_priority, test/test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_unrelated_tensors, test/test_python_dispatch.py::TestPythonDispatch::test_version, test/test_python_dispatch.py::TestPythonDispatch::test_with_mode_created_separately, test/test_python_dispatch.py::TestPythonDispatch::test_with_nested_modes, test/test_python_dispatch.py::TestPythonDispatch::test_wrapper_subclass_serializes, test/test_python_dispatch.py::TestPythonDispatcher::test_basic, test/test_python_dispatch.py::TestPythonDispatcher::test_lstsq 2023-03-31T05:57:56.9629004Z 2023-03-31T05:57:56.9629189Z test_python_dispatch.py::TestPythonRegistration::test_alias_analysis PASSED [ 1%] 2023-03-31T05:57:56.9629564Z test_python_dispatch.py::TestPythonRegistration::test_create_new_library PASSED [ 2%] 2023-03-31T05:57:56.9629959Z test_python_dispatch.py::TestPythonRegistration::test_error_for_unsupported_ns_or_kind PASSED [ 4%] 2023-03-31T05:57:56.9630360Z test_python_dispatch.py::TestPythonRegistration::test_error_if_fn_not_callable PASSED [ 5%] 2023-03-31T05:57:56.9630751Z test_python_dispatch.py::TestPythonRegistration::test_extend_library_with_dispatch_key_arg PASSED [ 6%] 2023-03-31T05:57:56.9631174Z test_python_dispatch.py::TestPythonRegistration::test_override_aten_ops_with_multiple_libraries PASSED [ 8%] 2023-03-31T05:57:56.9631652Z test_python_dispatch.py::TestPythonRegistration::test_override_cpu_sum PASSED [ 9%] 2023-03-31T05:57:56.9632157Z test_python_dispatch.py::TestPythonRegistration::test_override_cuda_with_jiterator PASSED [ 10%] 2023-03-31T05:57:56.9632528Z test_python_dispatch.py::TestPythonRegistration::test_returning_symint PASSED [ 12%] 2023-03-31T05:57:56.9632903Z test_python_dispatch.py::TestPythonDispatch::test_all_same_mode PASSED [ 13%] 2023-03-31T05:57:56.9633333Z test_python_dispatch.py::TestPythonDispatch::test_autograd_in_attr PASSED [ 15%] 2023-03-31T05:57:56.9633901Z test_python_dispatch.py::TestPythonDispatch::test_basic PASSED [ 16%] 2023-03-31T05:57:56.9634484Z test_python_dispatch.py::TestPythonDispatch::test_capture_logs_with_torch_dispatch_mode PASSED [ 17%] 2023-03-31T05:57:56.9635182Z test_python_dispatch.py::TestPythonDispatch::test_construct_int_tensor PASSED [ 19%] 2023-03-31T05:57:56.9635670Z test_python_dispatch.py::TestPythonDispatch::test_custom_autograd PASSED [ 20%] 2023-03-31T05:57:56.9636270Z test_python_dispatch.py::TestPythonDispatch::test_deepcopy_non_wrapper_subclass PASSED [ 21%] 2023-03-31T05:57:56.9636906Z test_python_dispatch.py::TestPythonDispatch::test_deepcopy_wrapper_subclass PASSED [ 23%] 2023-03-31T05:57:56.9637678Z test_python_dispatch.py::TestPythonDispatch::test_deepcopy_wrapper_subclass_with_clone_returning_different_type PASSED [ 24%] 2023-03-31T05:57:56.9638461Z test_python_dispatch.py::TestPythonDispatch::test_detach_appears_twice_when_called_once PASSED [ 26%] 2023-03-31T05:57:56.9639135Z test_python_dispatch.py::TestPythonDispatch::test_device_slowpath PASSED [ 27%] 2023-03-31T05:57:56.9639756Z test_python_dispatch.py::TestPythonDispatch::test_dim_slowpath PASSED [ 28%] 2023-03-31T05:57:56.9640411Z test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_call PASSED [ 30%] 2023-03-31T05:57:56.9640843Z test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_call_list_arg PASSED [ 31%] 2023-03-31T05:57:56.9641261Z test_python_dispatch.py::TestPythonDispatch::test_dispatch_super_dont_autograd PASSED [ 32%] 2023-03-31T05:57:56.9641643Z test_python_dispatch.py::TestPythonDispatch::test_error_using_class_method_on_mode PASSED [ 34%] 2023-03-31T05:57:56.9642014Z test_python_dispatch.py::TestPythonDispatch::test_exception_handling PASSED [ 35%] 2023-03-31T05:57:56.9642428Z test_python_dispatch.py::TestPythonDispatch::test_fancy_strides PASSED [ 36%] 2023-03-31T05:57:56.9642808Z test_python_dispatch.py::TestPythonDispatch::test_format PASSED [ 38%] 2023-03-31T05:57:56.9643460Z test_python_dispatch.py::TestPythonDispatch::test_get_cur_mode PASSED [ 39%] 2023-03-31T05:57:56.9643947Z test_python_dispatch.py::TestPythonDispatch::test_get_mode_stack PASSED [ 41%] 2023-03-31T05:57:56.9644619Z test_python_dispatch.py::TestPythonDispatch::test_index_put_where_only_index_is_subclass PASSED [ 42%] 2023-03-31T05:57:56.9645193Z test_python_dispatch.py::TestPythonDispatch::test_invalid_ret PASSED [ 43%] 2023-03-31T05:57:56.9645571Z test_python_dispatch.py::TestPythonDispatch::test_is_contiguous_slow_path PASSED [ 45%] 2023-03-31T05:57:56.9645989Z test_python_dispatch.py::TestPythonDispatch::test_kwarg_only PASSED [ 46%] 2023-03-31T05:57:56.9654177Z test_python_dispatch.py::TestPythonDispatch::test_kwarg_only_and_positional_default PASSED [ 47%] 2023-03-31T05:57:56.9654580Z test_python_dispatch.py::TestPythonDispatch::test_layout_slow_path PASSED [ 49%] 2023-03-31T05:57:56.9654932Z test_python_dispatch.py::TestPythonDispatch::test_like PASSED [ 50%] 2023-03-31T05:57:56.9655254Z test_python_dispatch.py::TestPythonDispatch::test_list_ret PASSED [ 52%] 2023-03-31T05:57:56.9655588Z test_python_dispatch.py::TestPythonDispatch::test_make_subclass_with_modes PASSED [ 53%] 2023-03-31T05:57:56.9655949Z test_python_dispatch.py::TestPythonDispatch::test_make_wrapper_subclass_noalloc PASSED [ 54%] 2023-03-31T05:57:56.9656341Z test_python_dispatch.py::TestPythonDispatch::test_make_wrapper_subclass_propagates_metadata PASSED [ 56%] 2023-03-31T05:57:56.9656820Z test_python_dispatch.py::TestPythonDispatch::test_maybe_tuple_bug PASSED [ 57%] 2023-03-31T05:57:56.9657182Z test_python_dispatch.py::TestPythonDispatch::test_mode_with_make_subclass PASSED [ 58%] 2023-03-31T05:57:56.9657548Z test_python_dispatch.py::TestPythonDispatch::test_multiple_ops_subclass PASSED [ 60%] 2023-03-31T05:57:56.9657929Z test_python_dispatch.py::TestPythonDispatch::test_nested_push_logging_tensor_mode PASSED [ 61%] 2023-03-31T05:57:56.9658288Z test_python_dispatch.py::TestPythonDispatch::test_nesting_same_mode PASSED [ 63%] 2023-03-31T05:57:56.9658631Z test_python_dispatch.py::TestPythonDispatch::test_new_ones PASSED [ 64%] 2023-03-31T05:57:56.9658972Z test_python_dispatch.py::TestPythonDispatch::test_none_wrapping PASSED [ 65%] 2023-03-31T05:57:56.9659318Z test_python_dispatch.py::TestPythonDispatch::test_notimplemented_mode PASSED [ 67%] 2023-03-31T05:57:56.9659733Z test_python_dispatch.py::TestPythonDispatch::test_optional_tensor_list PASSED [ 68%] 2023-03-31T05:57:56.9660080Z test_python_dispatch.py::TestPythonDispatch::test_out PASSED [ 69%] 2023-03-31T05:57:56.9660422Z test_python_dispatch.py::TestPythonDispatch::test_produce_real_type PASSED [ 71%] 2023-03-31T05:57:56.9660752Z test_python_dispatch.py::TestPythonDispatch::test_set_data PASSED [ 72%] 2023-03-31T05:57:56.9661104Z test_python_dispatch.py::TestPythonDispatch::test_shallow_copy_and_detach PASSED [ 73%] 2023-03-31T05:57:56.9661462Z test_python_dispatch.py::TestPythonDispatch::test_sizes_slow_path PASSED [ 75%] 2023-03-31T05:57:56.9661820Z test_python_dispatch.py::TestPythonDispatch::test_standard_is_not_subclass PASSED [ 76%] 2023-03-31T05:57:56.9662159Z test_python_dispatch.py::TestPythonDispatch::test_storage PASSED [ 78%] 2023-03-31T05:57:56.9662644Z test_python_dispatch.py::TestPythonDispatch::test_storage_can_be_converted_to_python_object PASSED [ 79%] 2023-03-31T05:57:56.9663023Z test_python_dispatch.py::TestPythonDispatch::test_strides_slow_path PASSED [ 80%] 2023-03-31T05:57:56.9663381Z test_python_dispatch.py::TestPythonDispatch::test_subclass_autograd_device_check PASSED [ 82%] 2023-03-31T05:57:56.9663749Z test_python_dispatch.py::TestPythonDispatch::test_subclass_creation PASSED [ 83%] 2023-03-31T05:57:56.9664105Z test_python_dispatch.py::TestPythonDispatch::test_subclass_priority PASSED [ 84%] 2023-03-31T05:57:56.9664547Z test_python_dispatch.py::TestPythonDispatch::test_tolist_numpy_with_torch_dispatch_mode PASSED [ 86%] 2023-03-31T05:57:56.9664957Z test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_basic PASSED [ 87%] 2023-03-31T05:57:56.9665350Z test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_respects_no_dispatch PASSED [ 89%] 2023-03-31T05:57:56.9665753Z test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_subclass_priority PASSED [ 90%] 2023-03-31T05:57:56.9666140Z test_python_dispatch.py::TestPythonDispatch::test_torch_dispatch_mode_unrelated_tensors PASSED [ 91%] 2023-03-31T05:57:56.9666508Z test_python_dispatch.py::TestPythonDispatch::test_version PASSED [ 93%] 2023-03-31T05:57:56.9666869Z test_python_dispatch.py::TestPythonDispatch::test_with_mode_created_separately PASSED [ 94%] 2023-03-31T05:57:56.9667237Z test_python_dispatch.py::TestPythonDispatch::test_with_nested_modes PASSED [ 95%] 2023-03-31T05:57:56.9667591Z test_python_dispatch.py::TestPythonDispatch::test_wrapper_subclass_serializes PASSED [ 97%] 2023-03-31T05:57:56.9667957Z test_python_dispatch.py::TestPythonDispatcher::test_basic PASSED [ 98%] 2023-03-31T05:57:56.9668300Z test_python_dispatch.py::TestPythonDispatcher::test_lstsq PASSED [100%] 2023-03-31T05:57:56.9668489Z 2023-03-31T05:57:56.9668944Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_python_dispatch/test_python_dispatch-0ca79c911b8030a4.xml - 2023-03-31T05:57:56.9671152Z ============================== 73 passed in 1.63s ============================== 2023-03-31T05:57:56.9671799Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:57:56.9672079Z 2023-03-31T05:57:56.9672456Z ##[endgroup] 2023-03-31T05:57:56.9673073Z FINISHED PRINTING LOG FILE of test_python_dispatch (/var/lib/jenkins/workspace/test/test-reports/test_python_dispatch_h99_luku.log) 2023-03-31T05:57:56.9673481Z 2023-03-31T05:57:59.0713387Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:57:59.0740839Z Ignoring disabled issues: [] 2023-03-31T05:57:59.0916193Z Running test_shape_ops ... [2023-03-31 05:57:59.091190] 2023-03-31T05:57:59.0918453Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_shape_ops.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:57:59.091585] 2023-03-31T05:58:02.6504021Z 2023-03-31T05:58:02.6504801Z Expand the folded group to see the log file of test_shape_ops 2023-03-31T05:58:02.6505696Z ##[group]PRINTING LOG FILE of test_shape_ops (/var/lib/jenkins/workspace/test/test-reports/test_shape_ops_5fiszwi9.log) 2023-03-31T05:58:02.6506729Z Test results will be stored in test-reports/python-pytest/test_shape_ops/test_shape_ops-9d35fd1a78d8eba1.xml 2023-03-31T05:58:02.6510882Z ============================= test session starts ============================== 2023-03-31T05:58:02.6511315Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:02.6511603Z cachedir: .pytest_cache 2023-03-31T05:58:02.6512045Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:02.6512410Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:02.6512862Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:02.6513180Z collecting ... collected 92 items 2023-03-31T05:58:02.6520826Z Running 92 items in this shard: test/test_shape_ops.py::TestShapeOpsCPU::test_clamp_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_clamp_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_clamp_propagates_nans_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_clamp_raises_arg_errors_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_complex_rot90_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_complex_rot90_cpu_complex64, test/test_shape_ops.py::TestShapeOpsCPU::test_diag_cpu_bool, test/test_shape_ops.py::TestShapeOpsCPU::test_diag_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_diagonal_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_diagonal_multidim_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_bfloat16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_bool, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_complex64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int8, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_uint8, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_bfloat16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_bool, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_complex64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int8, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_uint8, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_large_tensor_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_bfloat16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_bool, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_complex64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int16, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int32, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int8, test/test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_uint8, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_astuple_out_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_bfloat16, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_bool, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float16, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int16, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int32, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int8, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_uint8, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_discontiguous_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_no_warning_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_nonzero_non_diff_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_rot90_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_complex128, test/test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_tolist_cpu, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_float32, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_float64, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int16, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int32, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int64, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int8, test/test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_uint8, test/test_shape_ops.py::TestShapeOpsCPU::test_unbind_cpu 2023-03-31T05:58:02.6528143Z 2023-03-31T05:58:02.6528299Z test_shape_ops.py::TestShapeOpsCPU::test_clamp_cpu_float32 PASSED [ 1%] 2023-03-31T05:58:02.6528639Z test_shape_ops.py::TestShapeOpsCPU::test_clamp_cpu_int64 PASSED [ 2%] 2023-03-31T05:58:02.6529044Z test_shape_ops.py::TestShapeOpsCPU::test_clamp_propagates_nans_cpu PASSED [ 3%] 2023-03-31T05:58:02.6529397Z test_shape_ops.py::TestShapeOpsCPU::test_clamp_raises_arg_errors_cpu PASSED [ 4%] 2023-03-31T05:58:02.6529754Z test_shape_ops.py::TestShapeOpsCPU::test_complex_rot90_cpu_complex128 PASSED [ 5%] 2023-03-31T05:58:02.6530090Z test_shape_ops.py::TestShapeOpsCPU::test_complex_rot90_cpu_complex64 PASSED [ 6%] 2023-03-31T05:58:02.6530427Z test_shape_ops.py::TestShapeOpsCPU::test_diag_cpu_bool PASSED [ 7%] 2023-03-31T05:58:02.6530811Z test_shape_ops.py::TestShapeOpsCPU::test_diag_cpu_float32 PASSED [ 8%] 2023-03-31T05:58:02.6531128Z test_shape_ops.py::TestShapeOpsCPU::test_diagonal_cpu PASSED [ 9%] 2023-03-31T05:58:02.6531472Z test_shape_ops.py::TestShapeOpsCPU::test_diagonal_multidim_cpu_float32 PASSED [ 10%] 2023-03-31T05:58:02.6531810Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_bfloat16 PASSED [ 11%] 2023-03-31T05:58:02.6532134Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_bool PASSED [ 13%] 2023-03-31T05:58:02.6532450Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_complex128 PASSED [ 14%] 2023-03-31T05:58:02.6532775Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_complex64 PASSED [ 15%] 2023-03-31T05:58:02.6533096Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float16 PASSED [ 16%] 2023-03-31T05:58:02.6533426Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float32 PASSED [ 17%] 2023-03-31T05:58:02.6533837Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_float64 PASSED [ 18%] 2023-03-31T05:58:02.6534156Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int16 PASSED [ 19%] 2023-03-31T05:58:02.6534471Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int32 PASSED [ 20%] 2023-03-31T05:58:02.6534770Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int64 PASSED [ 21%] 2023-03-31T05:58:02.6535078Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_int8 PASSED [ 22%] 2023-03-31T05:58:02.6535448Z test_shape_ops.py::TestShapeOpsCPU::test_flip_cpu_uint8 PASSED [ 23%] 2023-03-31T05:58:02.6535789Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_bfloat16 PASSED [ 25%] 2023-03-31T05:58:02.6536115Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_bool PASSED [ 26%] 2023-03-31T05:58:02.6536449Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_complex128 PASSED [ 27%] 2023-03-31T05:58:02.6536795Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_complex64 PASSED [ 28%] 2023-03-31T05:58:02.6537118Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float16 PASSED [ 29%] 2023-03-31T05:58:02.6537445Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float32 PASSED [ 30%] 2023-03-31T05:58:02.6537768Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_float64 PASSED [ 31%] 2023-03-31T05:58:02.6538080Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int16 PASSED [ 32%] 2023-03-31T05:58:02.6538403Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int32 PASSED [ 33%] 2023-03-31T05:58:02.6538725Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int64 PASSED [ 34%] 2023-03-31T05:58:02.6539049Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_int8 PASSED [ 35%] 2023-03-31T05:58:02.6539358Z test_shape_ops.py::TestShapeOpsCPU::test_flip_errors_cpu_uint8 PASSED [ 36%] 2023-03-31T05:58:02.6539706Z test_shape_ops.py::TestShapeOpsCPU::test_flip_large_tensor_cpu SKIPPED (Only runs on cuda) [ 38%] 2023-03-31T05:58:02.6540063Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_bfloat16 PASSED [ 39%] 2023-03-31T05:58:02.6540430Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_bool PASSED [ 40%] 2023-03-31T05:58:02.6540771Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_complex128 PASSED [ 41%] 2023-03-31T05:58:02.6541107Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_complex64 PASSED [ 42%] 2023-03-31T05:58:02.6541441Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float16 PASSED [ 43%] 2023-03-31T05:58:02.6541760Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float32 PASSED [ 44%] 2023-03-31T05:58:02.6542084Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_float64 PASSED [ 45%] 2023-03-31T05:58:02.6542495Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int16 PASSED [ 46%] 2023-03-31T05:58:02.6542806Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int32 PASSED [ 47%] 2023-03-31T05:58:02.6543197Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int64 PASSED [ 48%] 2023-03-31T05:58:02.6543521Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_int8 PASSED [ 50%] 2023-03-31T05:58:02.6543845Z test_shape_ops.py::TestShapeOpsCPU::test_flip_numpy_cpu_uint8 PASSED [ 51%] 2023-03-31T05:58:02.6544157Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_complex128 PASSED [ 52%] 2023-03-31T05:58:02.6544478Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_float64 PASSED [ 53%] 2023-03-31T05:58:02.6544800Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_cpu_int64 PASSED [ 54%] 2023-03-31T05:58:02.6545126Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_complex128 PASSED [ 55%] 2023-03-31T05:58:02.6545479Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_float64 PASSED [ 56%] 2023-03-31T05:58:02.6545819Z test_shape_ops.py::TestShapeOpsCPU::test_fliplr_invalid_cpu_int64 PASSED [ 57%] 2023-03-31T05:58:02.6546152Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_complex128 PASSED [ 58%] 2023-03-31T05:58:02.6546466Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_float64 PASSED [ 59%] 2023-03-31T05:58:02.6546781Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_cpu_int64 PASSED [ 60%] 2023-03-31T05:58:02.6547120Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_complex128 PASSED [ 61%] 2023-03-31T05:58:02.6547452Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_float64 PASSED [ 63%] 2023-03-31T05:58:02.6547839Z test_shape_ops.py::TestShapeOpsCPU::test_flipud_invalid_cpu_int64 PASSED [ 64%] 2023-03-31T05:58:02.6548201Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_complex128 PASSED [ 65%] 2023-03-31T05:58:02.6548530Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_float32 PASSED [ 66%] 2023-03-31T05:58:02.6548841Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_cpu_int64 PASSED [ 67%] 2023-03-31T05:58:02.6549176Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_complex128 PASSED [ 68%] 2023-03-31T05:58:02.6549530Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_float32 PASSED [ 69%] 2023-03-31T05:58:02.6549861Z test_shape_ops.py::TestShapeOpsCPU::test_movedim_invalid_cpu_int64 PASSED [ 70%] 2023-03-31T05:58:02.6550201Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_astuple_out_cpu PASSED [ 71%] 2023-03-31T05:58:02.6550534Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_bfloat16 PASSED [ 72%] 2023-03-31T05:58:02.6550862Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_bool PASSED [ 73%] 2023-03-31T05:58:02.6551175Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float16 PASSED [ 75%] 2023-03-31T05:58:02.6551494Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float32 PASSED [ 76%] 2023-03-31T05:58:02.6551815Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_float64 PASSED [ 77%] 2023-03-31T05:58:02.6552121Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int16 PASSED [ 78%] 2023-03-31T05:58:02.6552440Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int32 PASSED [ 79%] 2023-03-31T05:58:02.6552793Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int64 PASSED [ 80%] 2023-03-31T05:58:02.6553111Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_int8 PASSED [ 81%] 2023-03-31T05:58:02.6553414Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_cpu_uint8 PASSED [ 82%] 2023-03-31T05:58:02.6553749Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_discontiguous_cpu PASSED [ 83%] 2023-03-31T05:58:02.6554088Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_no_warning_cpu PASSED [ 84%] 2023-03-31T05:58:02.6554401Z test_shape_ops.py::TestShapeOpsCPU::test_nonzero_non_diff_cpu PASSED [ 85%] 2023-03-31T05:58:02.6554719Z test_shape_ops.py::TestShapeOpsCPU::test_rot90_cpu PASSED [ 86%] 2023-03-31T05:58:02.6555050Z test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_complex128 PASSED [ 88%] 2023-03-31T05:58:02.6555440Z test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_float32 PASSED [ 89%] 2023-03-31T05:58:02.6555776Z test_shape_ops.py::TestShapeOpsCPU::test_sparse_dense_dim_cpu_int64 PASSED [ 90%] 2023-03-31T05:58:02.6556106Z test_shape_ops.py::TestShapeOpsCPU::test_tolist_cpu PASSED [ 91%] 2023-03-31T05:58:02.6556423Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_float32 PASSED [ 92%] 2023-03-31T05:58:02.6556731Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_float64 PASSED [ 93%] 2023-03-31T05:58:02.6557053Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int16 PASSED [ 94%] 2023-03-31T05:58:02.6557372Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int32 PASSED [ 95%] 2023-03-31T05:58:02.6557684Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int64 PASSED [ 96%] 2023-03-31T05:58:02.6557984Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_int8 PASSED [ 97%] 2023-03-31T05:58:02.6558295Z test_shape_ops.py::TestShapeOpsCPU::test_trace_cpu_uint8 PASSED [ 98%] 2023-03-31T05:58:02.6558608Z test_shape_ops.py::TestShapeOpsCPU::test_unbind_cpu PASSED [100%] 2023-03-31T05:58:02.6558790Z 2023-03-31T05:58:02.6559185Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_shape_ops/test_shape_ops-9d35fd1a78d8eba1.xml - 2023-03-31T05:58:02.6559564Z ======================== 91 passed, 1 skipped in 1.31s ========================= 2023-03-31T05:58:02.6559928Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:02.6560168Z 2023-03-31T05:58:02.6560493Z ##[endgroup] 2023-03-31T05:58:02.6560890Z FINISHED PRINTING LOG FILE of test_shape_ops (/var/lib/jenkins/workspace/test/test-reports/test_shape_ops_5fiszwi9.log) 2023-03-31T05:58:02.6561121Z 2023-03-31T05:58:04.7480186Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:04.7506917Z Ignoring disabled issues: [] 2023-03-31T05:58:04.7680558Z Running dynamo/test_backends ... [2023-03-31 05:58:04.767690] 2023-03-31T05:58:04.7683272Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_backends.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:04.768034] 2023-03-31T05:58:12.8155940Z 2023-03-31T05:58:12.8156495Z Expand the folded group to see the log file of dynamo/test_backends 2023-03-31T05:58:12.8157581Z ##[group]PRINTING LOG FILE of dynamo/test_backends (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_backends_d1qug9y2.log) 2023-03-31T05:58:12.8158451Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:12.8164558Z Test results will be stored in test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-cd2960c4786ef00b.xml 2023-03-31T05:58:12.8165335Z ============================= test session starts ============================== 2023-03-31T05:58:12.8166004Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:12.8166452Z cachedir: .pytest_cache 2023-03-31T05:58:12.8167435Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:12.8168076Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:12.8168854Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:12.8169374Z collecting ... collected 18 items 2023-03-31T05:58:12.8172062Z Running 18 items in this shard: test/dynamo/test_backends.py::TestOptimizations::test_aot_cudagraphs, test/dynamo/test_backends.py::TestOptimizations::test_aot_eager, test/dynamo/test_backends.py::TestOptimizations::test_aot_eager_decomp_partition, test/dynamo/test_backends.py::TestOptimizations::test_aot_ts, test/dynamo/test_backends.py::TestOptimizations::test_aot_ts_nvfuser, test/dynamo/test_backends.py::TestOptimizations::test_eager, test/dynamo/test_backends.py::TestOptimizations::test_example_inputs, test/dynamo/test_backends.py::TestOptimizations::test_example_inputs_runtime_use, test/dynamo/test_backends.py::TestOptimizations::test_ipex_bf16, test/dynamo/test_backends.py::TestOptimizations::test_ipex_fp32, test/dynamo/test_backends.py::TestOptimizations::test_list_backends, test/dynamo/test_backends.py::TestOptimizations::test_nvprims_aten, test/dynamo/test_backends.py::TestOptimizations::test_nvprims_nvfuser, test/dynamo/test_backends.py::TestOptimizations::test_onnxrt, test/dynamo/test_backends.py::TestOptimizations::test_torchscript, test/dynamo/test_backends.py::TestOptimizations::test_tvm, test/dynamo/test_backends.py::NormalizeIRTests::test_inplace_normalize, test/dynamo/test_backends.py::MPSNotSupportedTest::test_default_mps_to_aot_eager 2023-03-31T05:58:12.8174888Z 2023-03-31T05:58:12.8175221Z dynamo/test_backends.py::TestOptimizations::test_aot_cudagraphs SKIPPED (requires cuda) [ 5%] 2023-03-31T05:58:12.8175839Z dynamo/test_backends.py::TestOptimizations::test_aot_eager PASSED [ 11%] 2023-03-31T05:58:12.8176409Z dynamo/test_backends.py::TestOptimizations::test_aot_eager_decomp_partition PASSED [ 16%] 2023-03-31T05:58:12.8177019Z dynamo/test_backends.py::TestOptimizations::test_aot_ts PASSED [ 22%] 2023-03-31T05:58:12.8177566Z dynamo/test_backends.py::TestOptimizations::test_aot_ts_nvfuser SKIPPED (requires cuda) [ 27%] 2023-03-31T05:58:12.8178159Z dynamo/test_backends.py::TestOptimizations::test_eager PASSED [ 33%] 2023-03-31T05:58:12.8178863Z dynamo/test_backends.py::TestOptimizations::test_example_inputs PASSED [ 38%] 2023-03-31T05:58:12.8179541Z dynamo/test_backends.py::TestOptimizations::test_example_inputs_runtime_use PASSED [ 44%] 2023-03-31T05:58:12.8180228Z dynamo/test_backends.py::TestOptimizations::test_ipex_bf16 SKIPPED (requires ipex) [ 50%] 2023-03-31T05:58:12.8180992Z dynamo/test_backends.py::TestOptimizations::test_ipex_fp32 SKIPPED (requires ipex) [ 55%] 2023-03-31T05:58:12.8181878Z dynamo/test_backends.py::TestOptimizations::test_list_backends PASSED [ 61%] 2023-03-31T05:58:12.8182642Z dynamo/test_backends.py::TestOptimizations::test_nvprims_aten SKIPPED (requires cuda) [ 66%] 2023-03-31T05:58:12.8183219Z dynamo/test_backends.py::TestOptimizations::test_nvprims_nvfuser SKIPPED (requires cuda) [ 72%] 2023-03-31T05:58:12.8183953Z dynamo/test_backends.py::TestOptimizations::test_onnxrt SKIPPED (requires onnxruntime) [ 77%] 2023-03-31T05:58:12.8184644Z dynamo/test_backends.py::TestOptimizations::test_torchscript PASSED [ 83%] 2023-03-31T05:58:12.8185272Z dynamo/test_backends.py::TestOptimizations::test_tvm SKIPPED (requires tvm) [ 88%] 2023-03-31T05:58:12.8185877Z dynamo/test_backends.py::NormalizeIRTests::test_inplace_normalize PASSED [ 94%] 2023-03-31T05:58:12.8186548Z dynamo/test_backends.py::MPSNotSupportedTest::test_default_mps_to_aot_eager SKIPPED (requires mps) [100%] 2023-03-31T05:58:12.8186954Z 2023-03-31T05:58:12.8187650Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-cd2960c4786ef00b.xml - 2023-03-31T05:58:12.8188478Z ========================= 9 passed, 9 skipped in 1.23s ========================= 2023-03-31T05:58:12.8189059Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:12.8189438Z 2023-03-31T05:58:12.8189873Z ##[endgroup] 2023-03-31T05:58:12.8190704Z FINISHED PRINTING LOG FILE of dynamo/test_backends (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_backends_d1qug9y2.log) 2023-03-31T05:58:12.8191191Z 2023-03-31T05:58:14.9247447Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:14.9274309Z Ignoring disabled issues: [] 2023-03-31T05:58:14.9447179Z Running test_bundled_inputs ... [2023-03-31 05:58:14.944341] 2023-03-31T05:58:14.9448835Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_bundled_inputs.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:14.944663] 2023-03-31T05:58:18.0629757Z 2023-03-31T05:58:18.0630315Z Expand the folded group to see the log file of test_bundled_inputs 2023-03-31T05:58:18.0631368Z ##[group]PRINTING LOG FILE of test_bundled_inputs (/var/lib/jenkins/workspace/test/test-reports/test_bundled_inputs_ff3t0_ma.log) 2023-03-31T05:58:18.0632413Z Test results will be stored in test-reports/python-pytest/test_bundled_inputs/test_bundled_inputs-fdaf441b142eb3fe.xml 2023-03-31T05:58:18.0632995Z ============================= test session starts ============================== 2023-03-31T05:58:18.0633651Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:18.0634089Z cachedir: .pytest_cache 2023-03-31T05:58:18.0634823Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:18.0635430Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:18.0636191Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:18.0636698Z collecting ... collected 12 items 2023-03-31T05:58:18.0639057Z Running 12 items in this shard: test/test_bundled_inputs.py::TestBundledInputs::test_bad_inputs, test/test_bundled_inputs.py::TestBundledInputs::test_dict_args, test/test_bundled_inputs.py::TestBundledInputs::test_double_augment_fail, test/test_bundled_inputs.py::TestBundledInputs::test_double_augment_non_mutator, test/test_bundled_inputs.py::TestBundledInputs::test_double_augment_success, test/test_bundled_inputs.py::TestBundledInputs::test_large_tensor_with_inflation, test/test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs, test/test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs_both_defined_failure, test/test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs_neither_defined_failure, test/test_bundled_inputs.py::TestBundledInputs::test_non_tensors, test/test_bundled_inputs.py::TestBundledInputs::test_rejected_tensors, test/test_bundled_inputs.py::TestBundledInputs::test_single_tensors 2023-03-31T05:58:18.0641124Z 2023-03-31T05:58:18.0641386Z test_bundled_inputs.py::TestBundledInputs::test_bad_inputs PASSED [ 8%] 2023-03-31T05:58:18.0641963Z test_bundled_inputs.py::TestBundledInputs::test_dict_args PASSED [ 16%] 2023-03-31T05:58:18.0642540Z test_bundled_inputs.py::TestBundledInputs::test_double_augment_fail PASSED [ 25%] 2023-03-31T05:58:18.0643396Z test_bundled_inputs.py::TestBundledInputs::test_double_augment_non_mutator PASSED [ 33%] 2023-03-31T05:58:18.0643983Z test_bundled_inputs.py::TestBundledInputs::test_double_augment_success PASSED [ 41%] 2023-03-31T05:58:18.0644599Z test_bundled_inputs.py::TestBundledInputs::test_large_tensor_with_inflation PASSED [ 50%] 2023-03-31T05:58:18.0645218Z test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs PASSED [ 58%] 2023-03-31T05:58:18.0646022Z test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs_both_defined_failure PASSED [ 66%] 2023-03-31T05:58:18.0646711Z test_bundled_inputs.py::TestBundledInputs::test_multiple_methods_with_inputs_neither_defined_failure PASSED [ 75%] 2023-03-31T05:58:18.0647340Z test_bundled_inputs.py::TestBundledInputs::test_non_tensors PASSED [ 83%] 2023-03-31T05:58:18.0647907Z test_bundled_inputs.py::TestBundledInputs::test_rejected_tensors PASSED [ 91%] 2023-03-31T05:58:18.0648481Z test_bundled_inputs.py::TestBundledInputs::test_single_tensors PASSED [100%] 2023-03-31T05:58:18.0648816Z 2023-03-31T05:58:18.0649461Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_bundled_inputs/test_bundled_inputs-fdaf441b142eb3fe.xml - 2023-03-31T05:58:18.0650110Z ============================== 12 passed in 1.17s ============================== 2023-03-31T05:58:18.0650803Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:18.0651158Z 2023-03-31T05:58:18.0651613Z ##[endgroup] 2023-03-31T05:58:18.0652377Z FINISHED PRINTING LOG FILE of test_bundled_inputs (/var/lib/jenkins/workspace/test/test-reports/test_bundled_inputs_ff3t0_ma.log) 2023-03-31T05:58:18.0652808Z 2023-03-31T05:58:20.1777110Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:20.1807172Z Ignoring disabled issues: [] 2023-03-31T05:58:20.1980377Z Running test_futures ... [2023-03-31 05:58:20.197721] 2023-03-31T05:58:20.1982408Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_futures.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:20.198012] 2023-03-31T05:58:23.1947797Z 2023-03-31T05:58:23.1948290Z Expand the folded group to see the log file of test_futures 2023-03-31T05:58:23.1949362Z ##[group]PRINTING LOG FILE of test_futures (/var/lib/jenkins/workspace/test/test-reports/test_futures_7hxzild8.log) 2023-03-31T05:58:23.1950175Z Test results will be stored in test-reports/python-pytest/test_futures/test_futures-3d74c9457cb7d0ac.xml 2023-03-31T05:58:23.1950517Z ============================= test session starts ============================== 2023-03-31T05:58:23.1950910Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:23.1951187Z cachedir: .pytest_cache 2023-03-31T05:58:23.1951817Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:23.1952238Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:23.1952751Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:23.1953060Z collecting ... collected 22 items 2023-03-31T05:58:23.1954964Z Running 22 items in this shard: test/test_futures.py::TestFuture::test_add_done_callback_error_is_ignored, test/test_futures.py::TestFuture::test_add_done_callback_maintains_callback_order, test/test_futures.py::TestFuture::test_add_done_callback_no_arg_error_is_ignored, test/test_futures.py::TestFuture::test_add_done_callback_simple, test/test_futures.py::TestFuture::test_chained_then, test/test_futures.py::TestFuture::test_collect_all, test/test_futures.py::TestFuture::test_done, test/test_futures.py::TestFuture::test_done_exception, test/test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_maintains_callback_order, test/test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_propagates_error, test/test_futures.py::TestFuture::test_mark_future_twice, test/test_futures.py::TestFuture::test_pickle_future, test/test_futures.py::TestFuture::test_set_exception, test/test_futures.py::TestFuture::test_set_exception_multithreading, test/test_futures.py::TestFuture::test_then, test/test_futures.py::TestFuture::test_then_no_arg, test/test_futures.py::TestFuture::test_then_raise, test/test_futures.py::TestFuture::test_then_wrong_arg, test/test_futures.py::TestFuture::test_wait, test/test_futures.py::TestFuture::test_wait_all, test/test_futures.py::TestFuture::test_wait_multi_thread, test/test_futures.py::TestFuture::test_wait_none 2023-03-31T05:58:23.1956703Z 2023-03-31T05:58:23.1956987Z test_futures.py::TestFuture::test_add_done_callback_error_is_ignored [E pybind_utils.h:214] Got the following error when running the callback: ValueError: Expected error 2023-03-31T05:58:23.1957283Z 2023-03-31T05:58:23.1957351Z At: 2023-03-31T05:58:23.1957534Z test_futures.py(236): raise_value_error 2023-03-31T05:58:23.1957994Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/futures/__init__.py(244): set_result 2023-03-31T05:58:23.1958312Z test_futures.py(229): _test_add_done_callback_error_ignored 2023-03-31T05:58:23.1958652Z test_futures.py(238): test_add_done_callback_error_is_ignored 2023-03-31T05:58:23.1959024Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(633): _callTestMethod 2023-03-31T05:58:23.1959396Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(676): run 2023-03-31T05:58:23.1959848Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(2201): _run_with_retry 2023-03-31T05:58:23.1960379Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(2272): run 2023-03-31T05:58:23.1960714Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(736): __call__ 2023-03-31T05:58:23.1961182Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/unittest.py(330): runtest 2023-03-31T05:58:23.1961649Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(167): pytest_runtest_call 2023-03-31T05:58:23.1962095Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1962563Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1962976Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1963618Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(260): 2023-03-31T05:58:23.1964094Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(339): from_call 2023-03-31T05:58:23.1964520Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(259): call_runtest_hook 2023-03-31T05:58:23.1965079Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(220): call_and_report 2023-03-31T05:58:23.1965621Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(131): runtestprotocol 2023-03-31T05:58:23.1966090Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pytest_rerunfailures.py(578): pytest_runtest_protocol 2023-03-31T05:58:23.1966595Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1967015Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1967464Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1967960Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(349): pytest_runtestloop 2023-03-31T05:58:23.1968386Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1968851Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1969261Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1969734Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(324): _main 2023-03-31T05:58:23.1970185Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(270): wrap_session 2023-03-31T05:58:23.1970617Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(317): pytest_cmdline_main 2023-03-31T05:58:23.1971106Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1971635Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1972074Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1972698Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/config/__init__.py(167): main 2023-03-31T05:58:23.1973450Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(794): run_tests 2023-03-31T05:58:23.1973736Z test_futures.py(340): 2023-03-31T05:58:23.1973868Z 2023-03-31T05:58:23.1973942Z PASSED [ 4%] 2023-03-31T05:58:23.1974297Z test_futures.py::TestFuture::test_add_done_callback_maintains_callback_order PASSED [ 9%] 2023-03-31T05:58:23.1974813Z test_futures.py::TestFuture::test_add_done_callback_no_arg_error_is_ignored [E pybind_utils.h:214] Got the following error when running the callback: TypeError: no_arg() takes 0 positional arguments but 1 was given 2023-03-31T05:58:23.1975265Z PASSED [ 13%] 2023-03-31T05:58:23.1975540Z test_futures.py::TestFuture::test_add_done_callback_simple PASSED [ 18%] 2023-03-31T05:58:23.1975914Z test_futures.py::TestFuture::test_chained_then PASSED [ 22%] 2023-03-31T05:58:23.1976211Z test_futures.py::TestFuture::test_collect_all PASSED [ 27%] 2023-03-31T05:58:23.1976581Z test_futures.py::TestFuture::test_done PASSED [ 31%] 2023-03-31T05:58:23.1976884Z test_futures.py::TestFuture::test_done_exception PASSED [ 36%] 2023-03-31T05:58:23.1977314Z test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_maintains_callback_order PASSED [ 40%] 2023-03-31T05:58:23.1977800Z test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_propagates_error [E pybind_utils.h:214] Got the following error when running the callback: ValueError: Expected error 2023-03-31T05:58:23.1978153Z 2023-03-31T05:58:23.1978219Z At: 2023-03-31T05:58:23.1978412Z test_futures.py(280): raise_value_error 2023-03-31T05:58:23.1978862Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/futures/__init__.py(244): set_result 2023-03-31T05:58:23.1979202Z test_futures.py(285): test_interleaving_then_and_add_done_callback_propagates_error 2023-03-31T05:58:23.1979595Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(633): _callTestMethod 2023-03-31T05:58:23.1979903Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(676): run 2023-03-31T05:58:23.1980437Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(2201): _run_with_retry 2023-03-31T05:58:23.1980990Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(2272): run 2023-03-31T05:58:23.1981335Z /opt/conda/envs/py_3.8/lib/python3.8/unittest/case.py(736): __call__ 2023-03-31T05:58:23.1981776Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/unittest.py(330): runtest 2023-03-31T05:58:23.1982214Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(167): pytest_runtest_call 2023-03-31T05:58:23.1982697Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1983220Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1983675Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1984148Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(260): 2023-03-31T05:58:23.1984559Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(339): from_call 2023-03-31T05:58:23.1985003Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(259): call_runtest_hook 2023-03-31T05:58:23.1985452Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(220): call_and_report 2023-03-31T05:58:23.1985878Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/runner.py(131): runtestprotocol 2023-03-31T05:58:23.1986425Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pytest_rerunfailures.py(578): pytest_runtest_protocol 2023-03-31T05:58:23.1986867Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1987280Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1987687Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1988110Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(349): pytest_runtestloop 2023-03-31T05:58:23.1988548Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1988987Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1989389Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1989825Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(324): _main 2023-03-31T05:58:23.1990218Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(270): wrap_session 2023-03-31T05:58:23.1990640Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/main.py(317): pytest_cmdline_main 2023-03-31T05:58:23.1991057Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_callers.py(39): _multicall 2023-03-31T05:58:23.1991453Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_manager.py(80): _hookexec 2023-03-31T05:58:23.1991859Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/pluggy/_hooks.py(265): __call__ 2023-03-31T05:58:23.1992293Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/_pytest/config/__init__.py(167): main 2023-03-31T05:58:23.1992796Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/testing/_internal/common_utils.py(794): run_tests 2023-03-31T05:58:23.1993076Z test_futures.py(340): 2023-03-31T05:58:23.1993209Z 2023-03-31T05:58:23.1993285Z PASSED [ 45%] 2023-03-31T05:58:23.1993542Z test_futures.py::TestFuture::test_mark_future_twice PASSED [ 50%] 2023-03-31T05:58:23.1993845Z test_futures.py::TestFuture::test_pickle_future PASSED [ 54%] 2023-03-31T05:58:23.1994148Z test_futures.py::TestFuture::test_set_exception PASSED [ 59%] 2023-03-31T05:58:23.1994468Z test_futures.py::TestFuture::test_set_exception_multithreading PASSED [ 63%] 2023-03-31T05:58:23.1994819Z test_futures.py::TestFuture::test_then PASSED [ 68%] 2023-03-31T05:58:23.1995151Z test_futures.py::TestFuture::test_then_no_arg PASSED [ 72%] 2023-03-31T05:58:23.1995446Z test_futures.py::TestFuture::test_then_raise PASSED [ 77%] 2023-03-31T05:58:23.1995740Z test_futures.py::TestFuture::test_then_wrong_arg PASSED [ 81%] 2023-03-31T05:58:23.1996021Z test_futures.py::TestFuture::test_wait PASSED [ 86%] 2023-03-31T05:58:23.1996310Z test_futures.py::TestFuture::test_wait_all PASSED [ 90%] 2023-03-31T05:58:23.1996610Z test_futures.py::TestFuture::test_wait_multi_thread PASSED [ 95%] 2023-03-31T05:58:23.1996906Z test_futures.py::TestFuture::test_wait_none PASSED [100%] 2023-03-31T05:58:23.1997059Z 2023-03-31T05:58:23.1997397Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_futures/test_futures-3d74c9457cb7d0ac.xml - 2023-03-31T05:58:23.1997761Z ============================== 22 passed in 0.99s ============================== 2023-03-31T05:58:23.1998084Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:23.1998286Z 2023-03-31T05:58:23.1998574Z ##[endgroup] 2023-03-31T05:58:23.1998973Z FINISHED PRINTING LOG FILE of test_futures (/var/lib/jenkins/workspace/test/test-reports/test_futures_7hxzild8.log) 2023-03-31T05:58:23.1999200Z 2023-03-31T05:58:25.3050376Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:25.3078699Z Ignoring disabled issues: [] 2023-03-31T05:58:25.3256729Z Running test_namedtensor ... [2023-03-31 05:58:25.325333] 2023-03-31T05:58:25.3258720Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_namedtensor.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:25.325642] 2023-03-31T05:58:28.5281390Z 2023-03-31T05:58:28.5281942Z Expand the folded group to see the log file of test_namedtensor 2023-03-31T05:58:28.5282936Z ##[group]PRINTING LOG FILE of test_namedtensor (/var/lib/jenkins/workspace/test/test-reports/test_namedtensor_w9kosgfp.log) 2023-03-31T05:58:28.5284089Z Test results will be stored in test-reports/python-pytest/test_namedtensor/test_namedtensor-67adda638f28d19a.xml 2023-03-31T05:58:28.5284585Z ============================= test session starts ============================== 2023-03-31T05:58:28.5285411Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:28.5285822Z cachedir: .pytest_cache 2023-03-31T05:58:28.5286455Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:28.5286983Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:28.5287628Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:28.5288076Z collecting ... collected 86 items 2023-03-31T05:58:28.5298622Z Running 86 items in this shard: test/test_namedtensor.py::TestNamedTensor::test_aaa_must_run_first_check_experimental_warning, test/test_namedtensor.py::TestNamedTensor::test_addcmul_addcdiv, test/test_namedtensor.py::TestNamedTensor::test_addmm, test/test_namedtensor.py::TestNamedTensor::test_addmv, test/test_namedtensor.py::TestNamedTensor::test_align_as, test/test_namedtensor.py::TestNamedTensor::test_align_tensors, test/test_namedtensor.py::TestNamedTensor::test_align_tensors_two_inputs, test/test_namedtensor.py::TestNamedTensor::test_align_to, test/test_namedtensor.py::TestNamedTensor::test_align_to_ellipsis, test/test_namedtensor.py::TestNamedTensor::test_any_all, test/test_namedtensor.py::TestNamedTensor::test_as_strided, test/test_namedtensor.py::TestNamedTensor::test_as_strided_cuda, test/test_namedtensor.py::TestNamedTensor::test_autograd_ignores_names, test/test_namedtensor.py::TestNamedTensor::test_autograd_smoke, test/test_namedtensor.py::TestNamedTensor::test_autograd_warns_named_grad, test/test_namedtensor.py::TestNamedTensor::test_bernoulli, test/test_namedtensor.py::TestNamedTensor::test_big_tensor_repr_has_names, test/test_namedtensor.py::TestNamedTensor::test_binary_ops, test/test_namedtensor.py::TestNamedTensor::test_bitwise_not, test/test_namedtensor.py::TestNamedTensor::test_bmm, test/test_namedtensor.py::TestNamedTensor::test_cat, test/test_namedtensor.py::TestNamedTensor::test_cdist, test/test_namedtensor.py::TestNamedTensor::test_comparison_ops, test/test_namedtensor.py::TestNamedTensor::test_copy_transpose, test/test_namedtensor.py::TestNamedTensor::test_cummax_cummin, test/test_namedtensor.py::TestNamedTensor::test_detach, test/test_namedtensor.py::TestNamedTensor::test_diagonal, test/test_namedtensor.py::TestNamedTensor::test_dot, test/test_namedtensor.py::TestNamedTensor::test_equal, test/test_namedtensor.py::TestNamedTensor::test_expand, test/test_namedtensor.py::TestNamedTensor::test_factory_coverage, test/test_namedtensor.py::TestNamedTensor::test_factory_edge_cases, test/test_namedtensor.py::TestNamedTensor::test_flatten, test/test_namedtensor.py::TestNamedTensor::test_flatten_nodims, test/test_namedtensor.py::TestNamedTensor::test_has_names, test/test_namedtensor.py::TestNamedTensor::test_index_fill, test/test_namedtensor.py::TestNamedTensor::test_info_smoke, test/test_namedtensor.py::TestNamedTensor::test_logcumsumexp, test/test_namedtensor.py::TestNamedTensor::test_logical_not, test/test_namedtensor.py::TestNamedTensor::test_logical_ops, test/test_namedtensor.py::TestNamedTensor::test_masked_fill, test/test_namedtensor.py::TestNamedTensor::test_masked_select, test/test_namedtensor.py::TestNamedTensor::test_matmul, test/test_namedtensor.py::TestNamedTensor::test_max_pooling, test/test_namedtensor.py::TestNamedTensor::test_max_pooling_without_names_does_not_warn, test/test_namedtensor.py::TestNamedTensor::test_mm, test/test_namedtensor.py::TestNamedTensor::test_mv, test/test_namedtensor.py::TestNamedTensor::test_no_jit_script_support, test/test_namedtensor.py::TestNamedTensor::test_no_jit_tracer_support, test/test_namedtensor.py::TestNamedTensor::test_no_multiprocessing_support, test/test_namedtensor.py::TestNamedTensor::test_no_pickle_support, test/test_namedtensor.py::TestNamedTensor::test_no_save_support, test/test_namedtensor.py::TestNamedTensor::test_noncontig_contiguous, test/test_namedtensor.py::TestNamedTensor::test_none_names_refcount, test/test_namedtensor.py::TestNamedTensor::test_nyi_dimname_overload_msg, test/test_namedtensor.py::TestNamedTensor::test_out_fn_semantics, test/test_namedtensor.py::TestNamedTensor::test_pow_special, test/test_namedtensor.py::TestNamedTensor::test_py3_ellipsis, test/test_namedtensor.py::TestNamedTensor::test_reduction_fns, test/test_namedtensor.py::TestNamedTensor::test_refine_names, test/test_namedtensor.py::TestNamedTensor::test_rename, test/test_namedtensor.py::TestNamedTensor::test_rename_, test/test_namedtensor.py::TestNamedTensor::test_rename_globber, test/test_namedtensor.py::TestNamedTensor::test_rename_rename_map, test/test_namedtensor.py::TestNamedTensor::test_repr, test/test_namedtensor.py::TestNamedTensor::test_resize, test/test_namedtensor.py::TestNamedTensor::test_select, test/test_namedtensor.py::TestNamedTensor::test_select_cuda, test/test_namedtensor.py::TestNamedTensor::test_set_names_property, test/test_namedtensor.py::TestNamedTensor::test_size, test/test_namedtensor.py::TestNamedTensor::test_split_fns_propagates_names, test/test_namedtensor.py::TestNamedTensor::test_squeeze, test/test_namedtensor.py::TestNamedTensor::test_stride, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_lists, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_named_tensor, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_numpy, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_tensor, test/test_namedtensor.py::TestNamedTensor::test_tensor_grad_is_unnamed, test/test_namedtensor.py::TestNamedTensor::test_transpose_variants, test/test_namedtensor.py::TestNamedTensor::test_trivial, test/test_namedtensor.py::TestNamedTensor::test_unary_propagate_names_fns, test/test_namedtensor.py::TestNamedTensor::test_unflatten, test/test_namedtensor.py::TestNamedTensor::test_unsupported_op_error_msg, test/test_namedtensor.py::TestNamedTensor::test_using_seen_interned_string_doesnt_bump_refcount, test/test_namedtensor.py::TestNamedTensor::test_using_unseen_interned_string_bumps_refcount_permanently, test/test_namedtensor.py::TestNamedTensor::test_using_unseen_uninterned_string_refcounts 2023-03-31T05:58:28.5308471Z 2023-03-31T05:58:28.5308770Z test_namedtensor.py::TestNamedTensor::test_aaa_must_run_first_check_experimental_warning PASSED [ 1%] 2023-03-31T05:58:28.5309338Z test_namedtensor.py::TestNamedTensor::test_addcmul_addcdiv PASSED [ 2%] 2023-03-31T05:58:28.5309824Z test_namedtensor.py::TestNamedTensor::test_addmm PASSED [ 3%] 2023-03-31T05:58:28.5310295Z test_namedtensor.py::TestNamedTensor::test_addmv PASSED [ 4%] 2023-03-31T05:58:28.5310773Z test_namedtensor.py::TestNamedTensor::test_align_as PASSED [ 5%] 2023-03-31T05:58:28.5311307Z test_namedtensor.py::TestNamedTensor::test_align_tensors SKIPPED (Not implemented yet) [ 6%] 2023-03-31T05:58:28.5311877Z test_namedtensor.py::TestNamedTensor::test_align_tensors_two_inputs SKIPPED (Not implemented yet) [ 8%] 2023-03-31T05:58:28.5312411Z test_namedtensor.py::TestNamedTensor::test_align_to PASSED [ 9%] 2023-03-31T05:58:28.5312887Z test_namedtensor.py::TestNamedTensor::test_align_to_ellipsis PASSED [ 10%] 2023-03-31T05:58:28.5313451Z test_namedtensor.py::TestNamedTensor::test_any_all PASSED [ 11%] 2023-03-31T05:58:28.5313923Z test_namedtensor.py::TestNamedTensor::test_as_strided PASSED [ 12%] 2023-03-31T05:58:28.5314403Z test_namedtensor.py::TestNamedTensor::test_as_strided_cuda SKIPPED (no CUDA) [ 13%] 2023-03-31T05:58:28.5314916Z test_namedtensor.py::TestNamedTensor::test_autograd_ignores_names PASSED [ 15%] 2023-03-31T05:58:28.5315434Z test_namedtensor.py::TestNamedTensor::test_autograd_smoke PASSED [ 16%] 2023-03-31T05:58:28.5315948Z test_namedtensor.py::TestNamedTensor::test_autograd_warns_named_grad PASSED [ 17%] 2023-03-31T05:58:28.5316424Z test_namedtensor.py::TestNamedTensor::test_bernoulli PASSED [ 18%] 2023-03-31T05:58:28.5316950Z test_namedtensor.py::TestNamedTensor::test_big_tensor_repr_has_names PASSED [ 19%] 2023-03-31T05:58:28.5317516Z test_namedtensor.py::TestNamedTensor::test_binary_ops PASSED [ 20%] 2023-03-31T05:58:28.5317983Z test_namedtensor.py::TestNamedTensor::test_bitwise_not PASSED [ 22%] 2023-03-31T05:58:28.5318444Z test_namedtensor.py::TestNamedTensor::test_bmm PASSED [ 23%] 2023-03-31T05:58:28.5318895Z test_namedtensor.py::TestNamedTensor::test_cat PASSED [ 24%] 2023-03-31T05:58:28.5319372Z test_namedtensor.py::TestNamedTensor::test_cdist PASSED [ 25%] 2023-03-31T05:58:28.5319839Z test_namedtensor.py::TestNamedTensor::test_comparison_ops PASSED [ 26%] 2023-03-31T05:58:28.5320339Z test_namedtensor.py::TestNamedTensor::test_copy_transpose PASSED [ 27%] 2023-03-31T05:58:28.5320824Z test_namedtensor.py::TestNamedTensor::test_cummax_cummin PASSED [ 29%] 2023-03-31T05:58:28.5323765Z test_namedtensor.py::TestNamedTensor::test_detach PASSED [ 30%] 2023-03-31T05:58:28.5324236Z test_namedtensor.py::TestNamedTensor::test_diagonal PASSED [ 31%] 2023-03-31T05:58:28.5324713Z test_namedtensor.py::TestNamedTensor::test_dot PASSED [ 32%] 2023-03-31T05:58:28.5325210Z test_namedtensor.py::TestNamedTensor::test_equal PASSED [ 33%] 2023-03-31T05:58:28.5325683Z test_namedtensor.py::TestNamedTensor::test_expand PASSED [ 34%] 2023-03-31T05:58:28.5326154Z test_namedtensor.py::TestNamedTensor::test_factory_coverage PASSED [ 36%] 2023-03-31T05:58:28.5326767Z test_namedtensor.py::TestNamedTensor::test_factory_edge_cases PASSED [ 37%] 2023-03-31T05:58:28.5327305Z test_namedtensor.py::TestNamedTensor::test_flatten PASSED [ 38%] 2023-03-31T05:58:28.5327780Z test_namedtensor.py::TestNamedTensor::test_flatten_nodims PASSED [ 39%] 2023-03-31T05:58:28.5328267Z test_namedtensor.py::TestNamedTensor::test_has_names PASSED [ 40%] 2023-03-31T05:58:28.5328736Z test_namedtensor.py::TestNamedTensor::test_index_fill PASSED [ 41%] 2023-03-31T05:58:28.5329199Z test_namedtensor.py::TestNamedTensor::test_info_smoke PASSED [ 43%] 2023-03-31T05:58:28.5329672Z test_namedtensor.py::TestNamedTensor::test_logcumsumexp PASSED [ 44%] 2023-03-31T05:58:28.5330162Z test_namedtensor.py::TestNamedTensor::test_logical_not PASSED [ 45%] 2023-03-31T05:58:28.5330633Z test_namedtensor.py::TestNamedTensor::test_logical_ops PASSED [ 46%] 2023-03-31T05:58:28.5331083Z test_namedtensor.py::TestNamedTensor::test_masked_fill PASSED [ 47%] 2023-03-31T05:58:28.5331567Z test_namedtensor.py::TestNamedTensor::test_masked_select PASSED [ 48%] 2023-03-31T05:58:28.5332052Z test_namedtensor.py::TestNamedTensor::test_matmul PASSED [ 50%] 2023-03-31T05:58:28.5332517Z test_namedtensor.py::TestNamedTensor::test_max_pooling PASSED [ 51%] 2023-03-31T05:58:28.5333025Z test_namedtensor.py::TestNamedTensor::test_max_pooling_without_names_does_not_warn PASSED [ 52%] 2023-03-31T05:58:28.5333536Z test_namedtensor.py::TestNamedTensor::test_mm PASSED [ 53%] 2023-03-31T05:58:28.5334073Z test_namedtensor.py::TestNamedTensor::test_mv PASSED [ 54%] 2023-03-31T05:58:28.5334540Z test_namedtensor.py::TestNamedTensor::test_no_jit_script_support PASSED [ 55%] 2023-03-31T05:58:28.5335056Z test_namedtensor.py::TestNamedTensor::test_no_jit_tracer_support PASSED [ 56%] 2023-03-31T05:58:28.5335578Z test_namedtensor.py::TestNamedTensor::test_no_multiprocessing_support PASSED [ 58%] 2023-03-31T05:58:28.5336104Z test_namedtensor.py::TestNamedTensor::test_no_pickle_support PASSED [ 59%] 2023-03-31T05:58:28.5336588Z test_namedtensor.py::TestNamedTensor::test_no_save_support PASSED [ 60%] 2023-03-31T05:58:28.5337081Z test_namedtensor.py::TestNamedTensor::test_noncontig_contiguous PASSED [ 61%] 2023-03-31T05:58:28.5337565Z test_namedtensor.py::TestNamedTensor::test_none_names_refcount PASSED [ 62%] 2023-03-31T05:58:28.5338119Z test_namedtensor.py::TestNamedTensor::test_nyi_dimname_overload_msg PASSED [ 63%] 2023-03-31T05:58:28.5338621Z test_namedtensor.py::TestNamedTensor::test_out_fn_semantics PASSED [ 65%] 2023-03-31T05:58:28.5339098Z test_namedtensor.py::TestNamedTensor::test_pow_special PASSED [ 66%] 2023-03-31T05:58:28.5339581Z test_namedtensor.py::TestNamedTensor::test_py3_ellipsis PASSED [ 67%] 2023-03-31T05:58:28.5340054Z test_namedtensor.py::TestNamedTensor::test_reduction_fns PASSED [ 68%] 2023-03-31T05:58:28.5340532Z test_namedtensor.py::TestNamedTensor::test_refine_names PASSED [ 69%] 2023-03-31T05:58:28.5341013Z test_namedtensor.py::TestNamedTensor::test_rename PASSED [ 70%] 2023-03-31T05:58:28.5341455Z test_namedtensor.py::TestNamedTensor::test_rename_ PASSED [ 72%] 2023-03-31T05:58:28.5341944Z test_namedtensor.py::TestNamedTensor::test_rename_globber PASSED [ 73%] 2023-03-31T05:58:28.5342435Z test_namedtensor.py::TestNamedTensor::test_rename_rename_map PASSED [ 74%] 2023-03-31T05:58:28.5343012Z test_namedtensor.py::TestNamedTensor::test_repr PASSED [ 75%] 2023-03-31T05:58:28.5343461Z test_namedtensor.py::TestNamedTensor::test_resize PASSED [ 76%] 2023-03-31T05:58:28.5343930Z test_namedtensor.py::TestNamedTensor::test_select PASSED [ 77%] 2023-03-31T05:58:28.5344399Z test_namedtensor.py::TestNamedTensor::test_select_cuda SKIPPED (no CUDA) [ 79%] 2023-03-31T05:58:28.5344872Z test_namedtensor.py::TestNamedTensor::test_set_names_property PASSED [ 80%] 2023-03-31T05:58:28.5345446Z test_namedtensor.py::TestNamedTensor::test_size PASSED [ 81%] 2023-03-31T05:58:28.5345991Z test_namedtensor.py::TestNamedTensor::test_split_fns_propagates_names PASSED [ 82%] 2023-03-31T05:58:28.5346493Z test_namedtensor.py::TestNamedTensor::test_squeeze PASSED [ 83%] 2023-03-31T05:58:28.5347033Z test_namedtensor.py::TestNamedTensor::test_stride PASSED [ 84%] 2023-03-31T05:58:28.5347608Z test_namedtensor.py::TestNamedTensor::test_tensor_from_lists PASSED [ 86%] 2023-03-31T05:58:28.5348206Z test_namedtensor.py::TestNamedTensor::test_tensor_from_named_tensor PASSED [ 87%] 2023-03-31T05:58:28.5348786Z test_namedtensor.py::TestNamedTensor::test_tensor_from_numpy PASSED [ 88%] 2023-03-31T05:58:28.5349361Z test_namedtensor.py::TestNamedTensor::test_tensor_from_tensor PASSED [ 89%] 2023-03-31T05:58:28.5349941Z test_namedtensor.py::TestNamedTensor::test_tensor_grad_is_unnamed PASSED [ 90%] 2023-03-31T05:58:28.5350536Z test_namedtensor.py::TestNamedTensor::test_transpose_variants PASSED [ 91%] 2023-03-31T05:58:28.5351103Z test_namedtensor.py::TestNamedTensor::test_trivial PASSED [ 93%] 2023-03-31T05:58:28.5351685Z test_namedtensor.py::TestNamedTensor::test_unary_propagate_names_fns PASSED [ 94%] 2023-03-31T05:58:28.5352282Z test_namedtensor.py::TestNamedTensor::test_unflatten PASSED [ 95%] 2023-03-31T05:58:28.5352867Z test_namedtensor.py::TestNamedTensor::test_unsupported_op_error_msg PASSED [ 96%] 2023-03-31T05:58:28.5353618Z test_namedtensor.py::TestNamedTensor::test_using_seen_interned_string_doesnt_bump_refcount PASSED [ 97%] 2023-03-31T05:58:28.5354371Z test_namedtensor.py::TestNamedTensor::test_using_unseen_interned_string_bumps_refcount_permanently PASSED [ 98%] 2023-03-31T05:58:28.5355113Z test_namedtensor.py::TestNamedTensor::test_using_unseen_uninterned_string_refcounts PASSED [100%] 2023-03-31T05:58:28.5355483Z 2023-03-31T05:58:28.5356190Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_namedtensor/test_namedtensor-67adda638f28d19a.xml - 2023-03-31T05:58:28.5356859Z ======================== 82 passed, 4 skipped in 1.00s ========================= 2023-03-31T05:58:28.5357408Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:28.5357763Z 2023-03-31T05:58:28.5358388Z ##[endgroup] 2023-03-31T05:58:28.5359155Z FINISHED PRINTING LOG FILE of test_namedtensor (/var/lib/jenkins/workspace/test/test-reports/test_namedtensor_w9kosgfp.log) 2023-03-31T05:58:28.5359565Z 2023-03-31T05:58:30.6250958Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:30.6287309Z Ignoring disabled issues: [] 2023-03-31T05:58:30.6465815Z Running test_tensorexpr_pybind ... [2023-03-31 05:58:30.646248] 2023-03-31T05:58:30.6468483Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_tensorexpr_pybind.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:30.646590] 2023-03-31T05:58:33.5308534Z 2023-03-31T05:58:33.5309135Z Expand the folded group to see the log file of test_tensorexpr_pybind 2023-03-31T05:58:33.5310359Z ##[group]PRINTING LOG FILE of test_tensorexpr_pybind (/var/lib/jenkins/workspace/test/test-reports/test_tensorexpr_pybind_hjf77cap.log) 2023-03-31T05:58:33.5311363Z Test results will be stored in test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-c570424306e69389.xml 2023-03-31T05:58:33.5312025Z ============================= test session starts ============================== 2023-03-31T05:58:33.5312623Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:33.5312906Z cachedir: .pytest_cache 2023-03-31T05:58:33.5313332Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:33.5313919Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:33.5314432Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:33.5314748Z collecting ... collected 17 items 2023-03-31T05:58:33.5316485Z Running 17 items in this shard: test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_alloc_in_loop, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_call_raw, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dtype_error, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape_2d, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_external_calls, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop_module, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_custom_lowering, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_expand, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_permute, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_scalar_inputs, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_t, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_tensor_inputs, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_transpose, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_simple_sum, test/test_tensorexpr_pybind.py::TestExprHandlePyBind::test_unary_ops 2023-03-31T05:58:33.5318228Z 2023-03-31T05:58:33.5318405Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_alloc_in_loop PASSED [ 5%] 2023-03-31T05:58:33.5318750Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_call_raw PASSED [ 11%] 2023-03-31T05:58:33.5319100Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dtype_error PASSED [ 17%] 2023-03-31T05:58:33.5319460Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape PASSED [ 23%] 2023-03-31T05:58:33.5319829Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape_2d PASSED [ 29%] 2023-03-31T05:58:33.5320186Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_external_calls PASSED [ 35%] 2023-03-31T05:58:33.5320553Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop PASSED [ 41%] 2023-03-31T05:58:33.5320929Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop_module PASSED [ 47%] 2023-03-31T05:58:33.5321378Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_custom_lowering PASSED [ 52%] 2023-03-31T05:58:33.5321738Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_expand PASSED [ 58%] 2023-03-31T05:58:33.5322108Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_permute PASSED [ 64%] 2023-03-31T05:58:33.5322485Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_scalar_inputs PASSED [ 70%] 2023-03-31T05:58:33.5322841Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_t PASSED [ 76%] 2023-03-31T05:58:33.5323477Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_tensor_inputs PASSED [ 82%] 2023-03-31T05:58:33.5323957Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_transpose PASSED [ 88%] 2023-03-31T05:58:33.5324326Z test_tensorexpr_pybind.py::TestTensorExprPyBind::test_simple_sum PASSED [ 94%] 2023-03-31T05:58:33.5324669Z test_tensorexpr_pybind.py::TestExprHandlePyBind::test_unary_ops PASSED [100%] 2023-03-31T05:58:33.5324860Z 2023-03-31T05:58:33.5325245Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-c570424306e69389.xml - 2023-03-31T05:58:33.5325622Z ============================== 17 passed in 0.91s ============================== 2023-03-31T05:58:33.5325943Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:33.5326134Z 2023-03-31T05:58:33.5326484Z ##[endgroup] 2023-03-31T05:58:33.5326974Z FINISHED PRINTING LOG FILE of test_tensorexpr_pybind (/var/lib/jenkins/workspace/test/test-reports/test_tensorexpr_pybind_hjf77cap.log) 2023-03-31T05:58:33.5327228Z 2023-03-31T05:58:35.6086609Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:35.6114030Z Ignoring disabled issues: [] 2023-03-31T05:58:35.6289488Z Running test_model_dump ... [2023-03-31 05:58:35.628587] 2023-03-31T05:58:35.6291930Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_model_dump.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:35.628919] 2023-03-31T05:58:38.4386558Z 2023-03-31T05:58:38.4387062Z Expand the folded group to see the log file of test_model_dump 2023-03-31T05:58:38.4388687Z ##[group]PRINTING LOG FILE of test_model_dump (/var/lib/jenkins/workspace/test/test-reports/test_model_dump_qy_8ko3t.log) 2023-03-31T05:58:38.4389855Z Test results will be stored in test-reports/python-pytest/test_model_dump/test_model_dump-de886d8ecb5e46a4.xml 2023-03-31T05:58:38.4390623Z ============================= test session starts ============================== 2023-03-31T05:58:38.4391375Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:38.4391867Z cachedir: .pytest_cache 2023-03-31T05:58:38.4392690Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:38.4393648Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:38.4394502Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:38.4395061Z collecting ... collected 9 items 2023-03-31T05:58:38.4396639Z Running 9 items in this shard: test/test_model_dump.py::TestModelDump::test_inline_skeleton, test/test_model_dump.py::TestModelDump::test_invalid_json, test/test_model_dump.py::TestModelDump::test_main, test/test_model_dump.py::TestModelDump::test_memory_computation, test/test_model_dump.py::TestModelDump::test_model_with_lists, test/test_model_dump.py::TestModelDump::test_optimized_quantized_model, test/test_model_dump.py::TestModelDump::test_quantized_model, test/test_model_dump.py::TestModelDump::test_scripted_model, test/test_model_dump.py::TestModelDump::test_traced_model 2023-03-31T05:58:38.4398048Z 2023-03-31T05:58:38.4398338Z test_model_dump.py::TestModelDump::test_inline_skeleton PASSED [ 11%] 2023-03-31T05:58:38.4398943Z test_model_dump.py::TestModelDump::test_invalid_json PASSED [ 22%] 2023-03-31T05:58:38.4399529Z test_model_dump.py::TestModelDump::test_main PASSED [ 33%] 2023-03-31T05:58:38.4400171Z test_model_dump.py::TestModelDump::test_memory_computation SKIPPED (Webdriver not requested) [ 44%] 2023-03-31T05:58:38.4400846Z test_model_dump.py::TestModelDump::test_model_with_lists PASSED [ 55%] 2023-03-31T05:58:38.4401490Z test_model_dump.py::TestModelDump::test_optimized_quantized_model PASSED [ 66%] 2023-03-31T05:58:38.4402111Z test_model_dump.py::TestModelDump::test_quantized_model PASSED [ 77%] 2023-03-31T05:58:38.4402700Z test_model_dump.py::TestModelDump::test_scripted_model PASSED [ 88%] 2023-03-31T05:58:38.4403505Z test_model_dump.py::TestModelDump::test_traced_model PASSED [100%] 2023-03-31T05:58:38.4403847Z 2023-03-31T05:58:38.4404318Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_model_dump/test_model_dump-de886d8ecb5e46a4.xml - 2023-03-31T05:58:38.4404782Z ========================= 8 passed, 1 skipped in 0.84s ========================= 2023-03-31T05:58:38.4405104Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:38.4405307Z 2023-03-31T05:58:38.4405579Z ##[endgroup] 2023-03-31T05:58:38.4406101Z FINISHED PRINTING LOG FILE of test_model_dump (/var/lib/jenkins/workspace/test/test-reports/test_model_dump_qy_8ko3t.log) 2023-03-31T05:58:38.4406399Z 2023-03-31T05:58:40.5266009Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:40.5294540Z Ignoring disabled issues: [] 2023-03-31T05:58:40.5469573Z Running dynamo/test_python_autograd ... [2023-03-31 05:58:40.546625] 2023-03-31T05:58:40.5471763Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_python_autograd.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:40.546928] 2023-03-31T05:58:43.5694008Z 2023-03-31T05:58:43.5694591Z Expand the folded group to see the log file of dynamo/test_python_autograd 2023-03-31T05:58:43.5695662Z ##[group]PRINTING LOG FILE of dynamo/test_python_autograd (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_python_autograd_c9oyk_ik.log) 2023-03-31T05:58:43.5696744Z Test results will be stored in test-reports/python-pytest/dynamo.test_python_autograd/dynamo.test_python_autograd-0520caf607d27108.xml 2023-03-31T05:58:43.5697307Z ============================= test session starts ============================== 2023-03-31T05:58:43.5697901Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:43.5698344Z cachedir: .pytest_cache 2023-03-31T05:58:43.5699167Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:43.5700111Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:43.5700933Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:43.5701509Z collecting ... collected 5 items 2023-03-31T05:58:43.5702829Z Running 5 items in this shard: test/dynamo/test_python_autograd.py::TestPythonAutograd::test_backwards1, test/dynamo/test_python_autograd.py::TestPythonAutograd::test_backwards2, test/dynamo/test_python_autograd.py::TestPythonAutograd::test_forwards1, test/dynamo/test_python_autograd.py::TestPythonAutograd::test_forwards2, test/dynamo/test_python_autograd.py::TestPythonAutograd::test_split 2023-03-31T05:58:43.5703819Z 2023-03-31T05:58:43.5704117Z dynamo/test_python_autograd.py::TestPythonAutograd::test_backwards1 PASSED [ 20%] 2023-03-31T05:58:43.5704775Z dynamo/test_python_autograd.py::TestPythonAutograd::test_backwards2 PASSED [ 40%] 2023-03-31T05:58:43.5705554Z dynamo/test_python_autograd.py::TestPythonAutograd::test_forwards1 PASSED [ 60%] 2023-03-31T05:58:43.5706201Z dynamo/test_python_autograd.py::TestPythonAutograd::test_forwards2 PASSED [ 80%] 2023-03-31T05:58:43.5706823Z dynamo/test_python_autograd.py::TestPythonAutograd::test_split PASSED [100%] 2023-03-31T05:58:43.5707173Z 2023-03-31T05:58:43.5707906Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_python_autograd/dynamo.test_python_autograd-0520caf607d27108.xml - 2023-03-31T05:58:43.5708528Z ============================== 5 passed in 0.92s =============================== 2023-03-31T05:58:43.5709064Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:43.5709408Z 2023-03-31T05:58:43.5709853Z ##[endgroup] 2023-03-31T05:58:43.5710602Z FINISHED PRINTING LOG FILE of dynamo/test_python_autograd (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_python_autograd_c9oyk_ik.log) 2023-03-31T05:58:43.5710987Z 2023-03-31T05:58:45.6884838Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:45.6912731Z Ignoring disabled issues: [] 2023-03-31T05:58:45.7087513Z Running test_fx_reinplace_pass ... [2023-03-31 05:58:45.708396] 2023-03-31T05:58:45.7090469Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_fx_reinplace_pass.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:45.708737] 2023-03-31T05:58:48.3957995Z 2023-03-31T05:58:48.3958593Z Expand the folded group to see the log file of test_fx_reinplace_pass 2023-03-31T05:58:48.3959694Z ##[group]PRINTING LOG FILE of test_fx_reinplace_pass (/var/lib/jenkins/workspace/test/test-reports/test_fx_reinplace_pass_6l2mm107.log) 2023-03-31T05:58:48.3960405Z Test results will be stored in test-reports/python-pytest/test_fx_reinplace_pass/test_fx_reinplace_pass-1d9175640db7e35d.xml 2023-03-31T05:58:48.3960779Z ============================= test session starts ============================== 2023-03-31T05:58:48.3961189Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:48.3961457Z cachedir: .pytest_cache 2023-03-31T05:58:48.3961890Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:48.3962262Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:48.3962704Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:48.3963147Z collecting ... collected 11 items 2023-03-31T05:58:48.3964682Z Running 11 items in this shard: test/test_fx_reinplace_pass.py::TestReinplacePass::test_out_node_updated, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_basic, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_different_metadata, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_index_mutation, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_overlapping_memory, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_op, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_invalid, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_invalid2, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_valid, test/test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_with_view 2023-03-31T05:58:48.3965855Z 2023-03-31T05:58:48.3966014Z test_fx_reinplace_pass.py::TestReinplacePass::test_out_node_updated PASSED [ 9%] 2023-03-31T05:58:48.3966368Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_basic PASSED [ 18%] 2023-03-31T05:58:48.3966798Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_different_metadata PASSED [ 27%] 2023-03-31T05:58:48.3967184Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_index_mutation PASSED [ 36%] 2023-03-31T05:58:48.3967549Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_overlapping_memory PASSED [ 45%] 2023-03-31T05:58:48.3967928Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_op PASSED [ 54%] 2023-03-31T05:58:48.3968292Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice PASSED [ 63%] 2023-03-31T05:58:48.3968690Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_invalid PASSED [ 72%] 2023-03-31T05:58:48.3969137Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_invalid2 PASSED [ 81%] 2023-03-31T05:58:48.3969571Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_scatter_twice_with_different_view_op_valid PASSED [ 90%] 2023-03-31T05:58:48.3969962Z test_fx_reinplace_pass.py::TestReinplacePass::test_reinplace_with_view PASSED [100%] 2023-03-31T05:58:48.3970158Z 2023-03-31T05:58:48.3970520Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fx_reinplace_pass/test_fx_reinplace_pass-1d9175640db7e35d.xml - 2023-03-31T05:58:48.3970898Z ============================== 11 passed in 0.70s ============================== 2023-03-31T05:58:48.3971276Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:48.3971533Z 2023-03-31T05:58:48.3971818Z ##[endgroup] 2023-03-31T05:58:48.3972354Z FINISHED PRINTING LOG FILE of test_fx_reinplace_pass (/var/lib/jenkins/workspace/test/test-reports/test_fx_reinplace_pass_6l2mm107.log) 2023-03-31T05:58:48.3972600Z 2023-03-31T05:58:50.5314270Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:50.5342577Z Ignoring disabled issues: [] 2023-03-31T05:58:50.5520291Z Running test_function_schema ... [2023-03-31 05:58:50.551668] 2023-03-31T05:58:50.5523241Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_function_schema.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:50.551979] 2023-03-31T05:58:53.0665537Z 2023-03-31T05:58:53.0666037Z Expand the folded group to see the log file of test_function_schema 2023-03-31T05:58:53.0667187Z ##[group]PRINTING LOG FILE of test_function_schema (/var/lib/jenkins/workspace/test/test-reports/test_function_schema_iu1g23g8.log) 2023-03-31T05:58:53.0667967Z Test results will be stored in test-reports/python-pytest/test_function_schema/test_function_schema-607a676d8867d6ae.xml 2023-03-31T05:58:53.0668420Z ============================= test session starts ============================== 2023-03-31T05:58:53.0668819Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:53.0669208Z cachedir: .pytest_cache 2023-03-31T05:58:53.0669906Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:53.0670292Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:53.0670819Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:53.0671133Z collecting ... collected 15 items 2023-03-31T05:58:53.0673022Z Running 15 items in this shard: test/test_function_schema.py::TestFunctionSchema::test_backward_compatible_arguments, test/test_function_schema.py::TestFunctionSchema::test_backward_compatible_outputs, test/test_function_schema.py::TestFunctionSchema::test_backward_compatible_structure, test/test_function_schema.py::TestFunctionSchema::test_backward_compatible_with_smart_serialization, test/test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_real_use_case, test/test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_with_out, test/test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_without_out, test/test_function_schema.py::TestFunctionSchema::test_hash_schema, test/test_function_schema.py::TestFunctionSchema::test_out_schema, test/test_function_schema.py::TestFunctionSchema::test_schema_error, test/test_function_schema.py::TestFunctionSchema::test_serialize_and_deserialize, test/test_function_schema.py::TestFunctionSchema::test_string_optional_parameter_default_value, test/test_function_schema.py::TestFunctionSchema::test_sym_int_argument_properly_parsed, test/test_function_schema.py::TestFunctionSchema::test_tensor_list_alias_annotation_properly_parsed, test/test_function_schema.py::TestFunctionSchema::test_tensor_option_arguments_properly_parsed 2023-03-31T05:58:53.0674924Z 2023-03-31T05:58:53.0675118Z test_function_schema.py::TestFunctionSchema::test_backward_compatible_arguments PASSED [ 6%] 2023-03-31T05:58:53.0675583Z test_function_schema.py::TestFunctionSchema::test_backward_compatible_outputs PASSED [ 13%] 2023-03-31T05:58:53.0675974Z test_function_schema.py::TestFunctionSchema::test_backward_compatible_structure PASSED [ 20%] 2023-03-31T05:58:53.0676456Z test_function_schema.py::TestFunctionSchema::test_backward_compatible_with_smart_serialization PASSED [ 26%] 2023-03-31T05:58:53.0676869Z test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_real_use_case PASSED [ 33%] 2023-03-31T05:58:53.0677411Z test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_with_out PASSED [ 40%] 2023-03-31T05:58:53.0677950Z test_function_schema.py::TestFunctionSchema::test_forward_compatible_arguments_without_out PASSED [ 46%] 2023-03-31T05:58:53.0678327Z test_function_schema.py::TestFunctionSchema::test_hash_schema PASSED [ 53%] 2023-03-31T05:58:53.0678740Z test_function_schema.py::TestFunctionSchema::test_out_schema PASSED [ 60%] 2023-03-31T05:58:53.0679082Z test_function_schema.py::TestFunctionSchema::test_schema_error PASSED [ 66%] 2023-03-31T05:58:53.0679509Z test_function_schema.py::TestFunctionSchema::test_serialize_and_deserialize PASSED [ 73%] 2023-03-31T05:58:53.0679912Z test_function_schema.py::TestFunctionSchema::test_string_optional_parameter_default_value PASSED [ 80%] 2023-03-31T05:58:53.0680390Z test_function_schema.py::TestFunctionSchema::test_sym_int_argument_properly_parsed PASSED [ 86%] 2023-03-31T05:58:53.0680795Z test_function_schema.py::TestFunctionSchema::test_tensor_list_alias_annotation_properly_parsed PASSED [ 93%] 2023-03-31T05:58:53.0681294Z test_function_schema.py::TestFunctionSchema::test_tensor_option_arguments_properly_parsed PASSED [100%] 2023-03-31T05:58:53.0681516Z 2023-03-31T05:58:53.0681965Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_function_schema/test_function_schema-607a676d8867d6ae.xml - 2023-03-31T05:58:53.0682331Z ============================== 15 passed in 0.59s ============================== 2023-03-31T05:58:53.0682712Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:53.0682937Z 2023-03-31T05:58:53.0683530Z ##[endgroup] 2023-03-31T05:58:53.0683995Z FINISHED PRINTING LOG FILE of test_function_schema (/var/lib/jenkins/workspace/test/test-reports/test_function_schema_iu1g23g8.log) 2023-03-31T05:58:53.0684242Z 2023-03-31T05:58:55.1712385Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:55.1741731Z Ignoring disabled issues: [] 2023-03-31T05:58:55.1917336Z Running test_fx_passes ... [2023-03-31 05:58:55.191382] 2023-03-31T05:58:55.1919883Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_fx_passes.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:55.191742] 2023-03-31T05:58:57.8365767Z 2023-03-31T05:58:57.8366348Z Expand the folded group to see the log file of test_fx_passes 2023-03-31T05:58:57.8372832Z ##[group]PRINTING LOG FILE of test_fx_passes (/var/lib/jenkins/workspace/test/test-reports/test_fx_passes_rnk341p9.log) 2023-03-31T05:58:57.8373593Z Test results will be stored in test-reports/python-pytest/test_fx_passes/test_fx_passes-a8aac7079acdb8e3.xml 2023-03-31T05:58:57.8374003Z ============================= test session starts ============================== 2023-03-31T05:58:57.8374414Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:58:57.8374755Z cachedir: .pytest_cache 2023-03-31T05:58:57.8375186Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:58:57.8375631Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:58:57.8376085Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:58:57.8376449Z collecting ... collected 51 items 2023-03-31T05:58:57.8391295Z Running 51 items in this shard: test/test_fx_passes.py::TestFXGraphPasses::test_fuser_pass_deep_model, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'add_1', 'add_2']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'add_1'], ['add_5', 'add_6']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'linear', 'add_1', 'param', 'add_2', 'add_3', 'add_4', 'linear2', 'add_5', 'add_6', 'relu']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_2', 'add_3']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_3', 'add_4']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_4', 'add_1', 'add_3', 'add_2']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_5', 'add_6'], ['add_1', 'add_2', 'add_3', 'add_4']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_5', 'linear2']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_6', 'add_5']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_6', 'relu']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['param', 'add_1', 'linear']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['param', 'add_2']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add', 'add_1', 'add_3']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add', 'add_1'], ['add_1', 'add_5', 'add_6']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add_4', 'add_5']], test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['relu', 'add_5']], test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_7', 'add_6'], ['add_5', 'add_4', 'add_3'], ['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1'], ['add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2'], ['add_3', 'add_4', 'add_1'], ['add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add', 'std_mean', 'getitem', 'getitem_1']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add', 'permute_1', 'view', 'permute_2', 'permute_3', 'permute']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['permute_1', 'add_1', 'add']]_bookend_non_compute_pass_True, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add', 'permute_1', 'view', 'permute_2', 'permute_3', 'permute']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['permute_1', 'add_1', 'add']]_bookend_non_compute_pass_True, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2'], ['add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add_1', 'add']]_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ 2023-03-31T05:58:57.8400584Z 2023-03-31T05:58:57.8400822Z test_fx_passes.py::TestFXGraphPasses::test_fuser_pass_deep_model PASSED [ 1%] 2023-03-31T05:58:57.8401315Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'add_1', 'add_2']] PASSED [ 3%] 2023-03-31T05:58:57.8401876Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'add_1'], ['add_5', 'add_6']] PASSED [ 5%] 2023-03-31T05:58:57.8402514Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add', 'linear', 'add_1', 'param', 'add_2', 'add_3', 'add_4', 'linear2', 'add_5', 'add_6', 'relu']] PASSED [ 7%] 2023-03-31T05:58:57.8403224Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_2', 'add_3']] PASSED [ 9%] 2023-03-31T05:58:57.8403693Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_3', 'add_4']] PASSED [ 11%] 2023-03-31T05:58:57.8404251Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_4', 'add_1', 'add_3', 'add_2']] PASSED [ 13%] 2023-03-31T05:58:57.8404836Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_5', 'add_6'], ['add_1', 'add_2', 'add_3', 'add_4']] PASSED [ 15%] 2023-03-31T05:58:57.8405346Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_5', 'linear2']] PASSED [ 17%] 2023-03-31T05:58:57.8405805Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_6', 'add_5']] PASSED [ 19%] 2023-03-31T05:58:57.8406330Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['add_6', 'relu']] PASSED [ 21%] 2023-03-31T05:58:57.8406813Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['param', 'add_1', 'linear']] PASSED [ 23%] 2023-03-31T05:58:57.8407270Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition_[['param', 'add_2']] PASSED [ 25%] 2023-03-31T05:58:57.8407747Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add', 'add_1', 'add_3']] PASSED [ 27%] 2023-03-31T05:58:57.8408262Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add', 'add_1'], ['add_1', 'add_5', 'add_6']] PASSED [ 29%] 2023-03-31T05:58:57.8408757Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['add_4', 'add_5']] PASSED [ 31%] 2023-03-31T05:58:57.8409216Z test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition_[['relu', 'add_5']] PASSED [ 33%] 2023-03-31T05:58:57.8410021Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_7', 'add_6'], ['add_5', 'add_4', 'add_3'], ['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 35%] 2023-03-31T05:58:57.8410894Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False PASSED [ 37%] 2023-03-31T05:58:57.8411677Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1'], ['add']]_bookend_non_compute_pass_False PASSED [ 39%] 2023-03-31T05:58:57.8412471Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2'], ['add_3', 'add_4', 'add_1'], ['add']]_bookend_non_compute_pass_False PASSED [ 41%] 2023-03-31T05:58:57.8413237Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 43%] 2023-03-31T05:58:57.8414091Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add', 'std_mean', 'getitem', 'getitem_1']]_bookend_non_compute_pass_False PASSED [ 45%] 2023-03-31T05:58:57.8414956Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add', 'permute_1', 'view', 'permute_2', 'permute_3', 'permute']]_bookend_non_compute_pass_False PASSED [ 47%] 2023-03-31T05:58:57.8415775Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['permute_1', 'add_1', 'add']]_bookend_non_compute_pass_True PASSED [ 49%] 2023-03-31T05:58:57.8416621Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add', 'permute_1', 'view', 'permute_2', 'permute_3', 'permute']]_bookend_non_compute_pass_False PASSED [ 50%] 2023-03-31T05:58:57.8417528Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['permute_1', 'add_1', 'add']]_bookend_non_compute_pass_True PASSED [ 52%] 2023-03-31T05:58:57.8418342Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2'], ['add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 54%] 2023-03-31T05:58:57.8419141Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 56%] 2023-03-31T05:58:57.8419911Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_2', 'add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 58%] 2023-03-31T05:58:57.8420662Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 60%] 2023-03-31T05:58:57.8421402Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add']]_bookend_non_compute_pass_False PASSED [ 62%] 2023-03-31T05:58:57.8422229Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False PASSED [ 64%] 2023-03-31T05:58:57.8423016Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add', 'add_1']]_bookend_non_compute_pass_False PASSED [ 66%] 2023-03-31T05:58:57.8423846Z test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn__expected_partition_[['add_3', 'add_2', 'add_1', 'add']]_bookend_non_compute_pass_False PASSED [ 68%] 2023-03-31T05:58:57.8424479Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 70%] 2023-03-31T05:58:57.8425063Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 72%] 2023-03-31T05:58:57.8425658Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 74%] 2023-03-31T05:58:57.8426276Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 76%] 2023-03-31T05:58:57.8426885Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 78%] 2023-03-31T05:58:57.8427574Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 80%] 2023-03-31T05:58:57.8428236Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 82%] 2023-03-31T05:58:57.8428868Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 84%] 2023-03-31T05:58:57.8429473Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 86%] 2023-03-31T05:58:57.8430032Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 88%] 2023-03-31T05:58:57.8430642Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 90%] 2023-03-31T05:58:57.8431206Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 92%] 2023-03-31T05:58:57.8431755Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 94%] 2023-03-31T05:58:57.8432307Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 96%] 2023-03-31T05:58:57.8432883Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [ 98%] 2023-03-31T05:58:57.8433765Z test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model_ PASSED [100%]/var/lib/jenkins/workspace/test/test_fx_passes.py:22: PytestCollectionWarning: cannot collect test class 'TestModule' because it has a __init__ constructor (from: test/test_fx_passes.py) 2023-03-31T05:58:57.8434298Z class TestModule(torch.nn.Module): 2023-03-31T05:58:57.8434863Z /var/lib/jenkins/workspace/test/test_fx_passes.py:47: PytestCollectionWarning: cannot collect test class 'TestDeepModule' because it has a __init__ constructor (from: test/test_fx_passes.py) 2023-03-31T05:58:57.8435261Z class TestDeepModule(torch.nn.Module): 2023-03-31T05:58:57.8435807Z /var/lib/jenkins/workspace/test/test_fx_passes.py:366: PytestCollectionWarning: cannot collect test class 'TestCase' because it has a __init__ constructor (from: test/test_fx_passes.py) 2023-03-31T05:58:57.8436172Z class TestCase: 2023-03-31T05:58:57.8436290Z 2023-03-31T05:58:57.8436295Z 2023-03-31T05:58:57.8436642Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fx_passes/test_fx_passes-a8aac7079acdb8e3.xml - 2023-03-31T05:58:57.8436996Z ============================== 51 passed in 0.67s ============================== 2023-03-31T05:58:57.8437316Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:58:57.8437520Z 2023-03-31T05:58:57.8437821Z ##[endgroup] 2023-03-31T05:58:57.8438261Z FINISHED PRINTING LOG FILE of test_fx_passes (/var/lib/jenkins/workspace/test/test-reports/test_fx_passes_rnk341p9.log) 2023-03-31T05:58:57.8438491Z 2023-03-31T05:58:59.9123940Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:58:59.9151684Z Ignoring disabled issues: [] 2023-03-31T05:58:59.9327828Z Running dynamo/test_global ... [2023-03-31 05:58:59.932416] 2023-03-31T05:58:59.9329249Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_global.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:58:59.932709] 2023-03-31T05:59:02.5529917Z 2023-03-31T05:59:02.5530512Z Expand the folded group to see the log file of dynamo/test_global 2023-03-31T05:59:02.5531789Z ##[group]PRINTING LOG FILE of dynamo/test_global (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_global_vzcgkmyv.log) 2023-03-31T05:59:02.5533228Z Test results will be stored in test-reports/python-pytest/dynamo.test_global/dynamo.test_global-79ff908ccec5af93.xml 2023-03-31T05:59:02.5533840Z ============================= test session starts ============================== 2023-03-31T05:59:02.5534518Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:02.5534981Z cachedir: .pytest_cache 2023-03-31T05:59:02.5535772Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:02.5536432Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:02.5537176Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:02.5537735Z collecting ... collected 11 items 2023-03-31T05:59:02.5539564Z Running 11 items in this shard: test/dynamo/test_global.py::TestGlobals::test_store_global_1, test/dynamo/test_global.py::TestGlobals::test_store_global_2, test/dynamo/test_global.py::TestGlobals::test_store_global_cross_file, test/dynamo/test_global.py::TestGlobals::test_store_global_dict, test/dynamo/test_global.py::TestGlobals::test_store_global_dict_2, test/dynamo/test_global.py::TestGlobals::test_store_global_inline_1, test/dynamo/test_global.py::TestGlobals::test_store_global_inline_2, test/dynamo/test_global.py::TestGlobals::test_store_global_list, test/dynamo/test_global.py::TestGlobals::test_store_global_list_2, test/dynamo/test_global.py::TestGlobals::test_store_global_new, test/dynamo/test_global.py::TestGlobals::test_store_global_object 2023-03-31T05:59:02.5541124Z 2023-03-31T05:59:02.5541409Z dynamo/test_global.py::TestGlobals::test_store_global_1 PASSED [ 9%] 2023-03-31T05:59:02.5542121Z dynamo/test_global.py::TestGlobals::test_store_global_2 PASSED [ 18%] 2023-03-31T05:59:02.5542693Z dynamo/test_global.py::TestGlobals::test_store_global_cross_file PASSED [ 27%] 2023-03-31T05:59:02.5543272Z dynamo/test_global.py::TestGlobals::test_store_global_dict PASSED [ 36%] 2023-03-31T05:59:02.5543834Z dynamo/test_global.py::TestGlobals::test_store_global_dict_2 PASSED [ 45%] 2023-03-31T05:59:02.5544426Z dynamo/test_global.py::TestGlobals::test_store_global_inline_1 PASSED [ 54%] 2023-03-31T05:59:02.5545007Z dynamo/test_global.py::TestGlobals::test_store_global_inline_2 PASSED [ 63%] 2023-03-31T05:59:02.5545604Z dynamo/test_global.py::TestGlobals::test_store_global_list PASSED [ 72%] 2023-03-31T05:59:02.5546194Z dynamo/test_global.py::TestGlobals::test_store_global_list_2 PASSED [ 81%] 2023-03-31T05:59:02.5546752Z dynamo/test_global.py::TestGlobals::test_store_global_new PASSED [ 90%] 2023-03-31T05:59:02.5547305Z dynamo/test_global.py::TestGlobals::test_store_global_object PASSED [100%] 2023-03-31T05:59:02.5547615Z 2023-03-31T05:59:02.5548282Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_global/dynamo.test_global-79ff908ccec5af93.xml - 2023-03-31T05:59:02.5549093Z ============================== 11 passed in 0.55s ============================== 2023-03-31T05:59:02.5549639Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:02.5549962Z 2023-03-31T05:59:02.5550406Z ##[endgroup] 2023-03-31T05:59:02.5551198Z FINISHED PRINTING LOG FILE of dynamo/test_global (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_global_vzcgkmyv.log) 2023-03-31T05:59:02.5551643Z 2023-03-31T05:59:04.7166178Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:04.7193355Z Ignoring disabled issues: [] 2023-03-31T05:59:04.7368162Z Running export/test_export ... [2023-03-31 05:59:04.736512] 2023-03-31T05:59:04.7370787Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'export/test_export.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:04.736836] 2023-03-31T05:59:07.3361671Z 2023-03-31T05:59:07.3362275Z Expand the folded group to see the log file of export/test_export 2023-03-31T05:59:07.3363613Z ##[group]PRINTING LOG FILE of export/test_export (/var/lib/jenkins/workspace/test/test-reports/export-test_export_fagay4d7.log) 2023-03-31T05:59:07.3364751Z Test results will be stored in test-reports/python-pytest/export.test_export/export.test_export-3dec0aa1ac9f692a.xml 2023-03-31T05:59:07.3365395Z ============================= test session starts ============================== 2023-03-31T05:59:07.3366127Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:07.3366631Z cachedir: .pytest_cache 2023-03-31T05:59:07.3367414Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:07.3368098Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:07.3368939Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:07.3369515Z collecting ... collected 4 items 2023-03-31T05:59:07.3370479Z Running 4 items in this shard: test/export/test_export.py::TestExport::test_export_cond, test/export/test_export.py::TestExport::test_export_simple_model, test/export/test_export.py::TestExport::test_export_simple_model_buffer_mutation, test/export/test_export.py::TestExport::test_export_simple_model_with_attr 2023-03-31T05:59:07.3371265Z 2023-03-31T05:59:07.3372161Z export/test_export.py::TestExport::test_export_cond SKIPPED (dynamo failure -> RuntimeError: Could not infer dtype of SymBool) [ 25%] 2023-03-31T05:59:07.3372978Z export/test_export.py::TestExport::test_export_simple_model PASSED [ 50%] 2023-03-31T05:59:07.3373747Z export/test_export.py::TestExport::test_export_simple_model_buffer_mutation SKIPPED (TypeError: () missing 1 required positional argument) [ 75%] 2023-03-31T05:59:07.3374668Z export/test_export.py::TestExport::test_export_simple_model_with_attr SKIPPED (TypeError: () missing 1 required positional argument) [100%] 2023-03-31T05:59:07.3375144Z 2023-03-31T05:59:07.3375809Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_export/export.test_export-3dec0aa1ac9f692a.xml - 2023-03-31T05:59:07.3376510Z ========================= 1 passed, 3 skipped in 0.52s ========================= 2023-03-31T05:59:07.3377103Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:07.3377478Z 2023-03-31T05:59:07.3377895Z ##[endgroup] 2023-03-31T05:59:07.3378696Z FINISHED PRINTING LOG FILE of export/test_export (/var/lib/jenkins/workspace/test/test-reports/export-test_export_fagay4d7.log) 2023-03-31T05:59:07.3379131Z 2023-03-31T05:59:09.4104648Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:09.4133963Z Ignoring disabled issues: [] 2023-03-31T05:59:09.4308709Z Running dynamo/test_interop ... [2023-03-31 05:59:09.430448] 2023-03-31T05:59:09.4310277Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_interop.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:09.430776] 2023-03-31T05:59:11.9848395Z 2023-03-31T05:59:11.9848934Z Expand the folded group to see the log file of dynamo/test_interop 2023-03-31T05:59:11.9850217Z ##[group]PRINTING LOG FILE of dynamo/test_interop (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_interop_61gmdgpt.log) 2023-03-31T05:59:11.9850962Z Test results will be stored in test-reports/python-pytest/dynamo.test_interop/dynamo.test_interop-d1995c4b5c6edeee.xml 2023-03-31T05:59:11.9851330Z ============================= test session starts ============================== 2023-03-31T05:59:11.9851789Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:11.9852341Z cachedir: .pytest_cache 2023-03-31T05:59:11.9861474Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:11.9861978Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:11.9862433Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:11.9862847Z collecting ... collected 3 items 2023-03-31T05:59:11.9863283Z Running 3 items in this shard: test/dynamo/test_interop.py::InteropTests::test_fx_fn, test/dynamo/test_interop.py::InteropTests::test_script_fn, test/dynamo/test_interop.py::InteropTests::test_trace_fn 2023-03-31T05:59:11.9863587Z 2023-03-31T05:59:11.9863741Z dynamo/test_interop.py::InteropTests::test_fx_fn PASSED [ 33%] 2023-03-31T05:59:11.9864051Z dynamo/test_interop.py::InteropTests::test_script_fn PASSED [ 66%] 2023-03-31T05:59:11.9864379Z dynamo/test_interop.py::InteropTests::test_trace_fn PASSED [100%] 2023-03-31T05:59:11.9864557Z 2023-03-31T05:59:11.9864920Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_interop/dynamo.test_interop-d1995c4b5c6edeee.xml - 2023-03-31T05:59:11.9865301Z ============================== 3 passed in 0.52s =============================== 2023-03-31T05:59:11.9865606Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:11.9865808Z 2023-03-31T05:59:11.9866271Z ##[endgroup] 2023-03-31T05:59:11.9866762Z FINISHED PRINTING LOG FILE of dynamo/test_interop (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_interop_61gmdgpt.log) 2023-03-31T05:59:11.9867008Z 2023-03-31T05:59:14.0661003Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:14.0688441Z Ignoring disabled issues: [] 2023-03-31T05:59:14.0864176Z Running functorch/test_minifier ... [2023-03-31 05:59:14.086080] 2023-03-31T05:59:14.0867222Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'functorch/test_minifier.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:14.086423] 2023-03-31T05:59:16.6168861Z 2023-03-31T05:59:16.6169442Z Expand the folded group to see the log file of functorch/test_minifier 2023-03-31T05:59:16.6170843Z ##[group]PRINTING LOG FILE of functorch/test_minifier (/var/lib/jenkins/workspace/test/test-reports/functorch-test_minifier_5y1vins0.log) 2023-03-31T05:59:16.6172027Z Test results will be stored in test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-98c50059c8acff27.xml 2023-03-31T05:59:16.6172655Z ============================= test session starts ============================== 2023-03-31T05:59:16.6173355Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:16.6173912Z cachedir: .pytest_cache 2023-03-31T05:59:16.6174760Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:16.6175737Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:16.6176612Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:16.6177211Z collecting ... collected 5 items 2023-03-31T05:59:16.6178378Z Running 5 items in this shard: test/functorch/test_minifier.py::TestMinifier::test_has_add_mul, test/functorch/test_minifier.py::TestMinifier::test_has_mul_minifier, test/functorch/test_minifier.py::TestMinifier::test_input_returned, test/functorch/test_minifier.py::TestMinifier::test_module, test/functorch/test_minifier.py::TestMinifier::test_tup_use 2023-03-31T05:59:16.6179308Z 2023-03-31T05:59:16.6179590Z functorch/test_minifier.py::TestMinifier::test_has_add_mul PASSED [ 20%] 2023-03-31T05:59:16.6180366Z functorch/test_minifier.py::TestMinifier::test_has_mul_minifier PASSED [ 40%] 2023-03-31T05:59:16.6181020Z functorch/test_minifier.py::TestMinifier::test_input_returned PASSED [ 60%] 2023-03-31T05:59:16.6181651Z functorch/test_minifier.py::TestMinifier::test_module PASSED [ 80%] 2023-03-31T05:59:16.6182258Z functorch/test_minifier.py::TestMinifier::test_tup_use PASSED [100%] 2023-03-31T05:59:16.6182722Z 2023-03-31T05:59:16.6183488Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-98c50059c8acff27.xml - 2023-03-31T05:59:16.6184240Z ============================== 5 passed in 0.48s =============================== 2023-03-31T05:59:16.6184841Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:16.6185229Z 2023-03-31T05:59:16.6185698Z ##[endgroup] 2023-03-31T05:59:16.6186561Z FINISHED PRINTING LOG FILE of functorch/test_minifier (/var/lib/jenkins/workspace/test/test-reports/functorch-test_minifier_5y1vins0.log) 2023-03-31T05:59:16.6187058Z 2023-03-31T05:59:18.7020739Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:18.7048473Z Ignoring disabled issues: [] 2023-03-31T05:59:18.7225459Z Running test_subclass ... [2023-03-31 05:59:18.722179] 2023-03-31T05:59:18.7228274Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_subclass.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:18.722514] 2023-03-31T05:59:21.1645487Z 2023-03-31T05:59:21.1646007Z Expand the folded group to see the log file of test_subclass 2023-03-31T05:59:21.1646995Z ##[group]PRINTING LOG FILE of test_subclass (/var/lib/jenkins/workspace/test/test-reports/test_subclass_65qkgcbx.log) 2023-03-31T05:59:21.1648089Z Test results will be stored in test-reports/python-pytest/test_subclass/test_subclass-fb879363910a05f5.xml 2023-03-31T05:59:21.1648726Z ============================= test session starts ============================== 2023-03-31T05:59:21.1649472Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:21.1649965Z cachedir: .pytest_cache 2023-03-31T05:59:21.1650778Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:21.1651466Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:21.1652307Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:21.1652865Z collecting ... collected 71 items 2023-03-31T05:59:21.1667324Z Running 71 items in this shard: test/test_subclass.py::TestSubclass::test_deepcopy_base_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_deepcopy_base_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_deepcopy_diag_tensor_below_as_param_False, test/test_subclass.py::TestSubclass::test_deepcopy_diag_tensor_below_as_param_True, test/test_subclass.py::TestSubclass::test_deepcopy_logging_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_deepcopy_logging_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_deepcopy_non_wrapper_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_deepcopy_non_wrapper_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_deepcopy_sparse_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_deepcopy_sparse_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_lazy_module_base_tensor, test/test_subclass.py::TestSubclass::test_lazy_module_diag_tensor_below, test/test_subclass.py::TestSubclass::test_lazy_module_logging_tensor, test/test_subclass.py::TestSubclass::test_lazy_module_non_wrapper_tensor, test/test_subclass.py::TestSubclass::test_lazy_module_sparse_tensor, test/test_subclass.py::TestSubclass::test_module_optimization_base_tensor, test/test_subclass.py::TestSubclass::test_module_optimization_diag_tensor_below, test/test_subclass.py::TestSubclass::test_module_optimization_logging_tensor, test/test_subclass.py::TestSubclass::test_module_optimization_non_wrapper_tensor, test/test_subclass.py::TestSubclass::test_module_optimization_sparse_tensor, test/test_subclass.py::TestSubclass::test_non_rewrapping_torch_dispatch_subclass_as_parameter_throws_for_detach, test/test_subclass.py::TestSubclass::test_param_invariants_base_tensor_tensor_requires_grad_False, test/test_subclass.py::TestSubclass::test_param_invariants_base_tensor_tensor_requires_grad_True, test/test_subclass.py::TestSubclass::test_param_invariants_diag_tensor_below_tensor_requires_grad_False, test/test_subclass.py::TestSubclass::test_param_invariants_diag_tensor_below_tensor_requires_grad_True, test/test_subclass.py::TestSubclass::test_param_invariants_logging_tensor_tensor_requires_grad_False, test/test_subclass.py::TestSubclass::test_param_invariants_logging_tensor_tensor_requires_grad_True, test/test_subclass.py::TestSubclass::test_param_invariants_non_wrapper_tensor_tensor_requires_grad_False, test/test_subclass.py::TestSubclass::test_param_invariants_non_wrapper_tensor_tensor_requires_grad_True, test/test_subclass.py::TestSubclass::test_param_invariants_sparse_tensor_tensor_requires_grad_False, test/test_subclass.py::TestSubclass::test_param_invariants_sparse_tensor_tensor_requires_grad_True, test/test_subclass.py::TestSubclass::test_parametrization_base_tensor_leave_parametrized_False, test/test_subclass.py::TestSubclass::test_parametrization_base_tensor_leave_parametrized_True, test/test_subclass.py::TestSubclass::test_parametrization_diag_tensor_below_leave_parametrized_False, test/test_subclass.py::TestSubclass::test_parametrization_diag_tensor_below_leave_parametrized_True, test/test_subclass.py::TestSubclass::test_parametrization_logging_tensor_leave_parametrized_False, test/test_subclass.py::TestSubclass::test_parametrization_logging_tensor_leave_parametrized_True, test/test_subclass.py::TestSubclass::test_parametrization_non_wrapper_tensor_leave_parametrized_False, test/test_subclass.py::TestSubclass::test_parametrization_non_wrapper_tensor_leave_parametrized_True, test/test_subclass.py::TestSubclass::test_parametrization_sparse_tensor_leave_parametrized_False, test/test_subclass.py::TestSubclass::test_parametrization_sparse_tensor_leave_parametrized_True, test/test_subclass.py::TestSubclass::test_repr_base_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_repr_base_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_repr_diag_tensor_below_as_param_False, test/test_subclass.py::TestSubclass::test_repr_diag_tensor_below_as_param_True, test/test_subclass.py::TestSubclass::test_repr_logging_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_repr_logging_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_repr_non_wrapper_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_repr_non_wrapper_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_repr_sparse_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_repr_sparse_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_serialization_base_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_serialization_base_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_serialization_diag_tensor_below_as_param_False, test/test_subclass.py::TestSubclass::test_serialization_diag_tensor_below_as_param_True, test/test_subclass.py::TestSubclass::test_serialization_logging_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_serialization_logging_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_serialization_non_wrapper_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_serialization_non_wrapper_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_serialization_sparse_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_serialization_sparse_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_type_propagation_base_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_type_propagation_base_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_type_propagation_diag_tensor_below_as_param_False, test/test_subclass.py::TestSubclass::test_type_propagation_diag_tensor_below_as_param_True, test/test_subclass.py::TestSubclass::test_type_propagation_logging_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_type_propagation_logging_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_type_propagation_non_wrapper_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_type_propagation_non_wrapper_tensor_as_param_True, test/test_subclass.py::TestSubclass::test_type_propagation_sparse_tensor_as_param_False, test/test_subclass.py::TestSubclass::test_type_propagation_sparse_tensor_as_param_True 2023-03-31T05:59:21.1680701Z 2023-03-31T05:59:21.1681010Z test_subclass.py::TestSubclass::test_deepcopy_base_tensor_as_param_False PASSED [ 1%] 2023-03-31T05:59:21.1681697Z test_subclass.py::TestSubclass::test_deepcopy_base_tensor_as_param_True PASSED [ 2%] 2023-03-31T05:59:21.1682381Z test_subclass.py::TestSubclass::test_deepcopy_diag_tensor_below_as_param_False PASSED [ 4%] 2023-03-31T05:59:21.1683310Z test_subclass.py::TestSubclass::test_deepcopy_diag_tensor_below_as_param_True PASSED [ 5%] 2023-03-31T05:59:21.1683860Z test_subclass.py::TestSubclass::test_deepcopy_logging_tensor_as_param_False PASSED [ 7%] 2023-03-31T05:59:21.1684472Z test_subclass.py::TestSubclass::test_deepcopy_logging_tensor_as_param_True PASSED [ 8%] 2023-03-31T05:59:21.1684892Z test_subclass.py::TestSubclass::test_deepcopy_non_wrapper_tensor_as_param_False PASSED [ 9%] 2023-03-31T05:59:21.1685265Z test_subclass.py::TestSubclass::test_deepcopy_non_wrapper_tensor_as_param_True PASSED [ 11%] 2023-03-31T05:59:21.1685618Z test_subclass.py::TestSubclass::test_deepcopy_sparse_tensor_as_param_False PASSED [ 12%] 2023-03-31T05:59:21.1685978Z test_subclass.py::TestSubclass::test_deepcopy_sparse_tensor_as_param_True PASSED [ 14%] 2023-03-31T05:59:21.1686324Z test_subclass.py::TestSubclass::test_lazy_module_base_tensor XFAIL [ 15%] 2023-03-31T05:59:21.1686649Z test_subclass.py::TestSubclass::test_lazy_module_diag_tensor_below XFAIL [ 16%] 2023-03-31T05:59:21.1686978Z test_subclass.py::TestSubclass::test_lazy_module_logging_tensor XFAIL [ 18%] 2023-03-31T05:59:21.1687310Z test_subclass.py::TestSubclass::test_lazy_module_non_wrapper_tensor XFAIL [ 19%] 2023-03-31T05:59:21.1687645Z test_subclass.py::TestSubclass::test_lazy_module_sparse_tensor XFAIL [ 21%] 2023-03-31T05:59:21.1687974Z test_subclass.py::TestSubclass::test_module_optimization_base_tensor PASSED [ 22%] 2023-03-31T05:59:21.1688326Z test_subclass.py::TestSubclass::test_module_optimization_diag_tensor_below PASSED [ 23%] 2023-03-31T05:59:21.1688684Z test_subclass.py::TestSubclass::test_module_optimization_logging_tensor PASSED [ 25%] 2023-03-31T05:59:21.1689030Z test_subclass.py::TestSubclass::test_module_optimization_non_wrapper_tensor PASSED [ 26%] 2023-03-31T05:59:21.1689390Z test_subclass.py::TestSubclass::test_module_optimization_sparse_tensor PASSED [ 28%] 2023-03-31T05:59:21.1689828Z test_subclass.py::TestSubclass::test_non_rewrapping_torch_dispatch_subclass_as_parameter_throws_for_detach PASSED [ 29%] 2023-03-31T05:59:21.1690245Z test_subclass.py::TestSubclass::test_param_invariants_base_tensor_tensor_requires_grad_False PASSED [ 30%] 2023-03-31T05:59:21.1690627Z test_subclass.py::TestSubclass::test_param_invariants_base_tensor_tensor_requires_grad_True PASSED [ 32%] 2023-03-31T05:59:21.1691029Z test_subclass.py::TestSubclass::test_param_invariants_diag_tensor_below_tensor_requires_grad_False PASSED [ 33%] 2023-03-31T05:59:21.1691435Z test_subclass.py::TestSubclass::test_param_invariants_diag_tensor_below_tensor_requires_grad_True PASSED [ 35%] 2023-03-31T05:59:21.1691837Z test_subclass.py::TestSubclass::test_param_invariants_logging_tensor_tensor_requires_grad_False PASSED [ 36%] 2023-03-31T05:59:21.1692269Z test_subclass.py::TestSubclass::test_param_invariants_logging_tensor_tensor_requires_grad_True PASSED [ 38%] 2023-03-31T05:59:21.1692677Z test_subclass.py::TestSubclass::test_param_invariants_non_wrapper_tensor_tensor_requires_grad_False PASSED [ 39%] 2023-03-31T05:59:21.1693082Z test_subclass.py::TestSubclass::test_param_invariants_non_wrapper_tensor_tensor_requires_grad_True PASSED [ 40%] 2023-03-31T05:59:21.1693473Z test_subclass.py::TestSubclass::test_param_invariants_sparse_tensor_tensor_requires_grad_False PASSED [ 42%] 2023-03-31T05:59:21.1693871Z test_subclass.py::TestSubclass::test_param_invariants_sparse_tensor_tensor_requires_grad_True PASSED [ 43%] 2023-03-31T05:59:21.1694272Z test_subclass.py::TestSubclass::test_parametrization_base_tensor_leave_parametrized_False PASSED [ 45%] 2023-03-31T05:59:21.1694666Z test_subclass.py::TestSubclass::test_parametrization_base_tensor_leave_parametrized_True PASSED [ 46%] 2023-03-31T05:59:21.1695060Z test_subclass.py::TestSubclass::test_parametrization_diag_tensor_below_leave_parametrized_False PASSED [ 47%] 2023-03-31T05:59:21.1695471Z test_subclass.py::TestSubclass::test_parametrization_diag_tensor_below_leave_parametrized_True PASSED [ 49%] 2023-03-31T05:59:21.1695878Z test_subclass.py::TestSubclass::test_parametrization_logging_tensor_leave_parametrized_False PASSED [ 50%] 2023-03-31T05:59:21.1696280Z test_subclass.py::TestSubclass::test_parametrization_logging_tensor_leave_parametrized_True PASSED [ 52%] 2023-03-31T05:59:21.1696716Z test_subclass.py::TestSubclass::test_parametrization_non_wrapper_tensor_leave_parametrized_False PASSED [ 53%] 2023-03-31T05:59:21.1697159Z test_subclass.py::TestSubclass::test_parametrization_non_wrapper_tensor_leave_parametrized_True PASSED [ 54%] 2023-03-31T05:59:21.1697561Z test_subclass.py::TestSubclass::test_parametrization_sparse_tensor_leave_parametrized_False PASSED [ 56%] 2023-03-31T05:59:21.1697958Z test_subclass.py::TestSubclass::test_parametrization_sparse_tensor_leave_parametrized_True PASSED [ 57%] 2023-03-31T05:59:21.1698322Z test_subclass.py::TestSubclass::test_repr_base_tensor_as_param_False PASSED [ 59%] 2023-03-31T05:59:21.1698668Z test_subclass.py::TestSubclass::test_repr_base_tensor_as_param_True PASSED [ 60%] 2023-03-31T05:59:21.1699016Z test_subclass.py::TestSubclass::test_repr_diag_tensor_below_as_param_False PASSED [ 61%] 2023-03-31T05:59:21.1699359Z test_subclass.py::TestSubclass::test_repr_diag_tensor_below_as_param_True PASSED [ 63%] 2023-03-31T05:59:21.1699705Z test_subclass.py::TestSubclass::test_repr_logging_tensor_as_param_False PASSED [ 64%] 2023-03-31T05:59:21.1700056Z test_subclass.py::TestSubclass::test_repr_logging_tensor_as_param_True PASSED [ 66%] 2023-03-31T05:59:21.1700410Z test_subclass.py::TestSubclass::test_repr_non_wrapper_tensor_as_param_False PASSED [ 67%] 2023-03-31T05:59:21.1700751Z test_subclass.py::TestSubclass::test_repr_non_wrapper_tensor_as_param_True PASSED [ 69%] 2023-03-31T05:59:21.1701098Z test_subclass.py::TestSubclass::test_repr_sparse_tensor_as_param_False PASSED [ 70%] 2023-03-31T05:59:21.1701443Z test_subclass.py::TestSubclass::test_repr_sparse_tensor_as_param_True PASSED [ 71%] 2023-03-31T05:59:21.1701823Z test_subclass.py::TestSubclass::test_serialization_base_tensor_as_param_False PASSED [ 73%] 2023-03-31T05:59:21.1702183Z test_subclass.py::TestSubclass::test_serialization_base_tensor_as_param_True PASSED [ 74%] 2023-03-31T05:59:21.1702655Z test_subclass.py::TestSubclass::test_serialization_diag_tensor_below_as_param_False PASSED [ 76%] 2023-03-31T05:59:21.1703040Z test_subclass.py::TestSubclass::test_serialization_diag_tensor_below_as_param_True PASSED [ 77%] 2023-03-31T05:59:21.1703402Z test_subclass.py::TestSubclass::test_serialization_logging_tensor_as_param_False PASSED [ 78%] 2023-03-31T05:59:21.1703770Z test_subclass.py::TestSubclass::test_serialization_logging_tensor_as_param_True PASSED [ 80%] 2023-03-31T05:59:21.1704154Z test_subclass.py::TestSubclass::test_serialization_non_wrapper_tensor_as_param_False PASSED [ 81%] 2023-03-31T05:59:21.1704537Z test_subclass.py::TestSubclass::test_serialization_non_wrapper_tensor_as_param_True PASSED [ 83%] 2023-03-31T05:59:21.1704953Z test_subclass.py::TestSubclass::test_serialization_sparse_tensor_as_param_False PASSED [ 84%] 2023-03-31T05:59:21.1705320Z test_subclass.py::TestSubclass::test_serialization_sparse_tensor_as_param_True PASSED [ 85%] 2023-03-31T05:59:21.1705688Z test_subclass.py::TestSubclass::test_type_propagation_base_tensor_as_param_False PASSED [ 87%] 2023-03-31T05:59:21.1706048Z test_subclass.py::TestSubclass::test_type_propagation_base_tensor_as_param_True PASSED [ 88%] 2023-03-31T05:59:21.1706431Z test_subclass.py::TestSubclass::test_type_propagation_diag_tensor_below_as_param_False PASSED [ 90%] 2023-03-31T05:59:21.1706820Z test_subclass.py::TestSubclass::test_type_propagation_diag_tensor_below_as_param_True PASSED [ 91%] 2023-03-31T05:59:21.1707204Z test_subclass.py::TestSubclass::test_type_propagation_logging_tensor_as_param_False PASSED [ 92%] 2023-03-31T05:59:21.1707569Z test_subclass.py::TestSubclass::test_type_propagation_logging_tensor_as_param_True PASSED [ 94%] 2023-03-31T05:59:21.1707957Z test_subclass.py::TestSubclass::test_type_propagation_non_wrapper_tensor_as_param_False PASSED [ 95%] 2023-03-31T05:59:21.1708348Z test_subclass.py::TestSubclass::test_type_propagation_non_wrapper_tensor_as_param_True PASSED [ 97%] 2023-03-31T05:59:21.1708731Z test_subclass.py::TestSubclass::test_type_propagation_sparse_tensor_as_param_False PASSED [ 98%] 2023-03-31T05:59:21.1709096Z test_subclass.py::TestSubclass::test_type_propagation_sparse_tensor_as_param_True PASSED [100%] 2023-03-31T05:59:21.1709339Z 2023-03-31T05:59:21.1709804Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_subclass/test_subclass-fb879363910a05f5.xml - 2023-03-31T05:59:21.1710178Z ======================== 66 passed, 5 xfailed in 0.50s ========================= 2023-03-31T05:59:21.1710491Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:21.1710699Z 2023-03-31T05:59:21.1711038Z ##[endgroup] 2023-03-31T05:59:21.1711449Z FINISHED PRINTING LOG FILE of test_subclass (/var/lib/jenkins/workspace/test/test-reports/test_subclass_65qkgcbx.log) 2023-03-31T05:59:21.1711680Z 2023-03-31T05:59:23.2928897Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:23.2955907Z Ignoring disabled issues: [] 2023-03-31T05:59:23.3134209Z Running test_native_functions ... [2023-03-31 05:59:23.312925] 2023-03-31T05:59:23.3135411Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_native_functions.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:23.313254] 2023-03-31T05:59:25.7116165Z 2023-03-31T05:59:25.7116917Z Expand the folded group to see the log file of test_native_functions 2023-03-31T05:59:25.7118455Z ##[group]PRINTING LOG FILE of test_native_functions (/var/lib/jenkins/workspace/test/test-reports/test_native_functions_6i4_lwry.log) 2023-03-31T05:59:25.7119953Z Test results will be stored in test-reports/python-pytest/test_native_functions/test_native_functions-669237d2b8594e97.xml 2023-03-31T05:59:25.7121049Z ============================= test session starts ============================== 2023-03-31T05:59:25.7122158Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:25.7125554Z cachedir: .pytest_cache 2023-03-31T05:59:25.7126535Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:25.7127226Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:25.7127744Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:25.7128110Z collecting ... collected 11 items 2023-03-31T05:59:25.7129446Z Running 11 items in this shard: test/test_native_functions.py::TestNativeFunctions::test_intlist_error_with_overload, test/test_native_functions.py::TestNativeFunctions::test_optional_filled_intlist, test/test_native_functions.py::TestNativeFunctions::test_optional_floatlist, test/test_native_functions.py::TestNativeFunctions::test_optional_floatlist_invalid, test/test_native_functions.py::TestNativeFunctions::test_optional_intlist, test/test_native_functions.py::TestNativeFunctions::test_optional_intlist_invalid, test/test_native_functions.py::TestNativeFunctions::test_string_defaults, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload_but_is_unique, test/test_native_functions.py::TestNativeFunctions::test_vararg_symintlist_error 2023-03-31T05:59:25.7130795Z 2023-03-31T05:59:25.7130987Z test_native_functions.py::TestNativeFunctions::test_intlist_error_with_overload PASSED [ 9%] 2023-03-31T05:59:25.7131437Z test_native_functions.py::TestNativeFunctions::test_optional_filled_intlist PASSED [ 18%] 2023-03-31T05:59:25.7131831Z test_native_functions.py::TestNativeFunctions::test_optional_floatlist PASSED [ 27%] 2023-03-31T05:59:25.7132229Z test_native_functions.py::TestNativeFunctions::test_optional_floatlist_invalid PASSED [ 36%] 2023-03-31T05:59:25.7132646Z test_native_functions.py::TestNativeFunctions::test_optional_intlist PASSED [ 45%] 2023-03-31T05:59:25.7133021Z test_native_functions.py::TestNativeFunctions::test_optional_intlist_invalid PASSED [ 54%] 2023-03-31T05:59:25.7133497Z test_native_functions.py::TestNativeFunctions::test_string_defaults PASSED [ 63%] 2023-03-31T05:59:25.7133904Z test_native_functions.py::TestNativeFunctions::test_symintlist_error PASSED [ 72%] 2023-03-31T05:59:25.7134340Z test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload PASSED [ 81%] 2023-03-31T05:59:25.7134730Z test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload_but_is_unique PASSED [ 90%] 2023-03-31T05:59:25.7135198Z test_native_functions.py::TestNativeFunctions::test_vararg_symintlist_error PASSED [100%] 2023-03-31T05:59:25.7135403Z 2023-03-31T05:59:25.7135855Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_native_functions/test_native_functions-669237d2b8594e97.xml - 2023-03-31T05:59:25.7136237Z ============================== 11 passed in 0.45s ============================== 2023-03-31T05:59:25.7136615Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:25.7136825Z 2023-03-31T05:59:25.7137120Z ##[endgroup] 2023-03-31T05:59:25.7137611Z FINISHED PRINTING LOG FILE of test_native_functions (/var/lib/jenkins/workspace/test/test-reports/test_native_functions_6i4_lwry.log) 2023-03-31T05:59:25.7137852Z 2023-03-31T05:59:27.8384222Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:27.8411112Z Ignoring disabled issues: [] 2023-03-31T05:59:27.8587627Z Running test_autocast ... [2023-03-31 05:59:27.858386] 2023-03-31T05:59:27.8589622Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_autocast.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:27.858706] 2023-03-31T05:59:30.2286505Z 2023-03-31T05:59:30.2287024Z Expand the folded group to see the log file of test_autocast 2023-03-31T05:59:30.2288044Z ##[group]PRINTING LOG FILE of test_autocast (/var/lib/jenkins/workspace/test/test-reports/test_autocast_s9ortff0.log) 2023-03-31T05:59:30.2289061Z Test results will be stored in test-reports/python-pytest/test_autocast/test_autocast-cad3792541a15294.xml 2023-03-31T05:59:30.2289590Z ============================= test session starts ============================== 2023-03-31T05:59:30.2290208Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:30.2290631Z cachedir: .pytest_cache 2023-03-31T05:59:30.2291607Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:30.2292236Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:30.2292970Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:30.2293505Z collecting ... collected 9 items 2023-03-31T05:59:30.2295079Z Running 9 items in this shard: test/test_autocast.py::TestAutocastCPU::test_autocast_methods_expect_builtin_promote, test/test_autocast.py::TestAutocastCPU::test_autocast_nn_bf16, test/test_autocast.py::TestAutocastCPU::test_autocast_nn_fp32, test/test_autocast.py::TestAutocastCPU::test_autocast_torch_bf16, test/test_autocast.py::TestAutocastCPU::test_autocast_torch_expect_builtin_promote, test/test_autocast.py::TestAutocastCPU::test_autocast_torch_fp32, test/test_autocast.py::TestAutocastCPU::test_autocast_torch_need_autocast_promote, test/test_autocast.py::TestAutocastGPU::test_cast_cache_is_global, test/test_autocast.py::TestTorchAutocast::test_autocast_fast_dtype 2023-03-31T05:59:30.2296319Z 2023-03-31T05:59:30.2296618Z test_autocast.py::TestAutocastCPU::test_autocast_methods_expect_builtin_promote PASSED [ 11%] 2023-03-31T05:59:30.2297216Z test_autocast.py::TestAutocastCPU::test_autocast_nn_bf16 PASSED [ 22%] 2023-03-31T05:59:30.2297780Z test_autocast.py::TestAutocastCPU::test_autocast_nn_fp32 PASSED [ 33%] 2023-03-31T05:59:30.2298476Z test_autocast.py::TestAutocastCPU::test_autocast_torch_bf16 PASSED [ 44%] 2023-03-31T05:59:30.2299124Z test_autocast.py::TestAutocastCPU::test_autocast_torch_expect_builtin_promote PASSED [ 55%] 2023-03-31T05:59:30.2299734Z test_autocast.py::TestAutocastCPU::test_autocast_torch_fp32 PASSED [ 66%] 2023-03-31T05:59:30.2300286Z test_autocast.py::TestAutocastCPU::test_autocast_torch_need_autocast_promote PASSED [ 77%] 2023-03-31T05:59:30.2300857Z test_autocast.py::TestAutocastGPU::test_cast_cache_is_global SKIPPED (requires cuda) [ 88%] 2023-03-31T05:59:30.2301417Z test_autocast.py::TestTorchAutocast::test_autocast_fast_dtype PASSED [100%] 2023-03-31T05:59:30.2301731Z 2023-03-31T05:59:30.2302434Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_autocast/test_autocast-cad3792541a15294.xml - 2023-03-31T05:59:30.2303010Z ========================= 8 passed, 1 skipped in 0.43s ========================= 2023-03-31T05:59:30.2303520Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:30.2303852Z 2023-03-31T05:59:30.2304244Z ##[endgroup] 2023-03-31T05:59:30.2304909Z FINISHED PRINTING LOG FILE of test_autocast (/var/lib/jenkins/workspace/test/test-reports/test_autocast_s9ortff0.log) 2023-03-31T05:59:30.2305277Z 2023-03-31T05:59:32.3173614Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:32.3201497Z Ignoring disabled issues: [] 2023-03-31T05:59:32.3378479Z Running dynamo/test_nops ... [2023-03-31 05:59:32.337495] 2023-03-31T05:59:32.3381131Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_nops.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:32.337832] 2023-03-31T05:59:34.7479859Z 2023-03-31T05:59:34.7480962Z Expand the folded group to see the log file of dynamo/test_nops 2023-03-31T05:59:34.7482286Z ##[group]PRINTING LOG FILE of dynamo/test_nops (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_nops_jevid1nj.log) 2023-03-31T05:59:34.7483559Z Test results will be stored in test-reports/python-pytest/dynamo.test_nops/dynamo.test_nops-e432b0905306743b.xml 2023-03-31T05:59:34.7483918Z ============================= test session starts ============================== 2023-03-31T05:59:34.7484315Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:34.7484855Z cachedir: .pytest_cache 2023-03-31T05:59:34.7485298Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:34.7485683Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:34.7486133Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:34.7486448Z collecting ... collected 4 items 2023-03-31T05:59:34.7486879Z Running 4 items in this shard: test/dynamo/test_nops.py::NopTests::test1, test/dynamo/test_nops.py::NopTests::test2, test/dynamo/test_nops.py::NopTests::test3, test/dynamo/test_nops.py::NopTests::test_extended_args 2023-03-31T05:59:34.7487194Z 2023-03-31T05:59:34.7487334Z dynamo/test_nops.py::NopTests::test1 PASSED [ 25%] 2023-03-31T05:59:34.7487616Z dynamo/test_nops.py::NopTests::test2 PASSED [ 50%] 2023-03-31T05:59:34.7487905Z dynamo/test_nops.py::NopTests::test3 PASSED [ 75%] 2023-03-31T05:59:34.7488209Z dynamo/test_nops.py::NopTests::test_extended_args PASSED [100%] 2023-03-31T05:59:34.7488383Z 2023-03-31T05:59:34.7488732Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_nops/dynamo.test_nops-e432b0905306743b.xml - 2023-03-31T05:59:34.7489093Z ============================== 4 passed in 0.38s =============================== 2023-03-31T05:59:34.7489480Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:34.7489741Z 2023-03-31T05:59:34.7489997Z ##[endgroup] 2023-03-31T05:59:34.7490404Z FINISHED PRINTING LOG FILE of dynamo/test_nops (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_nops_jevid1nj.log) 2023-03-31T05:59:34.7490638Z 2023-03-31T05:59:36.8317057Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:36.8344685Z Ignoring disabled issues: [] 2023-03-31T05:59:36.8522796Z Running test_functional_optim ... [2023-03-31 05:59:36.851803] 2023-03-31T05:59:36.8524039Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_functional_optim.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:36.852119] 2023-03-31T05:59:39.4319923Z 2023-03-31T05:59:39.4320456Z Expand the folded group to see the log file of test_functional_optim 2023-03-31T05:59:39.4321596Z ##[group]PRINTING LOG FILE of test_functional_optim (/var/lib/jenkins/workspace/test/test-reports/test_functional_optim_snix2bz0.log) 2023-03-31T05:59:39.4322771Z Test results will be stored in test-reports/python-pytest/test_functional_optim/test_functional_optim-d28d870b6e4a8ab2.xml 2023-03-31T05:59:39.4323726Z ============================= test session starts ============================== 2023-03-31T05:59:39.4324424Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:39.4324892Z cachedir: .pytest_cache 2023-03-31T05:59:39.4325911Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:39.4326520Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:39.4327281Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:39.4327791Z collecting ... collected 4 items 2023-03-31T05:59:39.4328939Z Running 4 items in this shard: test/test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_adam, test/test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_adam_w, test/test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_sgd, test/test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_registration 2023-03-31T05:59:39.4329917Z 2023-03-31T05:59:39.4330277Z test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_adam PASSED [ 25%] 2023-03-31T05:59:39.4331161Z test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_adam_w PASSED [ 50%] 2023-03-31T05:59:39.4331885Z test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_parity_sgd PASSED [ 75%] 2023-03-31T05:59:39.4332579Z test_functional_optim.py::TestFunctionalOptimParity::test_functional_optim_registration PASSED [100%] 2023-03-31T05:59:39.4332975Z 2023-03-31T05:59:39.4333666Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functional_optim/test_functional_optim-d28d870b6e4a8ab2.xml - 2023-03-31T05:59:39.4334283Z ============================== 4 passed in 0.41s =============================== 2023-03-31T05:59:39.4334804Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:39.4335142Z 2023-03-31T05:59:39.4335552Z ##[endgroup] 2023-03-31T05:59:39.4336290Z FINISHED PRINTING LOG FILE of test_functional_optim (/var/lib/jenkins/workspace/test/test-reports/test_functional_optim_snix2bz0.log) 2023-03-31T05:59:39.4336692Z 2023-03-31T05:59:41.5167774Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:41.5194433Z Ignoring disabled issues: [] 2023-03-31T05:59:41.5370991Z Running nn/test_dropout ... [2023-03-31 05:59:41.536660] 2023-03-31T05:59:41.5372592Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_dropout.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:41.536979] 2023-03-31T05:59:44.0871268Z 2023-03-31T05:59:44.0872050Z Expand the folded group to see the log file of nn/test_dropout 2023-03-31T05:59:44.0872938Z ##[group]PRINTING LOG FILE of nn/test_dropout (/var/lib/jenkins/workspace/test/test-reports/nn-test_dropout_kfv633d5.log) 2023-03-31T05:59:44.0873609Z Test results will be stored in test-reports/python-pytest/nn.test_dropout/nn.test_dropout-5c381b2a31fd36ea.xml 2023-03-31T05:59:44.0873964Z ============================= test session starts ============================== 2023-03-31T05:59:44.0874377Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:44.0874663Z cachedir: .pytest_cache 2023-03-31T05:59:44.0875090Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:44.0875465Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:44.0875923Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:44.0876241Z collecting ... collected 9 items 2023-03-31T05:59:44.0877436Z Running 9 items in this shard: test/nn/test_dropout.py::TestDropoutNN::test_AlphaDropout, test/nn/test_dropout.py::TestDropoutNN::test_FeatureAlphaDropout, test/nn/test_dropout.py::TestDropoutNN::test_invalid_dropout_p, test/nn/test_dropout.py::TestDropoutNN::test_native_dropout_corner_case, test/nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout1d_cpu, test/nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout2d_cpu, test/nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout3d_cpu, test/nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout_cpu, test/nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_empty_dropout_cpu 2023-03-31T05:59:44.0878294Z 2023-03-31T05:59:44.0878454Z nn/test_dropout.py::TestDropoutNN::test_AlphaDropout PASSED [ 11%] 2023-03-31T05:59:44.0878789Z nn/test_dropout.py::TestDropoutNN::test_FeatureAlphaDropout PASSED [ 22%] 2023-03-31T05:59:44.0879133Z nn/test_dropout.py::TestDropoutNN::test_invalid_dropout_p PASSED [ 33%] 2023-03-31T05:59:44.0879501Z nn/test_dropout.py::TestDropoutNN::test_native_dropout_corner_case SKIPPED (CUDA unavailable) [ 44%] 2023-03-31T05:59:44.0879881Z nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout1d_cpu PASSED [ 55%] 2023-03-31T05:59:44.0880333Z nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout2d_cpu PASSED [ 66%] 2023-03-31T05:59:44.0880698Z nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout3d_cpu PASSED [ 77%] 2023-03-31T05:59:44.0881063Z nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_Dropout_cpu PASSED [ 88%] 2023-03-31T05:59:44.0881427Z nn/test_dropout.py::TestDropoutNNDeviceTypeCPU::test_empty_dropout_cpu PASSED [100%] 2023-03-31T05:59:44.0881635Z 2023-03-31T05:59:44.0881991Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-5c381b2a31fd36ea.xml - 2023-03-31T05:59:44.0882377Z ========================= 8 passed, 1 skipped in 0.38s ========================= 2023-03-31T05:59:44.0882700Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:44.0882910Z 2023-03-31T05:59:44.0883358Z ##[endgroup] 2023-03-31T05:59:44.0883779Z FINISHED PRINTING LOG FILE of nn/test_dropout (/var/lib/jenkins/workspace/test/test-reports/nn-test_dropout_kfv633d5.log) 2023-03-31T05:59:44.0884015Z 2023-03-31T05:59:46.1629090Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:46.1656761Z Ignoring disabled issues: [] 2023-03-31T05:59:46.1841106Z Running lazy/test_debug_util ... [2023-03-31 05:59:46.183575] 2023-03-31T05:59:46.1842417Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'lazy/test_debug_util.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:46.183909] 2023-03-31T05:59:48.5560612Z 2023-03-31T05:59:48.5561195Z Expand the folded group to see the log file of lazy/test_debug_util 2023-03-31T05:59:48.5562348Z ##[group]PRINTING LOG FILE of lazy/test_debug_util (/var/lib/jenkins/workspace/test/test-reports/lazy-test_debug_util_6iq8s8x3.log) 2023-03-31T05:59:48.5563604Z Test results will be stored in test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-f22e01707f730048.xml 2023-03-31T05:59:48.5563983Z ============================= test session starts ============================== 2023-03-31T05:59:48.5564399Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:48.5564668Z cachedir: .pytest_cache 2023-03-31T05:59:48.5565104Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:48.5565483Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:48.5565927Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:48.5566246Z collecting ... collected 1 item 2023-03-31T05:59:48.5566556Z Running 1 items in this shard: test/lazy/test_debug_util.py::DebugUtilTest::test_get_python_frames 2023-03-31T05:59:48.5566761Z 2023-03-31T05:59:48.5566920Z lazy/test_debug_util.py::DebugUtilTest::test_get_python_frames PASSED [100%] 2023-03-31T05:59:48.5567097Z 2023-03-31T05:59:48.5567673Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-f22e01707f730048.xml - 2023-03-31T05:59:48.5568051Z ============================== 1 passed in 0.38s =============================== 2023-03-31T05:59:48.5568374Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:48.5568579Z 2023-03-31T05:59:48.5568819Z ##[endgroup] 2023-03-31T05:59:48.5569250Z FINISHED PRINTING LOG FILE of lazy/test_debug_util (/var/lib/jenkins/workspace/test/test-reports/lazy-test_debug_util_6iq8s8x3.log) 2023-03-31T05:59:48.5569493Z 2023-03-31T05:59:50.6647916Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:50.6674916Z Ignoring disabled issues: [] 2023-03-31T05:59:50.6852730Z Running nn/test_parametrization ... [2023-03-31 05:59:50.684926] 2023-03-31T05:59:50.6854666Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_parametrization.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:50.685222] 2023-03-31T05:59:53.2621015Z 2023-03-31T05:59:53.2621608Z Expand the folded group to see the log file of nn/test_parametrization 2023-03-31T05:59:53.2622799Z ##[group]PRINTING LOG FILE of nn/test_parametrization (/var/lib/jenkins/workspace/test/test-reports/nn-test_parametrization_y4dtsa16.log) 2023-03-31T05:59:53.2623961Z Test results will be stored in test-reports/python-pytest/nn.test_parametrization/nn.test_parametrization-07c4b0561c36f470.xml 2023-03-31T05:59:53.2624348Z ============================= test session starts ============================== 2023-03-31T05:59:53.2624759Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:53.2625037Z cachedir: .pytest_cache 2023-03-31T05:59:53.2625463Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:53.2625848Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:53.2626298Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:53.2626599Z collecting ... collected 23 items 2023-03-31T05:59:53.2629809Z Running 23 items in this shard: test/nn/test_parametrization.py::TestNNParametrization::test_caching_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_caching_parametrization_with_transfer_parametrizations_and_params, test/nn/test_parametrization.py::TestNNParametrization::test_deepcopy_after_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_errors_parametrized_tensor_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_errors_unparametrized_tensor_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_initialization_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_multiple_inputs_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm, test/nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_dim, test/nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_forward, test/nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_load_state_dict, test/nn/test_parametrization.py::TestNNParametrization::test_orthogonal_errors, test/nn/test_parametrization.py::TestNNParametrization::test_orthogonal_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_parametrization_same_training_mode, test/nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_buffer_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_nested_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_serialization_parametrization, test/nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params, test/nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_many_to_one, test/nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_right_inverse, test/nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_single_param, test/nn/test_parametrization.py::TestNNParametrization::test_type_before_parametrizations 2023-03-31T05:59:53.2632692Z 2023-03-31T05:59:53.2632896Z nn/test_parametrization.py::TestNNParametrization::test_caching_parametrization PASSED [ 4%] 2023-03-31T05:59:53.2633354Z nn/test_parametrization.py::TestNNParametrization::test_caching_parametrization_with_transfer_parametrizations_and_params PASSED [ 8%] 2023-03-31T05:59:53.2633879Z nn/test_parametrization.py::TestNNParametrization::test_deepcopy_after_parametrization PASSED [ 13%] 2023-03-31T05:59:53.2634309Z nn/test_parametrization.py::TestNNParametrization::test_errors_parametrized_tensor_parametrization PASSED [ 17%] 2023-03-31T05:59:53.2634754Z nn/test_parametrization.py::TestNNParametrization::test_errors_unparametrized_tensor_parametrization PASSED [ 21%] 2023-03-31T05:59:53.2635187Z nn/test_parametrization.py::TestNNParametrization::test_initialization_parametrization PASSED [ 26%] 2023-03-31T05:59:53.2635598Z nn/test_parametrization.py::TestNNParametrization::test_multiple_inputs_parametrization PASSED [ 30%] 2023-03-31T05:59:53.2635999Z nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm PASSED [ 34%] 2023-03-31T05:59:53.2636381Z nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_dim PASSED [ 39%] 2023-03-31T05:59:53.2636770Z nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_forward PASSED [ 43%] 2023-03-31T05:59:53.2637162Z nn/test_parametrization.py::TestNNParametrization::test_new_spectral_norm_load_state_dict PASSED [ 47%] 2023-03-31T05:59:53.2637560Z nn/test_parametrization.py::TestNNParametrization::test_orthogonal_errors PASSED [ 52%] 2023-03-31T05:59:53.2637956Z nn/test_parametrization.py::TestNNParametrization::test_orthogonal_parametrization PASSED [ 56%] 2023-03-31T05:59:53.2638368Z nn/test_parametrization.py::TestNNParametrization::test_parametrization_same_training_mode PASSED [ 60%] 2023-03-31T05:59:53.2638824Z nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_buffer_parametrization PASSED [ 65%] 2023-03-31T05:59:53.2639289Z nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_nested_parametrization PASSED [ 69%] 2023-03-31T05:59:53.2639719Z nn/test_parametrization.py::TestNNParametrization::test_register_and_remove_parametrization PASSED [ 73%] 2023-03-31T05:59:53.2640122Z nn/test_parametrization.py::TestNNParametrization::test_serialization_parametrization PASSED [ 78%] 2023-03-31T05:59:53.2640551Z nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params PASSED [ 82%] 2023-03-31T05:59:53.2640993Z nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_many_to_one PASSED [ 86%] 2023-03-31T05:59:53.2641445Z nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_right_inverse PASSED [ 91%] 2023-03-31T05:59:53.2641892Z nn/test_parametrization.py::TestNNParametrization::test_transfer_parametrizations_and_params_single_param PASSED [ 95%] 2023-03-31T05:59:53.2642325Z nn/test_parametrization.py::TestNNParametrization::test_type_before_parametrizations PASSED [100%] 2023-03-31T05:59:53.2644575Z 2023-03-31T05:59:53.2645406Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_parametrization/nn.test_parametrization-07c4b0561c36f470.xml - 2023-03-31T05:59:53.2645944Z ============================== 23 passed in 0.40s ============================== 2023-03-31T05:59:53.2646387Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:53.2646633Z 2023-03-31T05:59:53.2647046Z ##[endgroup] 2023-03-31T05:59:53.2647581Z FINISHED PRINTING LOG FILE of nn/test_parametrization (/var/lib/jenkins/workspace/test/test-reports/nn-test_parametrization_y4dtsa16.log) 2023-03-31T05:59:53.2647838Z 2023-03-31T05:59:55.3521637Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:55.3551806Z Ignoring disabled issues: [] 2023-03-31T05:59:55.3729646Z Running dynamo/test_export_mutations ... [2023-03-31 05:59:55.372577] 2023-03-31T05:59:55.3731386Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'dynamo/test_export_mutations.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:55.372915] 2023-03-31T05:59:57.7846297Z 2023-03-31T05:59:57.7847111Z Expand the folded group to see the log file of dynamo/test_export_mutations 2023-03-31T05:59:57.7848312Z ##[group]PRINTING LOG FILE of dynamo/test_export_mutations (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_export_mutations_y61il8qf.log) 2023-03-31T05:59:57.7849142Z Test results will be stored in test-reports/python-pytest/dynamo.test_export_mutations/dynamo.test_export_mutations-a1308598ca2531c7.xml 2023-03-31T05:59:57.7849616Z ============================= test session starts ============================== 2023-03-31T05:59:57.7850216Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T05:59:57.7850617Z cachedir: .pytest_cache 2023-03-31T05:59:57.7851353Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T05:59:57.7851874Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T05:59:57.7852327Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T05:59:57.7852651Z collecting ... collected 8 items 2023-03-31T05:59:57.7853975Z Running 8 items in this shard: test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_1, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_2, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_3, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_4, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_1, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_2, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_3, test/dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_4 2023-03-31T05:59:57.7855069Z 2023-03-31T05:59:57.7855926Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_1 SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/88468 for platform(s) linux, macos, mac. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 12%] 2023-03-31T05:59:57.7857265Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_2 SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/88475 for platform(s) linux, mac, macos. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2023-03-31T05:59:57.7858657Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_3 SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/88466 for platform(s) linux, mac, macos. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 37%] 2023-03-31T05:59:57.7859975Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_negative_4 SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/88467 for platform(s) linux, macos. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 50%] 2023-03-31T05:59:57.7860708Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_1 PASSED [ 62%] 2023-03-31T05:59:57.7861186Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_2 PASSED [ 75%] 2023-03-31T05:59:57.7861626Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_3 PASSED [ 87%] 2023-03-31T05:59:57.7862058Z dynamo/test_export_mutations.py::MutationExportTests::test_module_attribute_mutation_violation_positive_4 PASSED [100%] 2023-03-31T05:59:57.7862389Z 2023-03-31T05:59:57.7862786Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_export_mutations/dynamo.test_export_mutations-a1308598ca2531c7.xml - 2023-03-31T05:59:57.7863197Z ========================= 4 passed, 4 skipped in 0.36s ========================= 2023-03-31T05:59:57.7863523Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T05:59:57.7863728Z 2023-03-31T05:59:57.7863986Z ##[endgroup] 2023-03-31T05:59:57.7864452Z FINISHED PRINTING LOG FILE of dynamo/test_export_mutations (/var/lib/jenkins/workspace/test/test-reports/dynamo-test_export_mutations_y61il8qf.log) 2023-03-31T05:59:57.7864712Z 2023-03-31T05:59:59.8645509Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T05:59:59.8674551Z Ignoring disabled issues: [] 2023-03-31T05:59:59.8854131Z Running test_pytree ... [2023-03-31 05:59:59.884957] 2023-03-31T05:59:59.8855503Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_pytree.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 05:59:59.885263] 2023-03-31T06:00:02.2497914Z 2023-03-31T06:00:02.2498496Z Expand the folded group to see the log file of test_pytree 2023-03-31T06:00:02.2499476Z ##[group]PRINTING LOG FILE of test_pytree (/var/lib/jenkins/workspace/test/test-reports/test_pytree_av7o56p3.log) 2023-03-31T06:00:02.2500135Z Test results will be stored in test-reports/python-pytest/test_pytree/test_pytree-7ecb7e01843e0b86.xml 2023-03-31T06:00:02.2500494Z ============================= test session starts ============================== 2023-03-31T06:00:02.2500973Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:02.2501255Z cachedir: .pytest_cache 2023-03-31T06:00:02.2501693Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:02.2502055Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:02.2502512Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:02.2502831Z collecting ... collected 15 items 2023-03-31T06:00:02.2504226Z Running 15 items in this shard: test/test_pytree.py::TestPytree::test_broadcast_to_and_flatten, test/test_pytree.py::TestPytree::test_flatten_unflatten_dict, test/test_pytree.py::TestPytree::test_flatten_unflatten_leaf, test/test_pytree.py::TestPytree::test_flatten_unflatten_list, test/test_pytree.py::TestPytree::test_flatten_unflatten_namedtuple, test/test_pytree.py::TestPytree::test_flatten_unflatten_nested, test/test_pytree.py::TestPytree::test_flatten_unflatten_odict, test/test_pytree.py::TestPytree::test_flatten_unflatten_return_type_max, test/test_pytree.py::TestPytree::test_flatten_unflatten_return_type_min, test/test_pytree.py::TestPytree::test_flatten_unflatten_tuple, test/test_pytree.py::TestPytree::test_tree_all_any, test/test_pytree.py::TestPytree::test_tree_only, test/test_pytree.py::TestPytree::test_treemap, test/test_pytree.py::TestPytree::test_treespec_equality, test/test_pytree.py::TestPytree::test_treespec_repr 2023-03-31T06:00:02.2505257Z 2023-03-31T06:00:02.2505415Z test_pytree.py::TestPytree::test_broadcast_to_and_flatten PASSED [ 6%] 2023-03-31T06:00:02.2505749Z test_pytree.py::TestPytree::test_flatten_unflatten_dict PASSED [ 13%] 2023-03-31T06:00:02.2506060Z test_pytree.py::TestPytree::test_flatten_unflatten_leaf PASSED [ 20%] 2023-03-31T06:00:02.2506430Z test_pytree.py::TestPytree::test_flatten_unflatten_list PASSED [ 26%] 2023-03-31T06:00:02.2506757Z test_pytree.py::TestPytree::test_flatten_unflatten_namedtuple PASSED [ 33%] 2023-03-31T06:00:02.2507070Z test_pytree.py::TestPytree::test_flatten_unflatten_nested PASSED [ 40%] 2023-03-31T06:00:02.2507387Z test_pytree.py::TestPytree::test_flatten_unflatten_odict PASSED [ 46%] 2023-03-31T06:00:02.2507715Z test_pytree.py::TestPytree::test_flatten_unflatten_return_type_max PASSED [ 53%] 2023-03-31T06:00:02.2508054Z test_pytree.py::TestPytree::test_flatten_unflatten_return_type_min PASSED [ 60%] 2023-03-31T06:00:02.2508371Z test_pytree.py::TestPytree::test_flatten_unflatten_tuple PASSED [ 66%] 2023-03-31T06:00:02.2508678Z test_pytree.py::TestPytree::test_tree_all_any PASSED [ 73%] 2023-03-31T06:00:02.2508978Z test_pytree.py::TestPytree::test_tree_only PASSED [ 80%] 2023-03-31T06:00:02.2509263Z test_pytree.py::TestPytree::test_treemap PASSED [ 86%] 2023-03-31T06:00:02.2509566Z test_pytree.py::TestPytree::test_treespec_equality PASSED [ 93%] 2023-03-31T06:00:02.2509871Z test_pytree.py::TestPytree::test_treespec_repr PASSED [100%] 2023-03-31T06:00:02.2510041Z 2023-03-31T06:00:02.2510378Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_pytree/test_pytree-7ecb7e01843e0b86.xml - 2023-03-31T06:00:02.2510785Z ============================== 15 passed in 0.38s ============================== 2023-03-31T06:00:02.2511155Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:02.2511358Z 2023-03-31T06:00:02.2511606Z ##[endgroup] 2023-03-31T06:00:02.2511995Z FINISHED PRINTING LOG FILE of test_pytree (/var/lib/jenkins/workspace/test/test-reports/test_pytree_av7o56p3.log) 2023-03-31T06:00:02.2512216Z 2023-03-31T06:00:04.3447994Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:04.3475492Z Ignoring disabled issues: [] 2023-03-31T06:00:04.3654259Z Running test_per_overload_api ... [2023-03-31 06:00:04.365119] 2023-03-31T06:00:04.3657049Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_per_overload_api.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:04.365434] 2023-03-31T06:00:06.6164259Z 2023-03-31T06:00:06.6164854Z Expand the folded group to see the log file of test_per_overload_api 2023-03-31T06:00:06.6166113Z ##[group]PRINTING LOG FILE of test_per_overload_api (/var/lib/jenkins/workspace/test/test-reports/test_per_overload_api_x9r574mj.log) 2023-03-31T06:00:06.6167249Z Test results will be stored in test-reports/python-pytest/test_per_overload_api/test_per_overload_api-b4332ef1533186c9.xml 2023-03-31T06:00:06.6167601Z ============================= test session starts ============================== 2023-03-31T06:00:06.6168018Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:06.6168495Z cachedir: .pytest_cache 2023-03-31T06:00:06.6168939Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:06.6169317Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:06.6169756Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:06.6170075Z collecting ... collected 3 items 2023-03-31T06:00:06.6170583Z Running 3 items in this shard: test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverload, test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverloadpacket, test/test_per_overload_api.py::TestPerOverloadAPI::test_decompose 2023-03-31T06:00:06.6170960Z 2023-03-31T06:00:06.6171134Z test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverload PASSED [ 33%] 2023-03-31T06:00:06.6171556Z test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverloadpacket PASSED [ 66%] 2023-03-31T06:00:06.6171922Z test_per_overload_api.py::TestPerOverloadAPI::test_decompose PASSED [100%] 2023-03-31T06:00:06.6172112Z 2023-03-31T06:00:06.6172471Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-b4332ef1533186c9.xml - 2023-03-31T06:00:06.6172848Z ============================== 3 passed in 0.34s =============================== 2023-03-31T06:00:06.6173154Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:06.6173359Z 2023-03-31T06:00:06.6173600Z ##[endgroup] 2023-03-31T06:00:06.6174030Z FINISHED PRINTING LOG FILE of test_per_overload_api (/var/lib/jenkins/workspace/test/test-reports/test_per_overload_api_x9r574mj.log) 2023-03-31T06:00:06.6174270Z 2023-03-31T06:00:08.6977159Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:08.7005590Z Ignoring disabled issues: [] 2023-03-31T06:00:08.7186146Z Running test_native_mha ... [2023-03-31 06:00:08.718156] 2023-03-31T06:00:08.7187164Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_native_mha.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:08.718456] 2023-03-31T06:00:11.1635485Z 2023-03-31T06:00:11.1636251Z Expand the folded group to see the log file of test_native_mha 2023-03-31T06:00:11.1637351Z ##[group]PRINTING LOG FILE of test_native_mha (/var/lib/jenkins/workspace/test/test-reports/test_native_mha_ptu4ehi9.log) 2023-03-31T06:00:11.1638129Z Test results will be stored in test-reports/python-pytest/test_native_mha/test_native_mha-d938d6850829e9a5.xml 2023-03-31T06:00:11.1638540Z ============================= test session starts ============================== 2023-03-31T06:00:11.1638956Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:11.1639362Z cachedir: .pytest_cache 2023-03-31T06:00:11.1640052Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:11.1640443Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:11.1640976Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:11.1641308Z collecting ... collected 28 items 2023-03-31T06:00:11.1648382Z Running 28 items in this shard: test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_attention_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_encoder_decoder_attention_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_False_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_True_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_transform_bias_rescale_qkv_cpu_float32, test/test_native_mha.py::TestMHADeviceTypeCPU::test_transform_bias_rescale_qkv_nested_cpu_float32 2023-03-31T06:00:11.1654849Z 2023-03-31T06:00:11.1655045Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_attention_cpu_float32 PASSED [ 3%] 2023-03-31T06:00:11.1655525Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_encoder_decoder_attention_cpu_float32 PASSED [ 7%] 2023-03-31T06:00:11.1656073Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 10%] 2023-03-31T06:00:11.1656800Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 14%] 2023-03-31T06:00:11.1657488Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 17%] 2023-03-31T06:00:11.1658153Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 21%] 2023-03-31T06:00:11.1658819Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 25%] 2023-03-31T06:00:11.1659505Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 28%] 2023-03-31T06:00:11.1660171Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 32%] 2023-03-31T06:00:11.1660852Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 35%] 2023-03-31T06:00:11.1661684Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 39%] 2023-03-31T06:00:11.1662298Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 42%] 2023-03-31T06:00:11.1662973Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 46%] 2023-03-31T06:00:11.1663650Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_False_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 50%] 2023-03-31T06:00:11.1666793Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 53%] 2023-03-31T06:00:11.1667742Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 57%] 2023-03-31T06:00:11.1668924Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 60%] 2023-03-31T06:00:11.1669619Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_False_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 64%] 2023-03-31T06:00:11.1670294Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 67%] 2023-03-31T06:00:11.1670928Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 71%] 2023-03-31T06:00:11.1671599Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 75%] 2023-03-31T06:00:11.1672247Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_False_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 78%] 2023-03-31T06:00:11.1672852Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_False_cpu_float32 PASSED [ 82%] 2023-03-31T06:00:11.1673526Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_False_fused_True_cpu_float32 PASSED [ 85%] 2023-03-31T06:00:11.1674120Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_False_cpu_float32 PASSED [ 89%] 2023-03-31T06:00:11.1674712Z test_native_mha.py::TestMHADeviceTypeCPU::test_native_multihead_self_attention_use_nt_True_use_padding_True_pad_all_True_need_weights_False_average_attn_weights_True_fused_True_cpu_float32 PASSED [ 92%] 2023-03-31T06:00:11.1675181Z test_native_mha.py::TestMHADeviceTypeCPU::test_transform_bias_rescale_qkv_cpu_float32 PASSED [ 96%] 2023-03-31T06:00:11.1675705Z test_native_mha.py::TestMHADeviceTypeCPU::test_transform_bias_rescale_qkv_nested_cpu_float32 SKIPPED (Only runs on cuda) [100%] 2023-03-31T06:00:11.1675981Z 2023-03-31T06:00:11.1676390Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_native_mha/test_native_mha-d938d6850829e9a5.xml - 2023-03-31T06:00:11.1676768Z ======================== 27 passed, 1 skipped in 0.27s ========================= 2023-03-31T06:00:11.1677087Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:11.1677296Z 2023-03-31T06:00:11.1677694Z ##[endgroup] 2023-03-31T06:00:11.1678108Z FINISHED PRINTING LOG FILE of test_native_mha (/var/lib/jenkins/workspace/test/test-reports/test_native_mha_ptu4ehi9.log) 2023-03-31T06:00:11.1678340Z 2023-03-31T06:00:13.2580402Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:13.2608625Z Ignoring disabled issues: [] 2023-03-31T06:00:13.2789054Z Running test_dlpack ... [2023-03-31 06:00:13.278571] 2023-03-31T06:00:13.2791724Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_dlpack.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:13.278880] 2023-03-31T06:00:15.6766051Z 2023-03-31T06:00:15.6766586Z Expand the folded group to see the log file of test_dlpack 2023-03-31T06:00:15.6767683Z ##[group]PRINTING LOG FILE of test_dlpack (/var/lib/jenkins/workspace/test/test-reports/test_dlpack_sa53punj.log) 2023-03-31T06:00:15.6769207Z Test results will be stored in test-reports/python-pytest/test_dlpack/test_dlpack-8aff06badb092ae9.xml 2023-03-31T06:00:15.6769817Z ============================= test session starts ============================== 2023-03-31T06:00:15.6770501Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:15.6771012Z cachedir: .pytest_cache 2023-03-31T06:00:15.6771555Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:15.6771933Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:15.6772372Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:15.6772682Z collecting ... collected 95 items 2023-03-31T06:00:15.6782446Z Running 95 items in this shard: test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_default_stream_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_error_on_bool_tensor_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_is_conj_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_non_strided_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_requires_grad_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_normalize_strides_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_shared_storage_cpu, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_uint8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_bfloat16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_complex128, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_complex64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int16, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int32, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int64, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int8, test/test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_uint8 2023-03-31T06:00:15.6791410Z 2023-03-31T06:00:15.6791606Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:00:15.6792042Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_complex128 PASSED [ 2%] 2023-03-31T06:00:15.6792430Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_complex64 PASSED [ 3%] 2023-03-31T06:00:15.6792814Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float16 PASSED [ 4%] 2023-03-31T06:00:15.6793192Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float32 PASSED [ 5%] 2023-03-31T06:00:15.6793568Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_float64 PASSED [ 6%] 2023-03-31T06:00:15.6793937Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int16 PASSED [ 7%] 2023-03-31T06:00:15.6794311Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int32 PASSED [ 8%] 2023-03-31T06:00:15.6794679Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int64 PASSED [ 9%] 2023-03-31T06:00:15.6795039Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_int8 PASSED [ 10%] 2023-03-31T06:00:15.6795413Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_capsule_conversion_cpu_uint8 PASSED [ 11%] 2023-03-31T06:00:15.6798064Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T06:00:15.6798885Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_complex128 SKIPPED (Only runs on cuda) [ 13%] 2023-03-31T06:00:15.6799784Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_complex64 SKIPPED (Only runs on cuda) [ 14%] 2023-03-31T06:00:15.6800513Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float16 SKIPPED (Only runs on cuda) [ 15%] 2023-03-31T06:00:15.6800954Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float32 SKIPPED (Only runs on cuda) [ 16%] 2023-03-31T06:00:15.6801366Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_float64 SKIPPED (Only runs on cuda) [ 17%] 2023-03-31T06:00:15.6801812Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int16 SKIPPED (Only runs on cuda) [ 18%] 2023-03-31T06:00:15.6802376Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int32 SKIPPED (Only runs on cuda) [ 20%] 2023-03-31T06:00:15.6803295Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int64 SKIPPED (Only runs on cuda) [ 21%] 2023-03-31T06:00:15.6803727Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_int8 SKIPPED (Only runs on cuda) [ 22%] 2023-03-31T06:00:15.6804160Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_diff_streams_cpu_uint8 SKIPPED (Only runs on cuda) [ 23%] 2023-03-31T06:00:15.6804615Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 24%] 2023-03-31T06:00:15.6805377Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_complex128 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T06:00:15.6806300Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_complex64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:00:15.6807049Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:00:15.6807474Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float32 SKIPPED (Only runs on cuda) [ 28%] 2023-03-31T06:00:15.6808161Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_float64 SKIPPED (Only runs on cuda) [ 29%] 2023-03-31T06:00:15.6808894Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int16 SKIPPED (Only runs on cuda) [ 30%] 2023-03-31T06:00:15.6809417Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int32 SKIPPED (Only runs on cuda) [ 31%] 2023-03-31T06:00:15.6809831Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int64 SKIPPED (Only runs on cuda) [ 32%] 2023-03-31T06:00:15.6810316Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_int8 SKIPPED (Only runs on cuda) [ 33%] 2023-03-31T06:00:15.6810724Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_conversion_with_streams_cpu_uint8 SKIPPED (Only runs on cuda) [ 34%] 2023-03-31T06:00:15.6811135Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_default_stream_cpu SKIPPED (Only runs on cuda) [ 35%] 2023-03-31T06:00:15.6811519Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_error_on_bool_tensor_cpu PASSED [ 36%] 2023-03-31T06:00:15.6811998Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_is_conj_cpu PASSED [ 37%] 2023-03-31T06:00:15.6812630Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_non_strided_cpu PASSED [ 38%] 2023-03-31T06:00:15.6813292Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_export_requires_grad_cpu PASSED [ 40%] 2023-03-31T06:00:15.6813740Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_normalize_strides_cpu PASSED [ 41%] 2023-03-31T06:00:15.6814107Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:00:15.6814497Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_complex128 PASSED [ 43%] 2023-03-31T06:00:15.6814886Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_complex64 PASSED [ 44%] 2023-03-31T06:00:15.6815269Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float16 PASSED [ 45%] 2023-03-31T06:00:15.6815898Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float32 PASSED [ 46%] 2023-03-31T06:00:15.6816467Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_float64 PASSED [ 47%] 2023-03-31T06:00:15.6816847Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int16 PASSED [ 48%] 2023-03-31T06:00:15.6817224Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int32 PASSED [ 49%] 2023-03-31T06:00:15.6817586Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int64 PASSED [ 50%] 2023-03-31T06:00:15.6817962Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_int8 PASSED [ 51%] 2023-03-31T06:00:15.6818335Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_protocol_conversion_cpu_uint8 PASSED [ 52%] 2023-03-31T06:00:15.6818743Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_shared_storage_cpu PASSED [ 53%] 2023-03-31T06:00:15.6819122Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:00:15.6819512Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_complex128 PASSED [ 55%] 2023-03-31T06:00:15.6819902Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_complex64 PASSED [ 56%] 2023-03-31T06:00:15.6820272Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float16 PASSED [ 57%] 2023-03-31T06:00:15.6820647Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float32 PASSED [ 58%] 2023-03-31T06:00:15.6821067Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_float64 PASSED [ 60%] 2023-03-31T06:00:15.6821431Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int16 PASSED [ 61%] 2023-03-31T06:00:15.6821896Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int32 PASSED [ 62%] 2023-03-31T06:00:15.6822274Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int64 PASSED [ 63%] 2023-03-31T06:00:15.6822652Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_int8 PASSED [ 64%] 2023-03-31T06:00:15.6823016Z test_dlpack.py::TestTorchDlPackCPU::test_dlpack_tensor_invalid_stream_cpu_uint8 PASSED [ 65%] 2023-03-31T06:00:15.6823379Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:00:15.6823734Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_complex128 PASSED [ 67%] 2023-03-31T06:00:15.6824121Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_complex64 PASSED [ 68%] 2023-03-31T06:00:15.6824455Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float16 PASSED [ 69%] 2023-03-31T06:00:15.6824810Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float32 PASSED [ 70%] 2023-03-31T06:00:15.6825140Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_float64 PASSED [ 71%] 2023-03-31T06:00:15.6825460Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int16 PASSED [ 72%] 2023-03-31T06:00:15.6825796Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int32 PASSED [ 73%] 2023-03-31T06:00:15.6826129Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int64 PASSED [ 74%] 2023-03-31T06:00:15.6826463Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_int8 PASSED [ 75%] 2023-03-31T06:00:15.6826782Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_cpu_uint8 PASSED [ 76%] 2023-03-31T06:00:15.6827132Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_bfloat16 PASSED [ 77%] 2023-03-31T06:00:15.6827502Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_complex128 PASSED [ 78%] 2023-03-31T06:00:15.6827855Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_complex64 PASSED [ 80%] 2023-03-31T06:00:15.6828220Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float16 PASSED [ 81%] 2023-03-31T06:00:15.6828582Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float32 PASSED [ 82%] 2023-03-31T06:00:15.6828972Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_float64 PASSED [ 83%] 2023-03-31T06:00:15.6829344Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int16 PASSED [ 84%] 2023-03-31T06:00:15.6829690Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int32 PASSED [ 85%] 2023-03-31T06:00:15.6830040Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int64 PASSED [ 86%] 2023-03-31T06:00:15.6830389Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_int8 PASSED [ 87%] 2023-03-31T06:00:15.6830729Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_dtype_cpu_uint8 PASSED [ 88%] 2023-03-31T06:00:15.6831090Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:00:15.6884752Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_complex128 PASSED [ 90%] 2023-03-31T06:00:15.6885349Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_complex64 PASSED [ 91%] 2023-03-31T06:00:15.6886107Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float16 PASSED [ 92%] 2023-03-31T06:00:15.6886822Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float32 PASSED [ 93%] 2023-03-31T06:00:15.6887481Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_float64 PASSED [ 94%] 2023-03-31T06:00:15.6887921Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int16 PASSED [ 95%] 2023-03-31T06:00:15.6888304Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int32 PASSED [ 96%] 2023-03-31T06:00:15.6888841Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int64 PASSED [ 97%] 2023-03-31T06:00:15.6889205Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_int8 PASSED [ 98%] 2023-03-31T06:00:15.6889585Z test_dlpack.py::TestTorchDlPackCPU::test_from_dlpack_noncontinguous_cpu_uint8 PASSED [100%] 2023-03-31T06:00:15.6889792Z 2023-03-31T06:00:15.6890253Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_dlpack/test_dlpack-8aff06badb092ae9.xml - 2023-03-31T06:00:15.6890626Z ======================== 72 passed, 23 skipped in 0.24s ======================== 2023-03-31T06:00:15.6890941Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:15.6891151Z 2023-03-31T06:00:15.6891617Z ##[endgroup] 2023-03-31T06:00:15.6892024Z FINISHED PRINTING LOG FILE of test_dlpack (/var/lib/jenkins/workspace/test/test-reports/test_dlpack_sa53punj.log) 2023-03-31T06:00:15.6892251Z 2023-03-31T06:00:17.7544133Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:17.7571686Z Ignoring disabled issues: [] 2023-03-31T06:00:17.7749441Z Running test_numpy_interop ... [2023-03-31 06:00:17.774625] 2023-03-31T06:00:17.7751718Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_numpy_interop.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:17.774935] 2023-03-31T06:00:20.1606677Z 2023-03-31T06:00:20.1607348Z Expand the folded group to see the log file of test_numpy_interop 2023-03-31T06:00:20.1608450Z ##[group]PRINTING LOG FILE of test_numpy_interop (/var/lib/jenkins/workspace/test/test-reports/test_numpy_interop_2ipxhs2g.log) 2023-03-31T06:00:20.1609175Z Test results will be stored in test-reports/python-pytest/test_numpy_interop/test_numpy_interop-2dbd5715b37db2e3.xml 2023-03-31T06:00:20.1609599Z ============================= test session starts ============================== 2023-03-31T06:00:20.1610126Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:20.1610588Z cachedir: .pytest_cache 2023-03-31T06:00:20.1611288Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:20.1612189Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:20.1612877Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:20.1613178Z collecting ... collected 26 items 2023-03-31T06:00:20.1616083Z Running 26 items in this shard: test/test_numpy_interop.py::TestNumPyInteropCPU::test_ctor_with_invalid_numpy_array_sequence_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_ctor_with_numpy_scalar_ctor_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_from_list_of_ndarray_warning_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_from_numpy_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_has_storage_numpy_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_multiplication_numpy_scalar_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_array_interface_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_index_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_non_writeable_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_bfloat16, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_bool, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_complex128, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_complex64, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float16, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float32, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float64, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int16, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int32, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int64, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int8, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_uint8, test/test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_unresizable_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_parse_numpy_int_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_bool_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_cpu, test/test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_force_argument_cpu 2023-03-31T06:00:20.1618848Z 2023-03-31T06:00:20.1619122Z test_numpy_interop.py::TestNumPyInteropCPU::test_ctor_with_invalid_numpy_array_sequence_cpu PASSED [ 3%] 2023-03-31T06:00:20.1619553Z test_numpy_interop.py::TestNumPyInteropCPU::test_ctor_with_numpy_scalar_ctor_cpu PASSED [ 7%] 2023-03-31T06:00:20.1619980Z test_numpy_interop.py::TestNumPyInteropCPU::test_from_list_of_ndarray_warning_cpu PASSED [ 11%] 2023-03-31T06:00:20.1620388Z test_numpy_interop.py::TestNumPyInteropCPU::test_from_numpy_cpu PASSED [ 15%] 2023-03-31T06:00:20.1620764Z test_numpy_interop.py::TestNumPyInteropCPU::test_has_storage_numpy_cpu PASSED [ 19%] 2023-03-31T06:00:20.1621220Z test_numpy_interop.py::TestNumPyInteropCPU::test_multiplication_numpy_scalar_cpu PASSED [ 23%] 2023-03-31T06:00:20.1621603Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_array_interface_cpu PASSED [ 26%] 2023-03-31T06:00:20.1622014Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_index_cpu PASSED [ 30%] 2023-03-31T06:00:20.1622463Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_non_writeable_cpu PASSED [ 34%] 2023-03-31T06:00:20.1622905Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:00:20.1623271Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_bool PASSED [ 42%] 2023-03-31T06:00:20.1623731Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_complex128 PASSED [ 46%] 2023-03-31T06:00:20.1624111Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_complex64 PASSED [ 50%] 2023-03-31T06:00:20.1624605Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float16 PASSED [ 53%] 2023-03-31T06:00:20.1625064Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float32 PASSED [ 57%] 2023-03-31T06:00:20.1625437Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_float64 PASSED [ 61%] 2023-03-31T06:00:20.1625873Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int16 PASSED [ 65%] 2023-03-31T06:00:20.1626229Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int32 PASSED [ 69%] 2023-03-31T06:00:20.1626662Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int64 PASSED [ 73%] 2023-03-31T06:00:20.1627025Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_int8 PASSED [ 76%] 2023-03-31T06:00:20.1627455Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_scalar_cmp_cpu_uint8 PASSED [ 80%] 2023-03-31T06:00:20.1627805Z test_numpy_interop.py::TestNumPyInteropCPU::test_numpy_unresizable_cpu PASSED [ 84%] 2023-03-31T06:00:20.1628230Z test_numpy_interop.py::TestNumPyInteropCPU::test_parse_numpy_int_cpu PASSED [ 88%] 2023-03-31T06:00:20.1628587Z test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_bool_cpu PASSED [ 92%] 2023-03-31T06:00:20.1628982Z test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_cpu PASSED [ 96%] 2023-03-31T06:00:20.1629346Z test_numpy_interop.py::TestNumPyInteropCPU::test_to_numpy_force_argument_cpu PASSED [100%] 2023-03-31T06:00:20.1629590Z 2023-03-31T06:00:20.1629983Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_numpy_interop/test_numpy_interop-2dbd5715b37db2e3.xml - 2023-03-31T06:00:20.1630465Z ============================== 26 passed in 0.20s ============================== 2023-03-31T06:00:20.1630776Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:20.1630979Z 2023-03-31T06:00:20.1631324Z ##[endgroup] 2023-03-31T06:00:20.1631759Z FINISHED PRINTING LOG FILE of test_numpy_interop (/var/lib/jenkins/workspace/test/test-reports/test_numpy_interop_2ipxhs2g.log) 2023-03-31T06:00:20.1632075Z 2023-03-31T06:00:22.2414413Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:22.2441247Z Ignoring disabled issues: [] 2023-03-31T06:00:22.2619606Z Running distributions/test_constraints ... [2023-03-31 06:00:22.261654] 2023-03-31T06:00:22.2622203Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'distributions/test_constraints.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:22.261950] 2023-03-31T06:00:24.6664261Z 2023-03-31T06:00:24.6678865Z Expand the folded group to see the log file of distributions/test_constraints 2023-03-31T06:00:24.6680162Z ##[group]PRINTING LOG FILE of distributions/test_constraints (/var/lib/jenkins/workspace/test/test-reports/distributions-test_constraints_acbye19v.log) 2023-03-31T06:00:24.6681479Z Test results will be stored in test-reports/python-pytest/distributions.test_constraints/distributions.test_constraints-c5c846935c57168a.xml 2023-03-31T06:00:24.6682184Z ============================= test session starts ============================== 2023-03-31T06:00:24.6682988Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:24.6683691Z cachedir: .pytest_cache 2023-03-31T06:00:24.6684465Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:24.6685091Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:24.6685904Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:24.6686461Z collecting ... collected 136 items 2023-03-31T06:00:24.6708188Z Running 136 items in this shard: test/distributions/test_constraints.py::test_constraint[False-constraint_fn0-False-value0], test/distributions/test_constraints.py::test_constraint[False-constraint_fn1-False-value1], test/distributions/test_constraints.py::test_constraint[False-constraint_fn2-False-value2], test/distributions/test_constraints.py::test_constraint[False-constraint_fn3-True-value3], test/distributions/test_constraints.py::test_constraint[False-constraint_fn4-False-value4], test/distributions/test_constraints.py::test_constraint[False-constraint_fn5-False-value5], test/distributions/test_constraints.py::test_constraint[False-constraint_fn6-True-value6], test/distributions/test_constraints.py::test_constraint[False-constraint_fn7-True-value7], test/distributions/test_constraints.py::test_constraint[False-constraint_fn8-False-value8], test/distributions/test_constraints.py::test_constraint[False-constraint_fn9-True-value9], test/distributions/test_constraints.py::test_constraint[False-constraint_fn10-False-value10], test/distributions/test_constraints.py::test_constraint[False-constraint_fn11-False-value11], test/distributions/test_constraints.py::test_constraint[False-constraint_fn12-True-value12], test/distributions/test_constraints.py::test_constraint[False-constraint_fn13-True-value13], test/distributions/test_constraints.py::test_constraint[False-constraint_fn14-False-value14], test/distributions/test_constraints.py::test_constraint[False-constraint_fn15-True-value15], test/distributions/test_constraints.py::test_constraint[False-constraint_fn16-True-value16], test/distributions/test_constraints.py::test_constraint[False-constraint_fn17-True-value17], test/distributions/test_constraints.py::test_constraint[True-constraint_fn0-False-value0], test/distributions/test_constraints.py::test_constraint[True-constraint_fn1-False-value1], test/distributions/test_constraints.py::test_constraint[True-constraint_fn2-False-value2], test/distributions/test_constraints.py::test_constraint[True-constraint_fn3-True-value3], test/distributions/test_constraints.py::test_constraint[True-constraint_fn4-False-value4], test/distributions/test_constraints.py::test_constraint[True-constraint_fn5-False-value5], test/distributions/test_constraints.py::test_constraint[True-constraint_fn6-True-value6], test/distributions/test_constraints.py::test_constraint[True-constraint_fn7-True-value7], test/distributions/test_constraints.py::test_constraint[True-constraint_fn8-False-value8], test/distributions/test_constraints.py::test_constraint[True-constraint_fn9-True-value9], test/distributions/test_constraints.py::test_constraint[True-constraint_fn10-False-value10], test/distributions/test_constraints.py::test_constraint[True-constraint_fn11-False-value11], test/distributions/test_constraints.py::test_constraint[True-constraint_fn12-True-value12], test/distributions/test_constraints.py::test_constraint[True-constraint_fn13-True-value13], test/distributions/test_constraints.py::test_constraint[True-constraint_fn14-False-value14], test/distributions/test_constraints.py::test_constraint[True-constraint_fn15-True-value15], test/distributions/test_constraints.py::test_constraint[True-constraint_fn16-True-value16], test/distributions/test_constraints.py::test_constraint[True-constraint_fn17-True-value17], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn0-args0], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn1-args1], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn2-args2], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args3], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args4], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args5], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args6], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args7], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args8], test/distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args9], test/distributions/test_constraints.py::test_biject_to[False-_LessThan-args10], test/distributions/test_constraints.py::test_biject_to[False-_LessThan-args11], test/distributions/test_constraints.py::test_biject_to[False-_LessThan-args12], test/distributions/test_constraints.py::test_biject_to[False-_LessThan-args13], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn14-args14], test/distributions/test_constraints.py::test_biject_to[False-_Interval-args15], test/distributions/test_constraints.py::test_biject_to[False-_Interval-args16], test/distributions/test_constraints.py::test_biject_to[False-_Interval-args17], test/distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args18], test/distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args19], test/distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args20], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn21-args21], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn22-args22], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn23-args23], test/distributions/test_constraints.py::test_biject_to[False-constraint_fn24-args24], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn0-args0], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn1-args1], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn2-args2], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args3], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args4], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args5], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args6], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args7], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args8], test/distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args9], test/distributions/test_constraints.py::test_biject_to[True-_LessThan-args10], test/distributions/test_constraints.py::test_biject_to[True-_LessThan-args11], test/distributions/test_constraints.py::test_biject_to[True-_LessThan-args12], test/distributions/test_constraints.py::test_biject_to[True-_LessThan-args13], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn14-args14], test/distributions/test_constraints.py::test_biject_to[True-_Interval-args15], test/distributions/test_constraints.py::test_biject_to[True-_Interval-args16], test/distributions/test_constraints.py::test_biject_to[True-_Interval-args17], test/distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args18], test/distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args19], test/distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args20], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn21-args21], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn22-args22], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn23-args23], test/distributions/test_constraints.py::test_biject_to[True-constraint_fn24-args24], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn0-args0], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn1-args1], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn2-args2], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args3], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args4], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args5], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args6], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args7], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args8], test/distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args9], test/distributions/test_constraints.py::test_transform_to[False-_LessThan-args10], test/distributions/test_constraints.py::test_transform_to[False-_LessThan-args11], test/distributions/test_constraints.py::test_transform_to[False-_LessThan-args12], test/distributions/test_constraints.py::test_transform_to[False-_LessThan-args13], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn14-args14], test/distributions/test_constraints.py::test_transform_to[False-_Interval-args15], test/distributions/test_constraints.py::test_transform_to[False-_Interval-args16], test/distributions/test_constraints.py::test_transform_to[False-_Interval-args17], test/distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args18], test/distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args19], test/distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args20], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn21-args21], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn22-args22], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn23-args23], test/distributions/test_constraints.py::test_transform_to[False-constraint_fn24-args24], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn0-args0], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn1-args1], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn2-args2], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args3], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args4], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args5], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args6], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args7], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args8], test/distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args9], test/distributions/test_constraints.py::test_transform_to[True-_LessThan-args10], test/distributions/test_constraints.py::test_transform_to[True-_LessThan-args11], test/distributions/test_constraints.py::test_transform_to[True-_LessThan-args12], test/distributions/test_constraints.py::test_transform_to[True-_LessThan-args13], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn14-args14], test/distributions/test_constraints.py::test_transform_to[True-_Interval-args15], test/distributions/test_constraints.py::test_transform_to[True-_Interval-args16], test/distributions/test_constraints.py::test_transform_to[True-_Interval-args17], test/distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args18], test/distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args19], test/distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args20], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn21-args21], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn22-args22], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn23-args23], test/distributions/test_constraints.py::test_transform_to[True-constraint_fn24-args24] 2023-03-31T06:00:24.6721964Z 2023-03-31T06:00:24.6722246Z distributions/test_constraints.py::test_constraint[False-constraint_fn0-False-value0] PASSED [ 0%] 2023-03-31T06:00:24.6722747Z distributions/test_constraints.py::test_constraint[False-constraint_fn1-False-value1] PASSED [ 1%] 2023-03-31T06:00:24.6723503Z distributions/test_constraints.py::test_constraint[False-constraint_fn2-False-value2] PASSED [ 2%] 2023-03-31T06:00:24.6724229Z distributions/test_constraints.py::test_constraint[False-constraint_fn3-True-value3] PASSED [ 2%] 2023-03-31T06:00:24.6724949Z distributions/test_constraints.py::test_constraint[False-constraint_fn4-False-value4] PASSED [ 3%] 2023-03-31T06:00:24.6725879Z distributions/test_constraints.py::test_constraint[False-constraint_fn5-False-value5] PASSED [ 4%] 2023-03-31T06:00:24.6726362Z distributions/test_constraints.py::test_constraint[False-constraint_fn6-True-value6] PASSED [ 5%] 2023-03-31T06:00:24.6726839Z distributions/test_constraints.py::test_constraint[False-constraint_fn7-True-value7] PASSED [ 5%] 2023-03-31T06:00:24.6727308Z distributions/test_constraints.py::test_constraint[False-constraint_fn8-False-value8] PASSED [ 6%] 2023-03-31T06:00:24.6727792Z distributions/test_constraints.py::test_constraint[False-constraint_fn9-True-value9] PASSED [ 7%] 2023-03-31T06:00:24.6728273Z distributions/test_constraints.py::test_constraint[False-constraint_fn10-False-value10] PASSED [ 8%] 2023-03-31T06:00:24.6728757Z distributions/test_constraints.py::test_constraint[False-constraint_fn11-False-value11] PASSED [ 8%] 2023-03-31T06:00:24.6729223Z distributions/test_constraints.py::test_constraint[False-constraint_fn12-True-value12] PASSED [ 9%] 2023-03-31T06:00:24.6729703Z distributions/test_constraints.py::test_constraint[False-constraint_fn13-True-value13] PASSED [ 10%] 2023-03-31T06:00:24.6730183Z distributions/test_constraints.py::test_constraint[False-constraint_fn14-False-value14] PASSED [ 11%] 2023-03-31T06:00:24.6730648Z distributions/test_constraints.py::test_constraint[False-constraint_fn15-True-value15] PASSED [ 11%] 2023-03-31T06:00:24.6731124Z distributions/test_constraints.py::test_constraint[False-constraint_fn16-True-value16] PASSED [ 12%] 2023-03-31T06:00:24.6731658Z distributions/test_constraints.py::test_constraint[False-constraint_fn17-True-value17] PASSED [ 13%] 2023-03-31T06:00:24.6732165Z distributions/test_constraints.py::test_constraint[True-constraint_fn0-False-value0] SKIPPED (CUDA not found.) [ 13%] 2023-03-31T06:00:24.6732676Z distributions/test_constraints.py::test_constraint[True-constraint_fn1-False-value1] SKIPPED (CUDA not found.) [ 14%] 2023-03-31T06:00:24.6733200Z distributions/test_constraints.py::test_constraint[True-constraint_fn2-False-value2] SKIPPED (CUDA not found.) [ 15%] 2023-03-31T06:00:24.6733717Z distributions/test_constraints.py::test_constraint[True-constraint_fn3-True-value3] SKIPPED (CUDA not found.) [ 16%] 2023-03-31T06:00:24.6734240Z distributions/test_constraints.py::test_constraint[True-constraint_fn4-False-value4] SKIPPED (CUDA not found.) [ 16%] 2023-03-31T06:00:24.6734741Z distributions/test_constraints.py::test_constraint[True-constraint_fn5-False-value5] SKIPPED (CUDA not found.) [ 17%] 2023-03-31T06:00:24.6735304Z distributions/test_constraints.py::test_constraint[True-constraint_fn6-True-value6] SKIPPED (CUDA not found.) [ 18%] 2023-03-31T06:00:24.6735817Z distributions/test_constraints.py::test_constraint[True-constraint_fn7-True-value7] SKIPPED (CUDA not found.) [ 19%] 2023-03-31T06:00:24.6736331Z distributions/test_constraints.py::test_constraint[True-constraint_fn8-False-value8] SKIPPED (CUDA not found.) [ 19%] 2023-03-31T06:00:24.6736835Z distributions/test_constraints.py::test_constraint[True-constraint_fn9-True-value9] SKIPPED (CUDA not found.) [ 20%] 2023-03-31T06:00:24.6737358Z distributions/test_constraints.py::test_constraint[True-constraint_fn10-False-value10] SKIPPED (CUDA not found.) [ 21%] 2023-03-31T06:00:24.6737881Z distributions/test_constraints.py::test_constraint[True-constraint_fn11-False-value11] SKIPPED (CUDA not found.) [ 22%] 2023-03-31T06:00:24.6738385Z distributions/test_constraints.py::test_constraint[True-constraint_fn12-True-value12] SKIPPED (CUDA not found.) [ 22%] 2023-03-31T06:00:24.6738908Z distributions/test_constraints.py::test_constraint[True-constraint_fn13-True-value13] SKIPPED (CUDA not found.) [ 23%] 2023-03-31T06:00:24.6739423Z distributions/test_constraints.py::test_constraint[True-constraint_fn14-False-value14] SKIPPED (CUDA not found.) [ 24%] 2023-03-31T06:00:24.6739941Z distributions/test_constraints.py::test_constraint[True-constraint_fn15-True-value15] SKIPPED (CUDA not found.) [ 25%] 2023-03-31T06:00:24.6740477Z distributions/test_constraints.py::test_constraint[True-constraint_fn16-True-value16] SKIPPED (CUDA not found.) [ 25%] 2023-03-31T06:00:24.6741023Z distributions/test_constraints.py::test_constraint[True-constraint_fn17-True-value17] SKIPPED (CUDA not found.) [ 26%] 2023-03-31T06:00:24.6741494Z distributions/test_constraints.py::test_biject_to[False-constraint_fn0-args0] PASSED [ 27%] 2023-03-31T06:00:24.6741947Z distributions/test_constraints.py::test_biject_to[False-constraint_fn1-args1] PASSED [ 27%] 2023-03-31T06:00:24.6742482Z distributions/test_constraints.py::test_biject_to[False-constraint_fn2-args2] PASSED [ 28%] 2023-03-31T06:00:24.6742941Z distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args3] PASSED [ 29%] 2023-03-31T06:00:24.6743386Z distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args4] PASSED [ 30%] 2023-03-31T06:00:24.6743835Z distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args5] PASSED [ 30%] 2023-03-31T06:00:24.6744263Z distributions/test_constraints.py::test_biject_to[False-_GreaterThan-args6] PASSED [ 31%] 2023-03-31T06:00:24.6744710Z distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args7] PASSED [ 32%] 2023-03-31T06:00:24.6745164Z distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args8] PASSED [ 33%] 2023-03-31T06:00:24.6745595Z distributions/test_constraints.py::test_biject_to[False-_GreaterThanEq-args9] PASSED [ 33%] 2023-03-31T06:00:24.6746035Z distributions/test_constraints.py::test_biject_to[False-_LessThan-args10] PASSED [ 34%] 2023-03-31T06:00:24.6746511Z distributions/test_constraints.py::test_biject_to[False-_LessThan-args11] PASSED [ 35%] 2023-03-31T06:00:24.6746942Z distributions/test_constraints.py::test_biject_to[False-_LessThan-args12] PASSED [ 36%] 2023-03-31T06:00:24.6747356Z distributions/test_constraints.py::test_biject_to[False-_LessThan-args13] PASSED [ 36%] 2023-03-31T06:00:24.6747796Z distributions/test_constraints.py::test_biject_to[False-constraint_fn14-args14] PASSED [ 37%] 2023-03-31T06:00:24.6748233Z distributions/test_constraints.py::test_biject_to[False-_Interval-args15] PASSED [ 38%] 2023-03-31T06:00:24.6748656Z distributions/test_constraints.py::test_biject_to[False-_Interval-args16] PASSED [ 38%] 2023-03-31T06:00:24.6749086Z distributions/test_constraints.py::test_biject_to[False-_Interval-args17] PASSED [ 39%] 2023-03-31T06:00:24.6749542Z distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args18] PASSED [ 40%] 2023-03-31T06:00:24.6750049Z distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args19] PASSED [ 41%] 2023-03-31T06:00:24.6750504Z distributions/test_constraints.py::test_biject_to[False-_HalfOpenInterval-args20] PASSED [ 41%] 2023-03-31T06:00:24.6750962Z distributions/test_constraints.py::test_biject_to[False-constraint_fn21-args21] PASSED [ 42%] 2023-03-31T06:00:24.6751412Z distributions/test_constraints.py::test_biject_to[False-constraint_fn22-args22] PASSED [ 43%] 2023-03-31T06:00:24.6751922Z distributions/test_constraints.py::test_biject_to[False-constraint_fn23-args23] SKIPPED (`biject_to` not implemented.) [ 44%] 2023-03-31T06:00:24.6752451Z distributions/test_constraints.py::test_biject_to[False-constraint_fn24-args24] SKIPPED (`biject_to` not implemented.) [ 44%] 2023-03-31T06:00:24.6752959Z distributions/test_constraints.py::test_biject_to[True-constraint_fn0-args0] SKIPPED (CUDA not found.) [ 45%] 2023-03-31T06:00:24.6753455Z distributions/test_constraints.py::test_biject_to[True-constraint_fn1-args1] SKIPPED (CUDA not found.) [ 46%] 2023-03-31T06:00:24.6753950Z distributions/test_constraints.py::test_biject_to[True-constraint_fn2-args2] SKIPPED (CUDA not found.) [ 47%] 2023-03-31T06:00:24.6754423Z distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args3] SKIPPED (CUDA not found.) [ 47%] 2023-03-31T06:00:24.6754916Z distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args4] SKIPPED (CUDA not found.) [ 48%] 2023-03-31T06:00:24.6755435Z distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args5] SKIPPED (CUDA not found.) [ 49%] 2023-03-31T06:00:24.6755956Z distributions/test_constraints.py::test_biject_to[True-_GreaterThan-args6] SKIPPED (CUDA not found.) [ 50%] 2023-03-31T06:00:24.6756447Z distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args7] SKIPPED (CUDA not found.) [ 50%] 2023-03-31T06:00:24.6756942Z distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args8] SKIPPED (CUDA not found.) [ 51%] 2023-03-31T06:00:24.6757441Z distributions/test_constraints.py::test_biject_to[True-_GreaterThanEq-args9] SKIPPED (CUDA not found.) [ 52%] 2023-03-31T06:00:24.6757915Z distributions/test_constraints.py::test_biject_to[True-_LessThan-args10] SKIPPED (CUDA not found.) [ 52%] 2023-03-31T06:00:24.6758394Z distributions/test_constraints.py::test_biject_to[True-_LessThan-args11] SKIPPED (CUDA not found.) [ 53%] 2023-03-31T06:00:24.6758867Z distributions/test_constraints.py::test_biject_to[True-_LessThan-args12] SKIPPED (CUDA not found.) [ 54%] 2023-03-31T06:00:24.6759343Z distributions/test_constraints.py::test_biject_to[True-_LessThan-args13] SKIPPED (CUDA not found.) [ 55%] 2023-03-31T06:00:24.6759821Z distributions/test_constraints.py::test_biject_to[True-constraint_fn14-args14] SKIPPED (CUDA not found.) [ 55%] 2023-03-31T06:00:24.6760308Z distributions/test_constraints.py::test_biject_to[True-_Interval-args15] SKIPPED (CUDA not found.) [ 56%] 2023-03-31T06:00:24.6760786Z distributions/test_constraints.py::test_biject_to[True-_Interval-args16] SKIPPED (CUDA not found.) [ 57%] 2023-03-31T06:00:24.6761295Z distributions/test_constraints.py::test_biject_to[True-_Interval-args17] SKIPPED (CUDA not found.) [ 58%] 2023-03-31T06:00:24.6761783Z distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args18] SKIPPED (CUDA not found.) [ 58%] 2023-03-31T06:00:24.6762286Z distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args19] SKIPPED (CUDA not found.) [ 59%] 2023-03-31T06:00:24.6762793Z distributions/test_constraints.py::test_biject_to[True-_HalfOpenInterval-args20] SKIPPED (CUDA not found.) [ 60%] 2023-03-31T06:00:24.6763469Z distributions/test_constraints.py::test_biject_to[True-constraint_fn21-args21] SKIPPED (CUDA not found.) [ 61%] 2023-03-31T06:00:24.6763961Z distributions/test_constraints.py::test_biject_to[True-constraint_fn22-args22] SKIPPED (CUDA not found.) [ 61%] 2023-03-31T06:00:24.6764440Z distributions/test_constraints.py::test_biject_to[True-constraint_fn23-args23] SKIPPED (CUDA not found.) [ 62%] 2023-03-31T06:00:24.6764995Z distributions/test_constraints.py::test_biject_to[True-constraint_fn24-args24] SKIPPED (CUDA not found.) [ 63%] 2023-03-31T06:00:24.6765455Z distributions/test_constraints.py::test_transform_to[False-constraint_fn0-args0] PASSED [ 63%] 2023-03-31T06:00:24.6765911Z distributions/test_constraints.py::test_transform_to[False-constraint_fn1-args1] PASSED [ 64%] 2023-03-31T06:00:24.6766371Z distributions/test_constraints.py::test_transform_to[False-constraint_fn2-args2] PASSED [ 65%] 2023-03-31T06:00:24.6766826Z distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args3] PASSED [ 66%] 2023-03-31T06:00:24.6767272Z distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args4] PASSED [ 66%] 2023-03-31T06:00:24.6767721Z distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args5] PASSED [ 67%] 2023-03-31T06:00:24.6768168Z distributions/test_constraints.py::test_transform_to[False-_GreaterThan-args6] PASSED [ 68%] 2023-03-31T06:00:24.6768604Z distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args7] PASSED [ 69%] 2023-03-31T06:00:24.6769057Z distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args8] PASSED [ 69%] 2023-03-31T06:00:24.6769504Z distributions/test_constraints.py::test_transform_to[False-_GreaterThanEq-args9] PASSED [ 70%] 2023-03-31T06:00:24.6770035Z distributions/test_constraints.py::test_transform_to[False-_LessThan-args10] PASSED [ 71%] 2023-03-31T06:00:24.6770679Z distributions/test_constraints.py::test_transform_to[False-_LessThan-args11] PASSED [ 72%] 2023-03-31T06:00:24.6780984Z distributions/test_constraints.py::test_transform_to[False-_LessThan-args12] PASSED [ 72%] 2023-03-31T06:00:24.6781465Z distributions/test_constraints.py::test_transform_to[False-_LessThan-args13] PASSED [ 73%] 2023-03-31T06:00:24.6781927Z distributions/test_constraints.py::test_transform_to[False-constraint_fn14-args14] PASSED [ 74%] 2023-03-31T06:00:24.6782496Z distributions/test_constraints.py::test_transform_to[False-_Interval-args15] PASSED [ 75%] 2023-03-31T06:00:24.6782939Z distributions/test_constraints.py::test_transform_to[False-_Interval-args16] PASSED [ 75%] 2023-03-31T06:00:24.6783375Z distributions/test_constraints.py::test_transform_to[False-_Interval-args17] PASSED [ 76%] 2023-03-31T06:00:24.6783833Z distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args18] PASSED [ 77%] 2023-03-31T06:00:24.6784313Z distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args19] PASSED [ 77%] 2023-03-31T06:00:24.6784785Z distributions/test_constraints.py::test_transform_to[False-_HalfOpenInterval-args20] PASSED [ 78%] 2023-03-31T06:00:24.6785258Z distributions/test_constraints.py::test_transform_to[False-constraint_fn21-args21] PASSED [ 79%] 2023-03-31T06:00:24.6785720Z distributions/test_constraints.py::test_transform_to[False-constraint_fn22-args22] PASSED [ 80%] 2023-03-31T06:00:24.6786166Z distributions/test_constraints.py::test_transform_to[False-constraint_fn23-args23] PASSED [ 80%] 2023-03-31T06:00:24.6786760Z distributions/test_constraints.py::test_transform_to[False-constraint_fn24-args24] PASSED [ 81%] 2023-03-31T06:00:24.6787250Z distributions/test_constraints.py::test_transform_to[True-constraint_fn0-args0] SKIPPED (CUDA not found.) [ 82%] 2023-03-31T06:00:24.6787755Z distributions/test_constraints.py::test_transform_to[True-constraint_fn1-args1] SKIPPED (CUDA not found.) [ 83%] 2023-03-31T06:00:24.6788249Z distributions/test_constraints.py::test_transform_to[True-constraint_fn2-args2] SKIPPED (CUDA not found.) [ 83%] 2023-03-31T06:00:24.6788748Z distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args3] SKIPPED (CUDA not found.) [ 84%] 2023-03-31T06:00:24.6789248Z distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args4] SKIPPED (CUDA not found.) [ 85%] 2023-03-31T06:00:24.6789742Z distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args5] SKIPPED (CUDA not found.) [ 86%] 2023-03-31T06:00:24.6790283Z distributions/test_constraints.py::test_transform_to[True-_GreaterThan-args6] SKIPPED (CUDA not found.) [ 86%] 2023-03-31T06:00:24.6790779Z distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args7] SKIPPED (CUDA not found.) [ 87%] 2023-03-31T06:00:24.6791283Z distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args8] SKIPPED (CUDA not found.) [ 88%] 2023-03-31T06:00:24.6791792Z distributions/test_constraints.py::test_transform_to[True-_GreaterThanEq-args9] SKIPPED (CUDA not found.) [ 88%] 2023-03-31T06:00:24.6792273Z distributions/test_constraints.py::test_transform_to[True-_LessThan-args10] SKIPPED (CUDA not found.) [ 89%] 2023-03-31T06:00:24.6792758Z distributions/test_constraints.py::test_transform_to[True-_LessThan-args11] SKIPPED (CUDA not found.) [ 90%] 2023-03-31T06:00:24.6793239Z distributions/test_constraints.py::test_transform_to[True-_LessThan-args12] SKIPPED (CUDA not found.) [ 91%] 2023-03-31T06:00:24.6793708Z distributions/test_constraints.py::test_transform_to[True-_LessThan-args13] SKIPPED (CUDA not found.) [ 91%] 2023-03-31T06:00:24.6794199Z distributions/test_constraints.py::test_transform_to[True-constraint_fn14-args14] SKIPPED (CUDA not found.) [ 92%] 2023-03-31T06:00:24.6794689Z distributions/test_constraints.py::test_transform_to[True-_Interval-args15] SKIPPED (CUDA not found.) [ 93%] 2023-03-31T06:00:24.6795170Z distributions/test_constraints.py::test_transform_to[True-_Interval-args16] SKIPPED (CUDA not found.) [ 94%] 2023-03-31T06:00:24.6795672Z distributions/test_constraints.py::test_transform_to[True-_Interval-args17] SKIPPED (CUDA not found.) [ 94%] 2023-03-31T06:00:24.6796203Z distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args18] SKIPPED (CUDA not found.) [ 95%] 2023-03-31T06:00:24.6796722Z distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args19] SKIPPED (CUDA not found.) [ 96%] 2023-03-31T06:00:24.6797249Z distributions/test_constraints.py::test_transform_to[True-_HalfOpenInterval-args20] SKIPPED (CUDA not found.) [ 97%] 2023-03-31T06:00:24.6797748Z distributions/test_constraints.py::test_transform_to[True-constraint_fn21-args21] SKIPPED (CUDA not found.) [ 97%] 2023-03-31T06:00:24.6798241Z distributions/test_constraints.py::test_transform_to[True-constraint_fn22-args22] SKIPPED (CUDA not found.) [ 98%] 2023-03-31T06:00:24.6798738Z distributions/test_constraints.py::test_transform_to[True-constraint_fn23-args23] SKIPPED (CUDA not found.) [ 99%] 2023-03-31T06:00:24.6799238Z distributions/test_constraints.py::test_transform_to[True-constraint_fn24-args24] SKIPPED (CUDA not found.) [100%] 2023-03-31T06:00:24.6799456Z 2023-03-31T06:00:24.6799848Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/distributions.test_constraints/distributions.test_constraints-c5c846935c57168a.xml - 2023-03-31T06:00:24.6800266Z ======================== 66 passed, 70 skipped in 0.26s ======================== 2023-03-31T06:00:24.6800590Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:24.6800791Z 2023-03-31T06:00:24.6801126Z ##[endgroup] 2023-03-31T06:00:24.6801601Z FINISHED PRINTING LOG FILE of distributions/test_constraints (/var/lib/jenkins/workspace/test/test-reports/distributions-test_constraints_acbye19v.log) 2023-03-31T06:00:24.6801878Z 2023-03-31T06:00:26.7440476Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:26.7470482Z Ignoring disabled issues: [] 2023-03-31T06:00:26.7648582Z Running nn/test_pruning ... [2023-03-31 06:00:26.764564] 2023-03-31T06:00:26.7651554Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_pruning.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:00:26.764860] 2023-03-31T06:00:29.1339612Z 2023-03-31T06:00:29.1340177Z Expand the folded group to see the log file of nn/test_pruning 2023-03-31T06:00:29.1341428Z ##[group]PRINTING LOG FILE of nn/test_pruning (/var/lib/jenkins/workspace/test/test-reports/nn-test_pruning_gjlazrm2.log) 2023-03-31T06:00:29.1342180Z Test results will be stored in test-reports/python-pytest/nn.test_pruning/nn.test_pruning-d14395771e0d5c57.xml 2023-03-31T06:00:29.1342675Z ============================= test session starts ============================== 2023-03-31T06:00:29.1343089Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:00:29.1343414Z cachedir: .pytest_cache 2023-03-31T06:00:29.1343877Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:00:29.1344305Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:00:29.1344814Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:00:29.1345129Z collecting ... collected 34 items 2023-03-31T06:00:29.1348350Z Running 34 items in this shard: test/nn/test_pruning.py::TestPruningNN::test_compute_nparams_to_prune, test/nn/test_pruning.py::TestPruningNN::test_custom_from_mask_pruning, test/nn/test_pruning.py::TestPruningNN::test_global_pruning, test/nn/test_pruning.py::TestPruningNN::test_global_pruning_importance_scores, test/nn/test_pruning.py::TestPruningNN::test_identity_pruning, test/nn/test_pruning.py::TestPruningNN::test_l1_unstructured_pruning, test/nn/test_pruning.py::TestPruningNN::test_l1_unstructured_pruning_with_importance_scores, test/nn/test_pruning.py::TestPruningNN::test_ln_structured_pruning, test/nn/test_pruning.py::TestPruningNN::test_ln_structured_pruning_importance_scores, test/nn/test_pruning.py::TestPruningNN::test_multiple_pruning_calls, test/nn/test_pruning.py::TestPruningNN::test_prune, test/nn/test_pruning.py::TestPruningNN::test_prune_importance_scores, test/nn/test_pruning.py::TestPruningNN::test_prune_importance_scores_mimic_default, test/nn/test_pruning.py::TestPruningNN::test_pruning_container, test/nn/test_pruning.py::TestPruningNN::test_pruning_container_compute_mask, test/nn/test_pruning.py::TestPruningNN::test_pruning_id_consistency, test/nn/test_pruning.py::TestPruningNN::test_pruning_rollback, test/nn/test_pruning.py::TestPruningNN::test_pruning_serialization_model, test/nn/test_pruning.py::TestPruningNN::test_pruning_serialization_state_dict, test/nn/test_pruning.py::TestPruningNN::test_random_pruning, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_0perc, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_forward, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_new_weight, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_orig, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_pickle, test/nn/test_pruning.py::TestPruningNN::test_random_pruning_sizes, test/nn/test_pruning.py::TestPruningNN::test_random_structured_pruning_amount, test/nn/test_pruning.py::TestPruningNN::test_remove_pruning, test/nn/test_pruning.py::TestPruningNN::test_remove_pruning_exception, test/nn/test_pruning.py::TestPruningNN::test_remove_pruning_forward, test/nn/test_pruning.py::TestPruningNN::test_rnn_pruning, test/nn/test_pruning.py::TestPruningNN::test_unstructured_pruning_same_magnitude, test/nn/test_pruning.py::TestPruningNN::test_validate_pruning_amount, test/nn/test_pruning.py::TestPruningNN::test_validate_pruning_amount_init 2023-03-31T06:00:29.1351278Z 2023-03-31T06:00:29.1351444Z nn/test_pruning.py::TestPruningNN::test_compute_nparams_to_prune PASSED [ 2%] 2023-03-31T06:00:29.1351851Z nn/test_pruning.py::TestPruningNN::test_custom_from_mask_pruning PASSED [ 5%] 2023-03-31T06:00:29.1352185Z nn/test_pruning.py::TestPruningNN::test_global_pruning PASSED [ 8%] 2023-03-31T06:00:29.1352588Z nn/test_pruning.py::TestPruningNN::test_global_pruning_importance_scores PASSED [ 11%] 2023-03-31T06:00:29.1352943Z nn/test_pruning.py::TestPruningNN::test_identity_pruning PASSED [ 14%] 2023-03-31T06:00:29.1353369Z nn/test_pruning.py::TestPruningNN::test_l1_unstructured_pruning PASSED [ 17%] 2023-03-31T06:00:29.1353733Z nn/test_pruning.py::TestPruningNN::test_l1_unstructured_pruning_with_importance_scores PASSED [ 20%] 2023-03-31T06:00:29.1354163Z nn/test_pruning.py::TestPruningNN::test_ln_structured_pruning PASSED [ 23%] 2023-03-31T06:00:29.1354504Z nn/test_pruning.py::TestPruningNN::test_ln_structured_pruning_importance_scores PASSED [ 26%] 2023-03-31T06:00:29.1354925Z nn/test_pruning.py::TestPruningNN::test_multiple_pruning_calls PASSED [ 29%] 2023-03-31T06:00:29.1355255Z nn/test_pruning.py::TestPruningNN::test_prune PASSED [ 32%] 2023-03-31T06:00:29.1355642Z nn/test_pruning.py::TestPruningNN::test_prune_importance_scores PASSED [ 35%] 2023-03-31T06:00:29.1355980Z nn/test_pruning.py::TestPruningNN::test_prune_importance_scores_mimic_default PASSED [ 38%] 2023-03-31T06:00:29.1356391Z nn/test_pruning.py::TestPruningNN::test_pruning_container PASSED [ 41%] 2023-03-31T06:00:29.1356738Z nn/test_pruning.py::TestPruningNN::test_pruning_container_compute_mask PASSED [ 44%] 2023-03-31T06:00:29.1357139Z nn/test_pruning.py::TestPruningNN::test_pruning_id_consistency PASSED [ 47%] 2023-03-31T06:00:29.1357470Z nn/test_pruning.py::TestPruningNN::test_pruning_rollback PASSED [ 50%] 2023-03-31T06:00:29.1357870Z nn/test_pruning.py::TestPruningNN::test_pruning_serialization_model PASSED [ 52%] 2023-03-31T06:00:29.1358227Z nn/test_pruning.py::TestPruningNN::test_pruning_serialization_state_dict PASSED [ 55%] 2023-03-31T06:00:29.1358653Z nn/test_pruning.py::TestPruningNN::test_random_pruning PASSED [ 58%] 2023-03-31T06:00:29.1359008Z nn/test_pruning.py::TestPruningNN::test_random_pruning_0perc PASSED [ 61%] 2023-03-31T06:00:29.1359408Z nn/test_pruning.py::TestPruningNN::test_random_pruning_forward PASSED [ 64%] 2023-03-31T06:00:29.1359729Z nn/test_pruning.py::TestPruningNN::test_random_pruning_new_weight PASSED [ 67%] 2023-03-31T06:00:29.1360092Z nn/test_pruning.py::TestPruningNN::test_random_pruning_orig PASSED [ 70%] 2023-03-31T06:00:29.1360443Z nn/test_pruning.py::TestPruningNN::test_random_pruning_pickle PASSED [ 73%] 2023-03-31T06:00:29.1360773Z nn/test_pruning.py::TestPruningNN::test_random_pruning_sizes PASSED [ 76%] 2023-03-31T06:00:29.1361178Z nn/test_pruning.py::TestPruningNN::test_random_structured_pruning_amount PASSED [ 79%] 2023-03-31T06:00:29.1361521Z nn/test_pruning.py::TestPruningNN::test_remove_pruning PASSED [ 82%] 2023-03-31T06:00:29.1361932Z nn/test_pruning.py::TestPruningNN::test_remove_pruning_exception PASSED [ 85%] 2023-03-31T06:00:29.1362255Z nn/test_pruning.py::TestPruningNN::test_remove_pruning_forward PASSED [ 88%] 2023-03-31T06:00:29.1362636Z nn/test_pruning.py::TestPruningNN::test_rnn_pruning PASSED [ 91%] 2023-03-31T06:00:29.1362977Z nn/test_pruning.py::TestPruningNN::test_unstructured_pruning_same_magnitude PASSED [ 94%] 2023-03-31T06:00:29.1363622Z nn/test_pruning.py::TestPruningNN::test_validate_pruning_amount PASSED [ 97%] 2023-03-31T06:00:29.1363964Z nn/test_pruning.py::TestPruningNN::test_validate_pruning_amount_init PASSED [100%] 2023-03-31T06:00:29.1364227Z 2023-03-31T06:00:29.1364693Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_pruning/nn.test_pruning-d14395771e0d5c57.xml - 2023-03-31T06:00:29.1365154Z ============================== 34 passed in 0.18s ============================== 2023-03-31T06:00:29.1365466Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:00:29.1365757Z 2023-03-31T06:00:29.1366034Z ##[endgroup] 2023-03-31T06:00:29.1366531Z FINISHED PRINTING LOG FILE of nn/test_pruning (/var/lib/jenkins/workspace/test/test-reports/nn-test_pruning_gjlazrm2.log) 2023-03-31T06:00:29.1366769Z 2023-03-31T06:00:31.2424671Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:00:31.2425886Z Ignoring disabled issues: [] 2023-03-31T06:22:41.3395134Z 2023-03-31T06:22:41.3395819Z Expand the folded group to see the log file of inductor/test_cpp_wrapper 2023-03-31T06:22:41.3397946Z ##[group]PRINTING LOG FILE of inductor/test_cpp_wrapper (/var/lib/jenkins/workspace/test/test-reports/inductor-test_cpp_wrapper_vfreyo6s.log) 2023-03-31T06:22:41.3405436Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:22:41.3406789Z Test results will be stored in test-reports/python-pytest/inductor.test_cpp_wrapper/inductor.test_cpp_wrapper-28bf4fee1260856b.xml 2023-03-31T06:22:41.3407595Z ============================= test session starts ============================== 2023-03-31T06:22:41.3408623Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:22:41.3409066Z cachedir: .pytest_cache 2023-03-31T06:22:41.3409904Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:22:41.3410578Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:22:41.3411294Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:22:41.3411798Z collecting ... collected 18 items 2023-03-31T06:22:41.3414668Z Running 18 items in this shard: test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_as_strided_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_bitwise_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_bmm1_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_bmm2_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_cat_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear1_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear2_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear_packed_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear_unary_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_lowmem_dropout1_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_mm_views_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_profiler_mark_wrapper_call_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_reduction1_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_relu_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_silu_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_sum_dtype_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_sum_int_cpu, test/inductor/test_cpp_wrapper.py::TestCppWrapper::test_transpose_cpu 2023-03-31T06:22:41.3417312Z 2023-03-31T06:22:41.3418072Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_as_strided_cpu [2023-03-31 05:31:30,838] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 0 2023-03-31T06:22:41.3419147Z [2023-03-31 05:31:33,527] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/dk/cdkr4vltsvywb2ay2a6qyoj72kcjfdleppgtpvdtakf7dmqrsfqe.py 2023-03-31T06:22:41.3420060Z [2023-03-31 05:31:33,528] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3420312Z 2023-03-31T06:22:41.3420469Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3420931Z import torch 2023-03-31T06:22:41.3421211Z import math 2023-03-31T06:22:41.3421460Z import random 2023-03-31T06:22:41.3421766Z import os 2023-03-31T06:22:41.3422019Z import tempfile 2023-03-31T06:22:41.3422325Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3422649Z 2023-03-31T06:22:41.3422832Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3423271Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3423755Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3424026Z 2023-03-31T06:22:41.3424165Z aten = torch.ops.aten 2023-03-31T06:22:41.3424586Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3425025Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3425260Z 2023-03-31T06:22:41.3425394Z 2023-03-31T06:22:41.3425684Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3426287Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3426903Z extern "C" void kernel(float* in_out_ptr0, 2023-03-31T06:22:41.3427267Z const float* in_ptr0) 2023-03-31T06:22:41.3427567Z { 2023-03-31T06:22:41.3427884Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3428219Z { 2023-03-31T06:22:41.3428476Z { 2023-03-31T06:22:41.3428763Z #pragma omp for 2023-03-31T06:22:41.3429204Z for(long i0=static_cast(0); i0(512); i0+=static_cast(1)) 2023-03-31T06:22:41.3429610Z { 2023-03-31T06:22:41.3430044Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3430589Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3430967Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3431387Z auto tmp3 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3431802Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.3432177Z tmp4.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3432522Z } 2023-03-31T06:22:41.3432853Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3433318Z for(long i0=static_cast(4096); i0(4096); i0+=static_cast(1)) 2023-03-31T06:22:41.3433726Z { 2023-03-31T06:22:41.3434157Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3434586Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3434934Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3435293Z auto tmp3 = static_cast(2); 2023-03-31T06:22:41.3435655Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.3436016Z in_out_ptr0[static_cast(i0)] = tmp4; 2023-03-31T06:22:41.3436342Z } 2023-03-31T06:22:41.3436614Z } 2023-03-31T06:22:41.3436846Z } 2023-03-31T06:22:41.3437120Z } 2023-03-31T06:22:41.3437411Z ''') 2023-03-31T06:22:41.3437568Z 2023-03-31T06:22:41.3437702Z async_compile.wait(globals()) 2023-03-31T06:22:41.3438038Z del async_compile 2023-03-31T06:22:41.3438407Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3438761Z wrapper = ( 2023-03-31T06:22:41.3439067Z ''' 2023-03-31T06:22:41.3439331Z #include 2023-03-31T06:22:41.3439619Z #include 2023-03-31T06:22:41.3439808Z 2023-03-31T06:22:41.3439969Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3440183Z 2023-03-31T06:22:41.3440331Z template 2023-03-31T06:22:41.3440742Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3441159Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3441500Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3441764Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3442155Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3442620Z return kernel_cpp; 2023-03-31T06:22:41.3442912Z } 2023-03-31T06:22:41.3443457Z std::vector call_0(std::vector args) { 2023-03-31T06:22:41.3443856Z at::Tensor arg0_1; 2023-03-31T06:22:41.3444182Z arg0_1 = args[0]; 2023-03-31T06:22:41.3444565Z auto buf0 = at::empty_strided({64, 64}, {64, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3445061Z auto buf1 = at::as_strided(buf0, {8, 8, 64}, {512, 64, 1}); buf0.reset(); // reuse 2023-03-31T06:22:41.3445809Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/tu/ctuexftsxfjy7ileno6nb5rem2atkxrpvh23rzqbdodo3lik7azd.so"); 2023-03-31T06:22:41.3446556Z kernel_cpp_0((float*)(buf1.data_ptr()), (float*)(arg0_1.data_ptr())); 2023-03-31T06:22:41.3447043Z return std::vector({at::as_strided(arg0_1, {8, 8, 64}, {512, 64, 1}), buf1}); 2023-03-31T06:22:41.3447548Z } 2023-03-31T06:22:41.3447841Z ''' 2023-03-31T06:22:41.3448090Z ) 2023-03-31T06:22:41.3448222Z 2023-03-31T06:22:41.3448347Z module = load_inline( 2023-03-31T06:22:41.3448998Z name='inline_extension_curlabymyuqckgxav5xqstz47v5d5u2ubnjbinvml3mpxvpdopzl', 2023-03-31T06:22:41.3449502Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3449906Z functions=['call_0'], 2023-03-31T06:22:41.3450647Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3451215Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3452726Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3453599Z 2023-03-31T06:22:41.3453725Z def _wrap_func(f): 2023-03-31T06:22:41.3454049Z def g(args): 2023-03-31T06:22:41.3454335Z return f(args) 2023-03-31T06:22:41.3454659Z return g 2023-03-31T06:22:41.3454977Z call = _wrap_func(module.call_0) 2023-03-31T06:22:41.3455196Z 2023-03-31T06:22:41.3455204Z 2023-03-31T06:22:41.3455362Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3455763Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3456342Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3457039Z arg0_1 = rand_strided({64, 64}, {64, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3457489Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.3457713Z 2023-03-31T06:22:41.3457722Z 2023-03-31T06:22:41.3457851Z if __name__ == "__main__": 2023-03-31T06:22:41.3458133Z import argparse 2023-03-31T06:22:41.3458518Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3458761Z 2023-03-31T06:22:41.3458935Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3459665Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3460537Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3461353Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3461812Z args = parser.parse_args() 2023-03-31T06:22:41.3462016Z 2023-03-31T06:22:41.3462144Z if args.benchmark_kernels: 2023-03-31T06:22:41.3462599Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3462978Z else: 2023-03-31T06:22:41.3463316Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3463555Z benchmark_compiled_module() 2023-03-31T06:22:41.3463727Z 2023-03-31T06:22:41.3463805Z if p: 2023-03-31T06:22:41.3464243Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3464658Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3465054Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3465233Z 2023-03-31T06:22:41.3465737Z [2023-03-31 05:31:33,528] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 0 2023-03-31T06:22:41.3466379Z [2023-03-31 05:31:33,712] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 1 2023-03-31T06:22:41.3467387Z [2023-03-31 05:31:36,318] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/6l/c6ldh4yw5egfkkepqxmomd5q65nfzi2hfavflexvxktpbxgh56lc.py 2023-03-31T06:22:41.3468233Z [2023-03-31 05:31:36,318] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3468609Z 2023-03-31T06:22:41.3468768Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3469177Z import torch 2023-03-31T06:22:41.3469558Z import math 2023-03-31T06:22:41.3469996Z import random 2023-03-31T06:22:41.3470375Z import os 2023-03-31T06:22:41.3470815Z import tempfile 2023-03-31T06:22:41.3501846Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3502149Z 2023-03-31T06:22:41.3502339Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3502838Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3503296Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3503574Z 2023-03-31T06:22:41.3503703Z aten = torch.ops.aten 2023-03-31T06:22:41.3504091Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3504502Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3504685Z 2023-03-31T06:22:41.3504695Z 2023-03-31T06:22:41.3504980Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3505538Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3506053Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.3506383Z float* out_ptr0, 2023-03-31T06:22:41.3506671Z float* out_ptr1) 2023-03-31T06:22:41.3506956Z { 2023-03-31T06:22:41.3507227Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3507515Z { 2023-03-31T06:22:41.3507747Z { 2023-03-31T06:22:41.3507995Z #pragma omp for 2023-03-31T06:22:41.3508618Z for(long i0=static_cast(0); i0(153600); i0+=static_cast(1)) 2023-03-31T06:22:41.3509440Z { 2023-03-31T06:22:41.3509890Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3510403Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3510791Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3511158Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3511447Z } 2023-03-31T06:22:41.3511777Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3512192Z for(long i0=static_cast(1228800); i0(1228800); i0+=static_cast(1)) 2023-03-31T06:22:41.3512543Z { 2023-03-31T06:22:41.3512861Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3513240Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3518779Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3519160Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3519466Z } 2023-03-31T06:22:41.3519692Z } 2023-03-31T06:22:41.3519933Z { 2023-03-31T06:22:41.3520203Z #pragma omp for 2023-03-31T06:22:41.3520600Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.3520959Z { 2023-03-31T06:22:41.3521227Z #pragma GCC ivdep 2023-03-31T06:22:41.3521799Z for(long i1=static_cast(0); i1(2); i1+=static_cast(1)) 2023-03-31T06:22:41.3522155Z { 2023-03-31T06:22:41.3522563Z for(long i2=static_cast(0); i2(11520); i2+=static_cast(1)) 2023-03-31T06:22:41.3522917Z { 2023-03-31T06:22:41.3523572Z auto tmp0 = at::vec::Vectorized::loadu(out_ptr0 + static_cast((8*i2) + (61440*i1) + (153600*i0))); 2023-03-31T06:22:41.3524094Z auto tmp1 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3524529Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3524945Z tmp2.store(out_ptr1 + static_cast((8*i2) + (92160*i1) + (184320*i0))); 2023-03-31T06:22:41.3525293Z } 2023-03-31T06:22:41.3525725Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3526196Z for(long i2=static_cast(92160); i2(92160); i2+=static_cast(1)) 2023-03-31T06:22:41.3526565Z { 2023-03-31T06:22:41.3526916Z auto tmp0 = out_ptr0[static_cast(i2 + (61440*i1) + (153600*i0))]; 2023-03-31T06:22:41.3527344Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3527694Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3528055Z out_ptr1[static_cast(i2 + (92160*i1) + (184320*i0))] = tmp2; 2023-03-31T06:22:41.3528370Z } 2023-03-31T06:22:41.3528615Z } 2023-03-31T06:22:41.3528840Z } 2023-03-31T06:22:41.3529092Z } 2023-03-31T06:22:41.3529313Z } 2023-03-31T06:22:41.3529526Z } 2023-03-31T06:22:41.3529815Z ''') 2023-03-31T06:22:41.3529984Z 2023-03-31T06:22:41.3530127Z async_compile.wait(globals()) 2023-03-31T06:22:41.3530418Z del async_compile 2023-03-31T06:22:41.3530750Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3531086Z wrapper = ( 2023-03-31T06:22:41.3531356Z ''' 2023-03-31T06:22:41.3531599Z #include 2023-03-31T06:22:41.3531854Z #include 2023-03-31T06:22:41.3532000Z 2023-03-31T06:22:41.3532127Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3532307Z 2023-03-31T06:22:41.3532453Z template 2023-03-31T06:22:41.3532835Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3533267Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3533694Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3534072Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3534452Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3534818Z return kernel_cpp; 2023-03-31T06:22:41.3535133Z } 2023-03-31T06:22:41.3535456Z std::vector call_1(std::vector args) { 2023-03-31T06:22:41.3535812Z at::Tensor arg0_1; 2023-03-31T06:22:41.3536093Z arg0_1 = args[0]; 2023-03-31T06:22:41.3536469Z auto buf0 = at::empty_strided({8, 384, 20, 20}, {153600, 1, 7680, 384}, at::ScalarType::Float); 2023-03-31T06:22:41.3536893Z auto buf1 = at::empty_strided({8, 384, 2, 20, 12}, {184320, 1, 92160, 384, 7680}, at::ScalarType::Float); 2023-03-31T06:22:41.3537645Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/g7/cg7sv2vjop6j4rv4wsnfdidz2eews3y5xclynpiawieupqob4d3i.so"); 2023-03-31T06:22:41.3538340Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf1.data_ptr())); 2023-03-31T06:22:41.3538888Z return std::vector({at::as_strided(arg0_1, {8, 384, 2, 20, 12}, {153600, 1, 61440, 384, 7680}), buf1}); 2023-03-31T06:22:41.3539266Z } 2023-03-31T06:22:41.3539582Z ''' 2023-03-31T06:22:41.3539836Z ) 2023-03-31T06:22:41.3539966Z 2023-03-31T06:22:41.3540070Z module = load_inline( 2023-03-31T06:22:41.3540659Z name='inline_extension_covxdet6pl6snnguh47eenhd5pjosuemouzh3mpz7t5cki2ht4ps', 2023-03-31T06:22:41.3541230Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3541627Z functions=['call_1'], 2023-03-31T06:22:41.3542565Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3543275Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3544777Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3545602Z 2023-03-31T06:22:41.3545738Z def _wrap_func(f): 2023-03-31T06:22:41.3546091Z def g(args): 2023-03-31T06:22:41.3546398Z return f(args) 2023-03-31T06:22:41.3546715Z return g 2023-03-31T06:22:41.3547013Z call = _wrap_func(module.call_1) 2023-03-31T06:22:41.3547242Z 2023-03-31T06:22:41.3547251Z 2023-03-31T06:22:41.3547401Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3547804Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3548286Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3549003Z arg0_1 = rand_strided({8, 384, 20, 20}, {153600, 1, 7680, 384}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3549534Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.3549814Z 2023-03-31T06:22:41.3549823Z 2023-03-31T06:22:41.3549971Z if __name__ == "__main__": 2023-03-31T06:22:41.3550284Z import argparse 2023-03-31T06:22:41.3550720Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3551010Z 2023-03-31T06:22:41.3551211Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3552056Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3553110Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3554075Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3554617Z args = parser.parse_args() 2023-03-31T06:22:41.3554851Z 2023-03-31T06:22:41.3555086Z if args.benchmark_kernels: 2023-03-31T06:22:41.3555730Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3556139Z else: 2023-03-31T06:22:41.3556459Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3556854Z benchmark_compiled_module() 2023-03-31T06:22:41.3557034Z 2023-03-31T06:22:41.3557140Z if p: 2023-03-31T06:22:41.3557566Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3557977Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3558386Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3558642Z 2023-03-31T06:22:41.3559114Z [2023-03-31 05:31:36,319] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 1 2023-03-31T06:22:41.3559538Z PASSED [ 5%] 2023-03-31T06:22:41.3560362Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_bitwise_cpu [2023-03-31 05:31:36,398] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 2 2023-03-31T06:22:41.3561285Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3562059Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_not 2023-03-31T06:22:41.3562910Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by store 2023-03-31T06:22:41.3564116Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3565058Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3565937Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_or 2023-03-31T06:22:41.3566817Z [2023-03-31 05:31:36,421] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by store 2023-03-31T06:22:41.3567702Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3568560Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3569442Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_xor 2023-03-31T06:22:41.3570266Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by store 2023-03-31T06:22:41.3571223Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3572174Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.3572777Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_and 2023-03-31T06:22:41.3573271Z [2023-03-31 05:31:36,422] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by store 2023-03-31T06:22:41.3573761Z [2023-03-31 05:31:36,423] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_not 2023-03-31T06:22:41.3574238Z [2023-03-31 05:31:36,423] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_or 2023-03-31T06:22:41.3574709Z [2023-03-31 05:31:36,423] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_xor 2023-03-31T06:22:41.3575164Z [2023-03-31 05:31:36,424] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: op: bitwise_and 2023-03-31T06:22:41.3575846Z [2023-03-31 05:31:39,090] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/oz/coz5kafzxbbxul7febrvkfypa5kjlufarr3wb4xs6qfggkur7lki.py 2023-03-31T06:22:41.3576405Z [2023-03-31 05:31:39,090] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3576585Z 2023-03-31T06:22:41.3576686Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3576877Z import torch 2023-03-31T06:22:41.3577055Z import math 2023-03-31T06:22:41.3577232Z import random 2023-03-31T06:22:41.3577442Z import os 2023-03-31T06:22:41.3577618Z import tempfile 2023-03-31T06:22:41.3577920Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3578138Z 2023-03-31T06:22:41.3578305Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3578720Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3579128Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3579297Z 2023-03-31T06:22:41.3579384Z aten = torch.ops.aten 2023-03-31T06:22:41.3579683Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3580111Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3580301Z 2023-03-31T06:22:41.3580306Z 2023-03-31T06:22:41.3580549Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3580957Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3581305Z extern "C" void kernel(const int* in_ptr0, 2023-03-31T06:22:41.3581534Z const int* in_ptr1, 2023-03-31T06:22:41.3581723Z int* out_ptr0, 2023-03-31T06:22:41.3581919Z int* out_ptr1, 2023-03-31T06:22:41.3582208Z int* out_ptr2, 2023-03-31T06:22:41.3582480Z int* out_ptr3) 2023-03-31T06:22:41.3582655Z { 2023-03-31T06:22:41.3582847Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3583032Z { 2023-03-31T06:22:41.3583190Z { 2023-03-31T06:22:41.3583365Z #pragma omp for 2023-03-31T06:22:41.3583641Z for(long i0=static_cast(0); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.3583873Z { 2023-03-31T06:22:41.3584087Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3584343Z auto tmp2 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.3584556Z auto tmp1 = ~tmp0; 2023-03-31T06:22:41.3584762Z auto tmp3 = tmp0 | tmp2; 2023-03-31T06:22:41.3585014Z auto tmp4 = tmp0 ^ tmp2; 2023-03-31T06:22:41.3585212Z auto tmp5 = tmp0 & tmp2; 2023-03-31T06:22:41.3585442Z out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3585683Z out_ptr1[static_cast(i0)] = tmp3; 2023-03-31T06:22:41.3585911Z out_ptr2[static_cast(i0)] = tmp4; 2023-03-31T06:22:41.3586149Z out_ptr3[static_cast(i0)] = tmp5; 2023-03-31T06:22:41.3586347Z } 2023-03-31T06:22:41.3586502Z } 2023-03-31T06:22:41.3586665Z } 2023-03-31T06:22:41.3586821Z } 2023-03-31T06:22:41.3586986Z ''') 2023-03-31T06:22:41.3587081Z 2023-03-31T06:22:41.3587173Z async_compile.wait(globals()) 2023-03-31T06:22:41.3587371Z del async_compile 2023-03-31T06:22:41.3587584Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3587804Z wrapper = ( 2023-03-31T06:22:41.3587987Z ''' 2023-03-31T06:22:41.3588140Z #include 2023-03-31T06:22:41.3588324Z #include 2023-03-31T06:22:41.3588439Z 2023-03-31T06:22:41.3588543Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3588672Z 2023-03-31T06:22:41.3588770Z template 2023-03-31T06:22:41.3589006Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3589240Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3589482Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3589714Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3589966Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3590234Z return kernel_cpp; 2023-03-31T06:22:41.3590427Z } 2023-03-31T06:22:41.3590642Z std::vector call_2(std::vector args) { 2023-03-31T06:22:41.3590882Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.3591058Z arg0_1 = args[0]; 2023-03-31T06:22:41.3591237Z arg1_1 = args[1]; 2023-03-31T06:22:41.3591475Z auto buf0 = at::empty_strided({64, }, {1, }, at::ScalarType::Int); 2023-03-31T06:22:41.3591765Z auto buf1 = at::empty_strided({64, }, {1, }, at::ScalarType::Int); 2023-03-31T06:22:41.3592038Z auto buf2 = at::empty_strided({64, }, {1, }, at::ScalarType::Int); 2023-03-31T06:22:41.3592319Z auto buf3 = at::empty_strided({64, }, {1, }, at::ScalarType::Int); 2023-03-31T06:22:41.3592767Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/zv/czveuk437w54gk75q555oojuehwhnck2gn52x6iaz3s5kb6rekdv.so"); 2023-03-31T06:22:41.3593291Z kernel_cpp_0((int*)(arg0_1.data_ptr()), (int*)(arg1_1.data_ptr()), (int*)(buf0.data_ptr()), (int*)(buf1.data_ptr()), (int*)(buf2.data_ptr()), (int*)(buf3.data_ptr())); 2023-03-31T06:22:41.3593587Z arg0_1.reset(); 2023-03-31T06:22:41.3593765Z arg1_1.reset(); 2023-03-31T06:22:41.3593992Z return std::vector({buf0, buf1, buf2, buf3}); 2023-03-31T06:22:41.3594190Z } 2023-03-31T06:22:41.3594364Z ''' 2023-03-31T06:22:41.3594516Z ) 2023-03-31T06:22:41.3594598Z 2023-03-31T06:22:41.3594677Z module = load_inline( 2023-03-31T06:22:41.3595060Z name='inline_extension_czr5eaopeh7agiq4afrfy3xpzl6ah74usjbg2q5enknjdyeuurgt', 2023-03-31T06:22:41.3595410Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3595632Z functions=['call_2'], 2023-03-31T06:22:41.3596135Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3596569Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3597422Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3597909Z 2023-03-31T06:22:41.3597973Z def _wrap_func(f): 2023-03-31T06:22:41.3598188Z def g(args): 2023-03-31T06:22:41.3598367Z return f(args) 2023-03-31T06:22:41.3598532Z return g 2023-03-31T06:22:41.3598728Z call = _wrap_func(module.call_2) 2023-03-31T06:22:41.3598859Z 2023-03-31T06:22:41.3598865Z 2023-03-31T06:22:41.3598960Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3599202Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3599465Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3599822Z arg0_1 = rand_strided({64, }, {1, }, device='cpu', dtype=torch.int32) 2023-03-31T06:22:41.3600184Z arg1_1 = rand_strided({64, }, {1, }, device='cpu', dtype=torch.int32) 2023-03-31T06:22:41.3600451Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.3600598Z 2023-03-31T06:22:41.3600604Z 2023-03-31T06:22:41.3600683Z if __name__ == "__main__": 2023-03-31T06:22:41.3600870Z import argparse 2023-03-31T06:22:41.3601110Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3601258Z 2023-03-31T06:22:41.3601370Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3601827Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3602397Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3602909Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3603531Z args = parser.parse_args() 2023-03-31T06:22:41.3603708Z 2023-03-31T06:22:41.3603801Z if args.benchmark_kernels: 2023-03-31T06:22:41.3604122Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3604336Z else: 2023-03-31T06:22:41.3604542Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3604779Z benchmark_compiled_module() 2023-03-31T06:22:41.3604914Z 2023-03-31T06:22:41.3604972Z if p: 2023-03-31T06:22:41.3605225Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3605495Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3605744Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3605913Z 2023-03-31T06:22:41.3606193Z [2023-03-31 05:31:39,090] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 2 2023-03-31T06:22:41.3606476Z PASSED [ 11%] 2023-03-31T06:22:41.3606968Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_bmm1_cpu [2023-03-31 05:31:39,126] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 3 2023-03-31T06:22:41.3607613Z [2023-03-31 05:31:41,850] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/f4/cf4hi37wo446e7rin2auo356gxtuy5jgoyinh4qnrrs6ot5nz4q3.py 2023-03-31T06:22:41.3608111Z [2023-03-31 05:31:41,850] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3608293Z 2023-03-31T06:22:41.3608389Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3608592Z import torch 2023-03-31T06:22:41.3608800Z import math 2023-03-31T06:22:41.3608978Z import random 2023-03-31T06:22:41.3609151Z import os 2023-03-31T06:22:41.3609308Z import tempfile 2023-03-31T06:22:41.3609528Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3609678Z 2023-03-31T06:22:41.3609797Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3610057Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3610346Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3610510Z 2023-03-31T06:22:41.3610590Z aten = torch.ops.aten 2023-03-31T06:22:41.3610830Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3611086Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3611217Z 2023-03-31T06:22:41.3611222Z 2023-03-31T06:22:41.3611411Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3611762Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3612105Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.3612338Z const float* in_ptr1, 2023-03-31T06:22:41.3612543Z float* out_ptr0, 2023-03-31T06:22:41.3612736Z float* out_ptr1) 2023-03-31T06:22:41.3612919Z { 2023-03-31T06:22:41.3613105Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3613285Z { 2023-03-31T06:22:41.3613442Z { 2023-03-31T06:22:41.3613622Z #pragma omp for 2023-03-31T06:22:41.3613877Z for(long i0=static_cast(0); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3614116Z { 2023-03-31T06:22:41.3614375Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3614693Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3614935Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3615176Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3615387Z } 2023-03-31T06:22:41.3615572Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3615859Z for(long i0=static_cast(128); i0(128); i0+=static_cast(1)) 2023-03-31T06:22:41.3616096Z { 2023-03-31T06:22:41.3616294Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3616573Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3616821Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3617036Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3617237Z } 2023-03-31T06:22:41.3617396Z } 2023-03-31T06:22:41.3617540Z { 2023-03-31T06:22:41.3617713Z #pragma omp for 2023-03-31T06:22:41.3617981Z for(long i0=static_cast(0); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3618218Z { 2023-03-31T06:22:41.3618464Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.3618782Z auto tmp1 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3619032Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3619262Z tmp2.store(out_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.3619474Z } 2023-03-31T06:22:41.3619676Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3619946Z for(long i0=static_cast(128); i0(128); i0+=static_cast(1)) 2023-03-31T06:22:41.3620182Z { 2023-03-31T06:22:41.3620396Z auto tmp0 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.3620637Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3620848Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3621076Z out_ptr1[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3621278Z } 2023-03-31T06:22:41.3621501Z } 2023-03-31T06:22:41.3621661Z } 2023-03-31T06:22:41.3621814Z } 2023-03-31T06:22:41.3621975Z ''') 2023-03-31T06:22:41.3622071Z 2023-03-31T06:22:41.3622076Z 2023-03-31T06:22:41.3622214Z kernel_cpp_1 = async_compile.cpp(''' 2023-03-31T06:22:41.3622662Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3623002Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3623201Z { 2023-03-31T06:22:41.3623388Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3623570Z { 2023-03-31T06:22:41.3623724Z { 2023-03-31T06:22:41.3623901Z #pragma omp for 2023-03-31T06:22:41.3624154Z for(long i0=static_cast(0); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3624440Z { 2023-03-31T06:22:41.3624702Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3625024Z auto tmp1 = at::vec::Vectorized(static_cast(3)); 2023-03-31T06:22:41.3625261Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3625502Z tmp2.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3625715Z } 2023-03-31T06:22:41.3625900Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3626184Z for(long i0=static_cast(128); i0(128); i0+=static_cast(1)) 2023-03-31T06:22:41.3626428Z { 2023-03-31T06:22:41.3626634Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3626882Z auto tmp1 = static_cast(3); 2023-03-31T06:22:41.3627103Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3627321Z in_out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3627524Z } 2023-03-31T06:22:41.3627686Z } 2023-03-31T06:22:41.3627833Z } 2023-03-31T06:22:41.3627983Z } 2023-03-31T06:22:41.3628159Z ''') 2023-03-31T06:22:41.3628257Z 2023-03-31T06:22:41.3628348Z async_compile.wait(globals()) 2023-03-31T06:22:41.3628533Z del async_compile 2023-03-31T06:22:41.3628763Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3628984Z wrapper = ( 2023-03-31T06:22:41.3629154Z ''' 2023-03-31T06:22:41.3629318Z #include 2023-03-31T06:22:41.3629552Z #include 2023-03-31T06:22:41.3629678Z 2023-03-31T06:22:41.3629770Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3629900Z 2023-03-31T06:22:41.3629995Z template 2023-03-31T06:22:41.3630243Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3630465Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3630707Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3630954Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3631196Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3631423Z return kernel_cpp; 2023-03-31T06:22:41.3631600Z } 2023-03-31T06:22:41.3631802Z std::vector call_3(std::vector args) { 2023-03-31T06:22:41.3632039Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.3632234Z arg0_1 = args[0]; 2023-03-31T06:22:41.3632403Z arg1_1 = args[1]; 2023-03-31T06:22:41.3632642Z auto buf0 = at::empty_strided({2, 8, 8}, {64, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3632898Z at::bmm_out(buf0, arg0_1, arg1_1); 2023-03-31T06:22:41.3633157Z auto buf1 = at::empty_strided({2, 8, 8}, {64, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3633441Z auto buf2 = at::empty_strided({2, 8, 8}, {64, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3633915Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/om/com25xfnmjsnbgfgowqbkswbuv2vlnpedvowzms3f7s2ubsdwblp.so"); 2023-03-31T06:22:41.3634475Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(arg1_1.data_ptr()), (float*)(buf1.data_ptr()), (float*)(buf2.data_ptr())); 2023-03-31T06:22:41.3634752Z arg0_1.reset(); 2023-03-31T06:22:41.3634920Z arg1_1.reset(); 2023-03-31T06:22:41.3635159Z auto buf3 = at::empty_strided({2, 8, 8}, {64, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3635406Z at::bmm_out(buf3, buf1, buf2); 2023-03-31T06:22:41.3635587Z buf1.reset(); 2023-03-31T06:22:41.3635764Z buf2.reset(); 2023-03-31T06:22:41.3635969Z auto buf4 = buf3; buf3.reset(); // reuse 2023-03-31T06:22:41.3636360Z static auto kernel_cpp_1 = load_cpp_kernel("/tmp/torchinductor_jenkins/4f/c4f5m2numoxe2kpicd3q2s2mkb776kp7ptxb3qpp3u3dlsxwq2rk.so"); 2023-03-31T06:22:41.3636752Z kernel_cpp_1((float*)(buf4.data_ptr())); 2023-03-31T06:22:41.3636996Z return std::vector({buf0, buf4}); 2023-03-31T06:22:41.3637227Z } 2023-03-31T06:22:41.3637386Z ''' 2023-03-31T06:22:41.3637537Z ) 2023-03-31T06:22:41.3637629Z 2023-03-31T06:22:41.3637709Z module = load_inline( 2023-03-31T06:22:41.3638080Z name='inline_extension_ceo7brj2m6v4xf47oklqvbetb2wshtoiw7otup3gtpctbfwjjybv', 2023-03-31T06:22:41.3638388Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3638621Z functions=['call_3'], 2023-03-31T06:22:41.3639103Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3639536Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3640397Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3640883Z 2023-03-31T06:22:41.3640960Z def _wrap_func(f): 2023-03-31T06:22:41.3641129Z def g(args): 2023-03-31T06:22:41.3641308Z return f(args) 2023-03-31T06:22:41.3641488Z return g 2023-03-31T06:22:41.3641668Z call = _wrap_func(module.call_3) 2023-03-31T06:22:41.3641796Z 2023-03-31T06:22:41.3641801Z 2023-03-31T06:22:41.3641895Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3642138Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3642410Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3642803Z arg0_1 = rand_strided({2, 8, 8}, {64, 8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3643393Z arg1_1 = rand_strided({2, 8, 8}, {64, 8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3643674Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.3643824Z 2023-03-31T06:22:41.3643829Z 2023-03-31T06:22:41.3643895Z if __name__ == "__main__": 2023-03-31T06:22:41.3644087Z import argparse 2023-03-31T06:22:41.3644326Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3644484Z 2023-03-31T06:22:41.3644598Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3645042Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3645606Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3646140Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3646430Z args = parser.parse_args() 2023-03-31T06:22:41.3646558Z 2023-03-31T06:22:41.3646647Z if args.benchmark_kernels: 2023-03-31T06:22:41.3646956Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3647181Z else: 2023-03-31T06:22:41.3647379Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3647617Z benchmark_compiled_module() 2023-03-31T06:22:41.3647750Z 2023-03-31T06:22:41.3647818Z if p: 2023-03-31T06:22:41.3648134Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3648404Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3648665Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3648838Z 2023-03-31T06:22:41.3649109Z [2023-03-31 05:31:41,851] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 3 2023-03-31T06:22:41.3649582Z [2023-03-31 05:31:41,889] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 4 2023-03-31T06:22:41.3650187Z [2023-03-31 05:31:44,648] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/tg/ctgaj6kdsbzecyi7mzmoxiqan3uzjtgr6zmwg3e7cblxgttmbgda.py 2023-03-31T06:22:41.3650701Z [2023-03-31 05:31:44,648] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3650920Z 2023-03-31T06:22:41.3651006Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3651209Z import torch 2023-03-31T06:22:41.3651389Z import math 2023-03-31T06:22:41.3651550Z import random 2023-03-31T06:22:41.3651720Z import os 2023-03-31T06:22:41.3651896Z import tempfile 2023-03-31T06:22:41.3652104Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3652255Z 2023-03-31T06:22:41.3652373Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3652646Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3652932Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3653082Z 2023-03-31T06:22:41.3653162Z aten = torch.ops.aten 2023-03-31T06:22:41.3653412Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3653669Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3653802Z 2023-03-31T06:22:41.3653807Z 2023-03-31T06:22:41.3653940Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3654292Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3654641Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.3654868Z const float* in_ptr1, 2023-03-31T06:22:41.3655061Z float* out_ptr0, 2023-03-31T06:22:41.3655260Z float* out_ptr1) 2023-03-31T06:22:41.3655438Z { 2023-03-31T06:22:41.3655610Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3655803Z { 2023-03-31T06:22:41.3656001Z { 2023-03-31T06:22:41.3656212Z #pragma omp for 2023-03-31T06:22:41.3656481Z for(long i0=static_cast(0); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3656719Z { 2023-03-31T06:22:41.3656964Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3657282Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3657536Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3657768Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3657978Z } 2023-03-31T06:22:41.3658177Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3658459Z for(long i0=static_cast(128); i0(128); i0+=static_cast(1)) 2023-03-31T06:22:41.3658686Z { 2023-03-31T06:22:41.3658899Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3659144Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3659356Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3659586Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3659786Z } 2023-03-31T06:22:41.3659936Z } 2023-03-31T06:22:41.3660093Z { 2023-03-31T06:22:41.3660266Z #pragma omp for 2023-03-31T06:22:41.3660517Z for(long i0=static_cast(0); i0(10); i0+=static_cast(1)) 2023-03-31T06:22:41.3660754Z { 2023-03-31T06:22:41.3661047Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.3661366Z auto tmp1 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3661605Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3661845Z tmp2.store(out_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.3662056Z } 2023-03-31T06:22:41.3662243Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3662615Z for(long i0=static_cast(80); i0(80); i0+=static_cast(1)) 2023-03-31T06:22:41.3662854Z { 2023-03-31T06:22:41.3663054Z auto tmp0 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.3663295Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3663553Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3663770Z out_ptr1[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3663969Z } 2023-03-31T06:22:41.3664138Z } 2023-03-31T06:22:41.3664281Z } 2023-03-31T06:22:41.3664434Z } 2023-03-31T06:22:41.3664611Z ''') 2023-03-31T06:22:41.3664707Z 2023-03-31T06:22:41.3664713Z 2023-03-31T06:22:41.3664853Z kernel_cpp_1 = async_compile.cpp(''' 2023-03-31T06:22:41.3665200Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3665550Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3665748Z { 2023-03-31T06:22:41.3665922Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3666117Z { 2023-03-31T06:22:41.3666272Z { 2023-03-31T06:22:41.3666432Z #pragma omp for 2023-03-31T06:22:41.3666699Z for(long i0=static_cast(0); i0(20); i0+=static_cast(1)) 2023-03-31T06:22:41.3666939Z { 2023-03-31T06:22:41.3667188Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3667510Z auto tmp1 = at::vec::Vectorized(static_cast(3)); 2023-03-31T06:22:41.3667762Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3668004Z tmp2.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3668206Z } 2023-03-31T06:22:41.3668406Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3668724Z for(long i0=static_cast(160); i0(160); i0+=static_cast(1)) 2023-03-31T06:22:41.3668977Z { 2023-03-31T06:22:41.3669192Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3669438Z auto tmp1 = static_cast(3); 2023-03-31T06:22:41.3669649Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3669882Z in_out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3670087Z } 2023-03-31T06:22:41.3670235Z } 2023-03-31T06:22:41.3670391Z } 2023-03-31T06:22:41.3670543Z } 2023-03-31T06:22:41.3670701Z ''') 2023-03-31T06:22:41.3670796Z 2023-03-31T06:22:41.3670888Z async_compile.wait(globals()) 2023-03-31T06:22:41.3671087Z del async_compile 2023-03-31T06:22:41.3671302Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3671521Z wrapper = ( 2023-03-31T06:22:41.3671704Z ''' 2023-03-31T06:22:41.3671857Z #include 2023-03-31T06:22:41.3672043Z #include 2023-03-31T06:22:41.3672163Z 2023-03-31T06:22:41.3672257Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3672387Z 2023-03-31T06:22:41.3672481Z template 2023-03-31T06:22:41.3672715Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3672951Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3673193Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3673428Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3673682Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3673962Z return kernel_cpp; 2023-03-31T06:22:41.3674127Z } 2023-03-31T06:22:41.3674343Z std::vector call_4(std::vector args) { 2023-03-31T06:22:41.3674581Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.3674760Z arg0_1 = args[0]; 2023-03-31T06:22:41.3674937Z arg1_1 = args[1]; 2023-03-31T06:22:41.3675183Z auto buf0 = at::empty_strided({1, 16, 10}, {160, 10, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3675429Z at::bmm_out(buf0, arg0_1, arg1_1); 2023-03-31T06:22:41.3675692Z auto buf1 = at::empty_strided({1, 16, 8}, {128, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3675987Z auto buf2 = at::empty_strided({1, 8, 10}, {80, 10, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3676467Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/it/citrikrerzwfoict5wvxisxrwm32lk3pzidlplpmkow54kmmuhor.so"); 2023-03-31T06:22:41.3677011Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(arg1_1.data_ptr()), (float*)(buf1.data_ptr()), (float*)(buf2.data_ptr())); 2023-03-31T06:22:41.3677286Z arg0_1.reset(); 2023-03-31T06:22:41.3677463Z arg1_1.reset(); 2023-03-31T06:22:41.3677704Z auto buf3 = at::empty_strided({1, 16, 10}, {160, 10, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3677944Z at::bmm_out(buf3, buf1, buf2); 2023-03-31T06:22:41.3678136Z buf1.reset(); 2023-03-31T06:22:41.3678315Z buf2.reset(); 2023-03-31T06:22:41.3678508Z auto buf4 = buf3; buf3.reset(); // reuse 2023-03-31T06:22:41.3678912Z static auto kernel_cpp_1 = load_cpp_kernel("/tmp/torchinductor_jenkins/xo/cxoi2ezy7jlmxiq7dqe4nbbxx5cboy4k2yj3mhxj4cr44czxc5ip.so"); 2023-03-31T06:22:41.3679303Z kernel_cpp_1((float*)(buf4.data_ptr())); 2023-03-31T06:22:41.3679531Z return std::vector({buf0, buf4}); 2023-03-31T06:22:41.3679731Z } 2023-03-31T06:22:41.3679902Z ''' 2023-03-31T06:22:41.3680042Z ) 2023-03-31T06:22:41.3680131Z 2023-03-31T06:22:41.3680209Z module = load_inline( 2023-03-31T06:22:41.3680599Z name='inline_extension_clwkpn3fgybq3o7izrgzedd3nfvhzsnryzc3qmajosbtx2qtq5x4', 2023-03-31T06:22:41.3680910Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3681131Z functions=['call_4'], 2023-03-31T06:22:41.3681627Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3682130Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3682992Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3683618Z 2023-03-31T06:22:41.3683695Z def _wrap_func(f): 2023-03-31T06:22:41.3683876Z def g(args): 2023-03-31T06:22:41.3684056Z return f(args) 2023-03-31T06:22:41.3684220Z return g 2023-03-31T06:22:41.3684410Z call = _wrap_func(module.call_4) 2023-03-31T06:22:41.3684538Z 2023-03-31T06:22:41.3684543Z 2023-03-31T06:22:41.3684636Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3684866Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3685138Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3685517Z arg0_1 = rand_strided({1, 16, 8}, {128, 8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3685891Z arg1_1 = rand_strided({1, 8, 10}, {80, 10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3686153Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.3686307Z 2023-03-31T06:22:41.3686312Z 2023-03-31T06:22:41.3686389Z if __name__ == "__main__": 2023-03-31T06:22:41.3686579Z import argparse 2023-03-31T06:22:41.3686807Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3686965Z 2023-03-31T06:22:41.3687141Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3687599Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3688169Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3688688Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3688991Z args = parser.parse_args() 2023-03-31T06:22:41.3689121Z 2023-03-31T06:22:41.3689210Z if args.benchmark_kernels: 2023-03-31T06:22:41.3689508Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3689732Z else: 2023-03-31T06:22:41.3689986Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3690208Z benchmark_compiled_module() 2023-03-31T06:22:41.3690340Z 2023-03-31T06:22:41.3690408Z if p: 2023-03-31T06:22:41.3690661Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3690930Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3691217Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3691382Z 2023-03-31T06:22:41.3691672Z [2023-03-31 05:31:44,648] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 4 2023-03-31T06:22:41.3691953Z PASSED [ 16%] 2023-03-31T06:22:41.3692428Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_bmm2_cpu [2023-03-31 05:31:44,677] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 5 2023-03-31T06:22:41.3693088Z [2023-03-31 05:31:46,198] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/7v/c7vu52mmn33t63qz6vh5hxo6msxwczyxkm3luygul6qqyx6fjkxq.py 2023-03-31T06:22:41.3693587Z [2023-03-31 05:31:46,198] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3693768Z 2023-03-31T06:22:41.3693866Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3694053Z import torch 2023-03-31T06:22:41.3694228Z import math 2023-03-31T06:22:41.3694400Z import random 2023-03-31T06:22:41.3694558Z import os 2023-03-31T06:22:41.3694729Z import tempfile 2023-03-31T06:22:41.3694950Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3695098Z 2023-03-31T06:22:41.3695245Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3695551Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3695838Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3696002Z 2023-03-31T06:22:41.3696081Z aten = torch.ops.aten 2023-03-31T06:22:41.3696321Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3696582Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3696713Z 2023-03-31T06:22:41.3696803Z async_compile.wait(globals()) 2023-03-31T06:22:41.3696988Z del async_compile 2023-03-31T06:22:41.3697216Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3697437Z wrapper = ( 2023-03-31T06:22:41.3697612Z ''' 2023-03-31T06:22:41.3697779Z #include 2023-03-31T06:22:41.3697962Z #include 2023-03-31T06:22:41.3698451Z 2023-03-31T06:22:41.3698536Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3698667Z 2023-03-31T06:22:41.3698764Z template 2023-03-31T06:22:41.3699016Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3699240Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3699483Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3699726Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3699982Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3700200Z return kernel_cpp; 2023-03-31T06:22:41.3700373Z } 2023-03-31T06:22:41.3700585Z std::vector call_5(std::vector args) { 2023-03-31T06:22:41.3700847Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.3701040Z arg0_1 = args[0]; 2023-03-31T06:22:41.3701219Z arg1_1 = args[1]; 2023-03-31T06:22:41.3701451Z auto buf0 = at::empty_strided({1, 8, 8}, {64, 8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3701736Z at::bmm_out(buf0, at::as_strided(arg0_1, {1, 8, 8}, {64, 1, 8}), arg1_1); 2023-03-31T06:22:41.3701957Z arg0_1.reset(); 2023-03-31T06:22:41.3702122Z arg1_1.reset(); 2023-03-31T06:22:41.3702334Z return std::vector({buf0}); 2023-03-31T06:22:41.3702610Z } 2023-03-31T06:22:41.3702771Z ''' 2023-03-31T06:22:41.3702924Z ) 2023-03-31T06:22:41.3703016Z 2023-03-31T06:22:41.3703094Z module = load_inline( 2023-03-31T06:22:41.3703468Z name='inline_extension_cuoby27fw4is7ou4r4r5ktzn42om2p4u2bdjofibp52rvgiexvl3', 2023-03-31T06:22:41.3703753Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3704024Z functions=['call_5'], 2023-03-31T06:22:41.3704525Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3704950Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3705805Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3706281Z 2023-03-31T06:22:41.3706359Z def _wrap_func(f): 2023-03-31T06:22:41.3706536Z def g(args): 2023-03-31T06:22:41.3706703Z return f(args) 2023-03-31T06:22:41.3706880Z return g 2023-03-31T06:22:41.3707072Z call = _wrap_func(module.call_5) 2023-03-31T06:22:41.3707205Z 2023-03-31T06:22:41.3707210Z 2023-03-31T06:22:41.3707291Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3707532Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3707806Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3708165Z arg0_1 = rand_strided({1, 8, 8}, {64, 8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3708538Z arg1_1 = rand_strided({1, 8, 8}, {64, 8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3708818Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.3709012Z 2023-03-31T06:22:41.3709043Z 2023-03-31T06:22:41.3709123Z if __name__ == "__main__": 2023-03-31T06:22:41.3709300Z import argparse 2023-03-31T06:22:41.3709537Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3709696Z 2023-03-31T06:22:41.3709806Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3710252Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3710947Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3711470Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3711850Z args = parser.parse_args() 2023-03-31T06:22:41.3711979Z 2023-03-31T06:22:41.3712057Z if args.benchmark_kernels: 2023-03-31T06:22:41.3712512Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3712739Z else: 2023-03-31T06:22:41.3712934Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3713173Z benchmark_compiled_module() 2023-03-31T06:22:41.3713440Z 2023-03-31T06:22:41.3713550Z if p: 2023-03-31T06:22:41.3713802Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3714060Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3714326Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3714491Z 2023-03-31T06:22:41.3714818Z [2023-03-31 05:31:46,199] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 5 2023-03-31T06:22:41.3715090Z PASSED [ 22%] 2023-03-31T06:22:41.3715586Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_cat_cpu [2023-03-31 05:31:46,251] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 6 2023-03-31T06:22:41.3716156Z [2023-03-31 05:31:46,307] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3716691Z [2023-03-31 05:31:46,307] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.3717202Z [2023-03-31 05:31:46,307] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3717752Z [2023-03-31 05:31:46,307] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.3718266Z [2023-03-31 05:31:46,309] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3718779Z [2023-03-31 05:31:46,310] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3719583Z [2023-03-31 05:31:48,902] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/db/cdbj6uzu6uasxhdo6rzz7uma5237lalrh6n6jieujq52jhwhf5gs.py 2023-03-31T06:22:41.3720093Z [2023-03-31 05:31:48,902] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3720271Z 2023-03-31T06:22:41.3720370Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3720572Z import torch 2023-03-31T06:22:41.3720735Z import math 2023-03-31T06:22:41.3720910Z import random 2023-03-31T06:22:41.3721085Z import os 2023-03-31T06:22:41.3721243Z import tempfile 2023-03-31T06:22:41.3721464Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3721615Z 2023-03-31T06:22:41.3721737Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3721997Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3722282Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3722445Z 2023-03-31T06:22:41.3722526Z aten = torch.ops.aten 2023-03-31T06:22:41.3722815Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3723339Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3723475Z 2023-03-31T06:22:41.3723480Z 2023-03-31T06:22:41.3723629Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3723983Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3724320Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.3724547Z float* out_ptr0, 2023-03-31T06:22:41.3724748Z float* out_ptr1, 2023-03-31T06:22:41.3724935Z float* out_ptr2, 2023-03-31T06:22:41.3725135Z float* out_ptr3, 2023-03-31T06:22:41.3725331Z float* out_ptr4, 2023-03-31T06:22:41.3725519Z double* out_ptr5, 2023-03-31T06:22:41.3725723Z double* out_ptr6) 2023-03-31T06:22:41.3725903Z { 2023-03-31T06:22:41.3726077Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3726271Z { 2023-03-31T06:22:41.3726430Z { 2023-03-31T06:22:41.3726593Z #pragma omp for 2023-03-31T06:22:41.3726863Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.3727102Z { 2023-03-31T06:22:41.3727352Z for(long i1=static_cast(0); i1(2); i1+=static_cast(1)) 2023-03-31T06:22:41.3727581Z { 2023-03-31T06:22:41.3727860Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast((8*i1) + (16*i0))); 2023-03-31T06:22:41.3728257Z auto tmp1 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3728505Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3728757Z tmp0.store(out_ptr0 + static_cast((8*i1) + (36*i0))); 2023-03-31T06:22:41.3729035Z tmp2.store(out_ptr1 + static_cast((8*i1) + (36*i0))); 2023-03-31T06:22:41.3729252Z } 2023-03-31T06:22:41.3729442Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3729724Z for(long i1=static_cast(16); i1(16); i1+=static_cast(1)) 2023-03-31T06:22:41.3729966Z { 2023-03-31T06:22:41.3730175Z auto tmp0 = in_ptr0[static_cast(i1 + (16*i0))]; 2023-03-31T06:22:41.3730431Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3730714Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3730941Z out_ptr0[static_cast(i1 + (36*i0))] = tmp0; 2023-03-31T06:22:41.3731200Z out_ptr1[static_cast(i1 + (36*i0))] = tmp2; 2023-03-31T06:22:41.3731404Z } 2023-03-31T06:22:41.3731557Z } 2023-03-31T06:22:41.3731719Z } 2023-03-31T06:22:41.3731876Z { 2023-03-31T06:22:41.3732039Z #pragma omp for 2023-03-31T06:22:41.3732305Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.3732546Z { 2023-03-31T06:22:41.3732795Z for(long i1=static_cast(0); i1(0); i1+=static_cast(1)) 2023-03-31T06:22:41.3733021Z { 2023-03-31T06:22:41.3733289Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast((8*i1) + (16*i0))); 2023-03-31T06:22:41.3733611Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3733855Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3734106Z tmp2.store(out_ptr2 + static_cast((8*i1) + (36*i0))); 2023-03-31T06:22:41.3734326Z } 2023-03-31T06:22:41.3734513Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3734792Z for(long i1=static_cast(0); i1(4); i1+=static_cast(1)) 2023-03-31T06:22:41.3735030Z { 2023-03-31T06:22:41.3735288Z auto tmp0 = in_ptr0[static_cast(i1 + (16*i0))]; 2023-03-31T06:22:41.3735567Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3735793Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3736034Z out_ptr2[static_cast(i1 + (36*i0))] = tmp2; 2023-03-31T06:22:41.3736229Z } 2023-03-31T06:22:41.3736392Z } 2023-03-31T06:22:41.3736557Z } 2023-03-31T06:22:41.3736701Z { 2023-03-31T06:22:41.3736872Z #pragma omp for 2023-03-31T06:22:41.3737140Z for(long i0=static_cast(0); i0(128); i0+=static_cast(1)) 2023-03-31T06:22:41.3737364Z { 2023-03-31T06:22:41.3737637Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3737951Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3738478Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.3738776Z auto tmp3 = static_cast(tmp2); 2023-03-31T06:22:41.3739232Z out_ptr3[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3739539Z out_ptr4[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3739853Z out_ptr5[static_cast(i0)] = tmp3; 2023-03-31T06:22:41.3740110Z out_ptr6[static_cast(i0)] = tmp3; 2023-03-31T06:22:41.3740425Z } 2023-03-31T06:22:41.3740654Z } 2023-03-31T06:22:41.3740832Z } 2023-03-31T06:22:41.3741048Z } 2023-03-31T06:22:41.3741302Z ''') 2023-03-31T06:22:41.3741388Z 2023-03-31T06:22:41.3741575Z async_compile.wait(globals()) 2023-03-31T06:22:41.3741847Z del async_compile 2023-03-31T06:22:41.3742146Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3742467Z wrapper = ( 2023-03-31T06:22:41.3742755Z ''' 2023-03-31T06:22:41.3743001Z #include 2023-03-31T06:22:41.3743212Z #include 2023-03-31T06:22:41.3743357Z 2023-03-31T06:22:41.3743487Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3743651Z 2023-03-31T06:22:41.3743776Z template 2023-03-31T06:22:41.3744042Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3744365Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3744673Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3744981Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3745290Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3745642Z return kernel_cpp; 2023-03-31T06:22:41.3745878Z } 2023-03-31T06:22:41.3746116Z std::vector call_6(std::vector args) { 2023-03-31T06:22:41.3746407Z at::Tensor arg0_1; 2023-03-31T06:22:41.3746657Z arg0_1 = args[0]; 2023-03-31T06:22:41.3746934Z auto buf3 = at::empty_strided({8, 36}, {36, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3747269Z auto buf0 = at::as_strided(buf3, {8, 16}, {36, 1}); // alias 2023-03-31T06:22:41.3747587Z auto buf2 = at::as_strided(buf3, {8, 16}, {36, 1}, 20); // alias 2023-03-31T06:22:41.3747909Z auto buf1 = at::as_strided(buf3, {8, 4}, {36, 1}, 16); // alias 2023-03-31T06:22:41.3748243Z auto buf6 = at::empty_strided({16, 16}, {16, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3748594Z auto buf4 = at::as_strided(buf6, {8, 16}, {16, 1}); // alias 2023-03-31T06:22:41.3748914Z auto buf5 = at::as_strided(buf6, {8, 16}, {16, 1}, 128); // alias 2023-03-31T06:22:41.3749220Z auto buf9 = at::empty_strided({16, 16}, {16, 1}, at::ScalarType::Double); 2023-03-31T06:22:41.3749666Z auto buf7 = at::as_strided(buf9, {8, 16}, {16, 1}); // alias 2023-03-31T06:22:41.3750005Z auto buf8 = at::as_strided(buf9, {8, 16}, {16, 1}, 128); // alias 2023-03-31T06:22:41.3750550Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/se/cseivitrsolvv2k6opepkqo43hk2ogblbx4csyn5tali3tzbyjq7.so"); 2023-03-31T06:22:41.3751252Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf2.data_ptr()), (float*)(buf1.data_ptr()), (float*)(buf4.data_ptr()), (float*)(buf5.data_ptr()), (double*)(buf7.data_ptr()), (double*)(buf8.data_ptr())); 2023-03-31T06:22:41.3751743Z arg0_1.reset(); 2023-03-31T06:22:41.3777225Z return std::vector({buf3, buf6, buf9}); 2023-03-31T06:22:41.3777434Z } 2023-03-31T06:22:41.3777659Z ''' 2023-03-31T06:22:41.3777821Z ) 2023-03-31T06:22:41.3777902Z 2023-03-31T06:22:41.3777982Z module = load_inline( 2023-03-31T06:22:41.3778393Z name='inline_extension_cswywia7wxuaptmguqtukza6gthq2rc4ra7ohsuor5ecpcmpbf2h', 2023-03-31T06:22:41.3778704Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3778919Z functions=['call_6'], 2023-03-31T06:22:41.3779405Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3779825Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3780660Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3781118Z 2023-03-31T06:22:41.3781185Z def _wrap_func(f): 2023-03-31T06:22:41.3781350Z def g(args): 2023-03-31T06:22:41.3781513Z return f(args) 2023-03-31T06:22:41.3781667Z return g 2023-03-31T06:22:41.3782034Z call = _wrap_func(module.call_6) 2023-03-31T06:22:41.3782164Z 2023-03-31T06:22:41.3782169Z 2023-03-31T06:22:41.3782255Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3782600Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3782855Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3783215Z arg0_1 = rand_strided({8, 16}, {16, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3783488Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.3783620Z 2023-03-31T06:22:41.3783632Z 2023-03-31T06:22:41.3783698Z if __name__ == "__main__": 2023-03-31T06:22:41.3783875Z import argparse 2023-03-31T06:22:41.3784105Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3784258Z 2023-03-31T06:22:41.3784416Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3784871Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3785432Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3785943Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3786228Z args = parser.parse_args() 2023-03-31T06:22:41.3786358Z 2023-03-31T06:22:41.3786445Z if args.benchmark_kernels: 2023-03-31T06:22:41.3786750Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3786957Z else: 2023-03-31T06:22:41.3787159Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3787391Z benchmark_compiled_module() 2023-03-31T06:22:41.3787523Z 2023-03-31T06:22:41.3787592Z if p: 2023-03-31T06:22:41.3787830Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3788085Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3788346Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3788514Z 2023-03-31T06:22:41.3788778Z [2023-03-31 05:31:48,903] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 6 2023-03-31T06:22:41.3789240Z [2023-03-31 05:31:48,959] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 7 2023-03-31T06:22:41.3789826Z [2023-03-31 05:31:49,050] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: load with indirect indexing or non-contigous: i0 + 3*i1 2023-03-31T06:22:41.3790400Z [2023-03-31 05:31:49,064] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3790909Z [2023-03-31 05:31:49,064] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.3791423Z [2023-03-31 05:31:49,064] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3791936Z [2023-03-31 05:31:49,065] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.3792435Z [2023-03-31 05:31:49,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3792930Z [2023-03-31 05:31:49,067] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.3793484Z [2023-03-31 05:31:49,073] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: store with indirect indexing or non-contigous: 144*i0 + i1 + 3*i2 2023-03-31T06:22:41.3794030Z [2023-03-31 05:31:49,083] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by load 2023-03-31T06:22:41.3794546Z [2023-03-31 05:31:49,083] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.3795204Z [2023-03-31 05:31:51,748] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/q3/cq3gar4wkwzl2w744o6x344d5gnud5d2stbtbrnwzaosryi75enh.py 2023-03-31T06:22:41.3795706Z [2023-03-31 05:31:51,748] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3795883Z 2023-03-31T06:22:41.3795983Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3796180Z import torch 2023-03-31T06:22:41.3796337Z import math 2023-03-31T06:22:41.3796503Z import random 2023-03-31T06:22:41.3796663Z import os 2023-03-31T06:22:41.3796820Z import tempfile 2023-03-31T06:22:41.3797045Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3797196Z 2023-03-31T06:22:41.3797314Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3797611Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3797896Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3798054Z 2023-03-31T06:22:41.3798133Z aten = torch.ops.aten 2023-03-31T06:22:41.3798374Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3798634Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3798761Z 2023-03-31T06:22:41.3798766Z 2023-03-31T06:22:41.3798904Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3799247Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3799587Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.3799808Z const float* in_ptr1, 2023-03-31T06:22:41.3800012Z const double* in_ptr2, 2023-03-31T06:22:41.3800200Z float* out_ptr0, 2023-03-31T06:22:41.3800399Z float* out_ptr1, 2023-03-31T06:22:41.3800599Z float* out_ptr2, 2023-03-31T06:22:41.3800783Z float* out_ptr3, 2023-03-31T06:22:41.3800972Z float* out_ptr4, 2023-03-31T06:22:41.3801173Z double* out_ptr5, 2023-03-31T06:22:41.3801369Z double* out_ptr6, 2023-03-31T06:22:41.3801569Z float* out_ptr7, 2023-03-31T06:22:41.3801762Z double* out_ptr8) 2023-03-31T06:22:41.3801932Z { 2023-03-31T06:22:41.3802103Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3802285Z { 2023-03-31T06:22:41.3802463Z { 2023-03-31T06:22:41.3802649Z #pragma omp for 2023-03-31T06:22:41.3802904Z for(long i0=static_cast(0); i0(3); i0+=static_cast(1)) 2023-03-31T06:22:41.3803409Z { 2023-03-31T06:22:41.3803646Z for(long i1=static_cast(0); i1(6); i1+=static_cast(1)) 2023-03-31T06:22:41.3803888Z { 2023-03-31T06:22:41.3804168Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast((8*i1) + (48*i0))); 2023-03-31T06:22:41.3804484Z auto tmp1 = at::vec::Vectorized(static_cast(1)); 2023-03-31T06:22:41.3804745Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3805005Z auto tmp3 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3805255Z auto tmp4 = tmp0 + tmp3; 2023-03-31T06:22:41.3805492Z tmp0.store(out_ptr0 + static_cast((8*i1) + (144*i0))); 2023-03-31T06:22:41.3805619Z tmp2.store(out_ptr1 + static_cast((8*i1) + (144*i0))); 2023-03-31T06:22:41.3805737Z tmp4.store(out_ptr2 + static_cast((8*i1) + (144*i0))); 2023-03-31T06:22:41.3805801Z } 2023-03-31T06:22:41.3805902Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3806053Z for(long i1=static_cast(48); i1(48); i1+=static_cast(1)) 2023-03-31T06:22:41.3806112Z { 2023-03-31T06:22:41.3806289Z auto tmp0 = in_ptr0[static_cast(i1 + (48*i0))]; 2023-03-31T06:22:41.3806400Z auto tmp1 = static_cast(1); 2023-03-31T06:22:41.3806493Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3806595Z auto tmp3 = static_cast(2); 2023-03-31T06:22:41.3806683Z auto tmp4 = tmp0 + tmp3; 2023-03-31T06:22:41.3806798Z out_ptr0[static_cast(i1 + (144*i0))] = tmp0; 2023-03-31T06:22:41.3806915Z out_ptr1[static_cast(i1 + (144*i0))] = tmp2; 2023-03-31T06:22:41.3807019Z out_ptr2[static_cast(i1 + (144*i0))] = tmp4; 2023-03-31T06:22:41.3807086Z } 2023-03-31T06:22:41.3807152Z } 2023-03-31T06:22:41.3807217Z } 2023-03-31T06:22:41.3807275Z { 2023-03-31T06:22:41.3807396Z #pragma omp for 2023-03-31T06:22:41.3807546Z for(long i0=static_cast(0); i0(0); i0+=static_cast(1)) 2023-03-31T06:22:41.3807601Z { 2023-03-31T06:22:41.3807687Z #pragma GCC ivdep 2023-03-31T06:22:41.3807841Z for(long i1=static_cast(0); i1(6); i1+=static_cast(1)) 2023-03-31T06:22:41.3807904Z { 2023-03-31T06:22:41.3808019Z float tmp0[8*8] __attribute__ ((aligned (8))); 2023-03-31T06:22:41.3808204Z at::vec::transpose_mxn(in_ptr0 + static_cast((8*i0) + (24*i1)), static_cast(3), tmp0, 8); 2023-03-31T06:22:41.3808324Z for (long i0_inner = 0; i0_inner < 8; i0_inner++) 2023-03-31T06:22:41.3808394Z { 2023-03-31T06:22:41.3808546Z auto tmp1 = at::vec::Vectorized::loadu(tmp0 + static_cast(8*i0_inner)); 2023-03-31T06:22:41.3808694Z auto tmp2 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.3808792Z auto tmp3 = tmp1 * tmp2; 2023-03-31T06:22:41.3808936Z tmp3.store(out_ptr3 + static_cast((8*i1) + (48*i0_inner) + (384*i0))); 2023-03-31T06:22:41.3809006Z } 2023-03-31T06:22:41.3809071Z } 2023-03-31T06:22:41.3809148Z #pragma GCC ivdep 2023-03-31T06:22:41.3809291Z for(long i1=static_cast(48); i1(48); i1+=static_cast(1)) 2023-03-31T06:22:41.3809393Z { 2023-03-31T06:22:41.3809542Z for (long i0_inner = 0; i0_inner < 8; i0_inner++) 2023-03-31T06:22:41.3809607Z { 2023-03-31T06:22:41.3809741Z auto tmp0 = in_ptr0[static_cast(i0_inner + (3*i1) + (8*i0))]; 2023-03-31T06:22:41.3809846Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3809946Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.3810077Z out_ptr3[static_cast(i1 + (48*i0_inner) + (384*i0))] = tmp2; 2023-03-31T06:22:41.3810137Z } 2023-03-31T06:22:41.3810204Z } 2023-03-31T06:22:41.3810270Z } 2023-03-31T06:22:41.3810351Z #pragma omp for 2023-03-31T06:22:41.3810500Z for(long i0=static_cast(0); i0(3); i0+=static_cast(1)) 2023-03-31T06:22:41.3810565Z { 2023-03-31T06:22:41.3810649Z #pragma GCC ivdep 2023-03-31T06:22:41.3810790Z for(long i1=static_cast(0); i1(48); i1+=static_cast(1)) 2023-03-31T06:22:41.3810859Z { 2023-03-31T06:22:41.3810983Z auto tmp0 = in_ptr0[static_cast(i0 + (3*i1))]; 2023-03-31T06:22:41.3811091Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.3811185Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.3811305Z out_ptr3[static_cast(i1 + (48*i0))] = tmp2; 2023-03-31T06:22:41.3811372Z } 2023-03-31T06:22:41.3811424Z } 2023-03-31T06:22:41.3811517Z } 2023-03-31T06:22:41.3811583Z { 2023-03-31T06:22:41.3811663Z #pragma omp for 2023-03-31T06:22:41.3811808Z for(long i0=static_cast(0); i0(144); i0+=static_cast(1)) 2023-03-31T06:22:41.3811870Z { 2023-03-31T06:22:41.3811984Z auto tmp0 = out_ptr3[static_cast(i0)]; 2023-03-31T06:22:41.3812085Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.3812192Z out_ptr4[static_cast(i0)] = tmp0; 2023-03-31T06:22:41.3812295Z out_ptr5[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3812399Z out_ptr6[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3812459Z } 2023-03-31T06:22:41.3812520Z } 2023-03-31T06:22:41.3812610Z { 2023-03-31T06:22:41.3812678Z #pragma omp for 2023-03-31T06:22:41.3812826Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3812886Z { 2023-03-31T06:22:41.3812968Z #pragma GCC ivdep 2023-03-31T06:22:41.3813116Z for(long i1=static_cast(0); i1(0); i1+=static_cast(1)) 2023-03-31T06:22:41.3813181Z { 2023-03-31T06:22:41.3813263Z #pragma GCC ivdep 2023-03-31T06:22:41.3813405Z for(long i2=static_cast(0); i2(6); i2+=static_cast(1)) 2023-03-31T06:22:41.3813469Z { 2023-03-31T06:22:41.3813588Z float tmp1[8*8] __attribute__ ((aligned (8))); 2023-03-31T06:22:41.3813704Z for (long i1_inner = 0; i1_inner < 8; i1_inner++) 2023-03-31T06:22:41.3813773Z { 2023-03-31T06:22:41.3813962Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr1 + static_cast((8*i2) + (48*i1_inner) + (144*i0) + (384*i1))); 2023-03-31T06:22:41.3814088Z tmp0.store(tmp1 + static_cast(8*i1_inner)); 2023-03-31T06:22:41.3814159Z } 2023-03-31T06:22:41.3814347Z at::vec::transpose_mxn(tmp1, 8, out_ptr7 + static_cast((8*i1) + (24*i2) + (144*i0)), static_cast(3)); 2023-03-31T06:22:41.3814404Z } 2023-03-31T06:22:41.3814518Z #pragma GCC ivdep 2023-03-31T06:22:41.3814702Z for(long i2=static_cast(48); i2(48); i2+=static_cast(1)) 2023-03-31T06:22:41.3814771Z { 2023-03-31T06:22:41.3814891Z for (long i1_inner = 0; i1_inner < 8; i1_inner++) 2023-03-31T06:22:41.3814961Z { 2023-03-31T06:22:41.3815102Z auto tmp0 = in_ptr1[static_cast(i2 + (48*i1_inner) + (144*i0) + (384*i1))]; 2023-03-31T06:22:41.3815236Z out_ptr7[static_cast(i1_inner + (3*i2) + (8*i1) + (144*i0))] = tmp0; 2023-03-31T06:22:41.3815302Z } 2023-03-31T06:22:41.3815364Z } 2023-03-31T06:22:41.3815427Z } 2023-03-31T06:22:41.3815508Z #pragma GCC ivdep 2023-03-31T06:22:41.3815655Z for(long i1=static_cast(0); i1(3); i1+=static_cast(1)) 2023-03-31T06:22:41.3815718Z { 2023-03-31T06:22:41.3815794Z #pragma GCC ivdep 2023-03-31T06:22:41.3815943Z for(long i2=static_cast(0); i2(48); i2+=static_cast(1)) 2023-03-31T06:22:41.3816012Z { 2023-03-31T06:22:41.3816143Z auto tmp0 = in_ptr1[static_cast(i2 + (48*i1) + (144*i0))]; 2023-03-31T06:22:41.3816272Z out_ptr7[static_cast(i1 + (3*i2) + (144*i0))] = tmp0; 2023-03-31T06:22:41.3816345Z } 2023-03-31T06:22:41.3816411Z } 2023-03-31T06:22:41.3816507Z } 2023-03-31T06:22:41.3816560Z } 2023-03-31T06:22:41.3816623Z { 2023-03-31T06:22:41.3816723Z #pragma omp for collapse(3) 2023-03-31T06:22:41.3816874Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3816939Z { 2023-03-31T06:22:41.3817095Z for(long i1=static_cast(0); i1(3); i1+=static_cast(1)) 2023-03-31T06:22:41.3817165Z { 2023-03-31T06:22:41.3817305Z for(long i2=static_cast(0); i2(48); i2+=static_cast(1)) 2023-03-31T06:22:41.3817375Z { 2023-03-31T06:22:41.3817504Z auto tmp0 = in_ptr2[static_cast(i2 + (48*i1) + (144*i0))]; 2023-03-31T06:22:41.3817662Z out_ptr8[static_cast(i1 + (3*i2) + (144*i0))] = tmp0; 2023-03-31T06:22:41.3817728Z } 2023-03-31T06:22:41.3817792Z } 2023-03-31T06:22:41.3817855Z } 2023-03-31T06:22:41.3817908Z } 2023-03-31T06:22:41.3817968Z } 2023-03-31T06:22:41.3818025Z } 2023-03-31T06:22:41.3818130Z ''') 2023-03-31T06:22:41.3818137Z 2023-03-31T06:22:41.3818228Z async_compile.wait(globals()) 2023-03-31T06:22:41.3818299Z del async_compile 2023-03-31T06:22:41.3818419Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3818482Z wrapper = ( 2023-03-31T06:22:41.3818564Z ''' 2023-03-31T06:22:41.3818638Z #include 2023-03-31T06:22:41.3818714Z #include 2023-03-31T06:22:41.3818720Z 2023-03-31T06:22:41.3818810Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3818815Z 2023-03-31T06:22:41.3818908Z template 2023-03-31T06:22:41.3819028Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3819106Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3819226Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3819320Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3819441Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3819521Z return kernel_cpp; 2023-03-31T06:22:41.3819582Z } 2023-03-31T06:22:41.3819704Z std::vector call_7(std::vector args) { 2023-03-31T06:22:41.3819770Z at::Tensor arg0_1; 2023-03-31T06:22:41.3819840Z arg0_1 = args[0]; 2023-03-31T06:22:41.3820026Z auto buf3 = at::empty_strided({1, 3, 3, 48}, {432, 1, 144, 3}, at::ScalarType::Float); 2023-03-31T06:22:41.3820176Z auto buf0 = at::as_strided(buf3, {1, 3, 3, 16}, {432, 1, 144, 3}); // alias 2023-03-31T06:22:41.3820299Z auto buf1 = at::as_strided(buf3, {1, 3, 3, 16}, {432, 1, 144, 3}, 48); // alias 2023-03-31T06:22:41.3820420Z auto buf2 = at::as_strided(buf3, {1, 3, 3, 16}, {432, 1, 144, 3}, 96); // alias 2023-03-31T06:22:41.3820556Z auto buf6 = at::empty_strided({2, 3, 3, 16}, {144, 48, 16, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3820679Z auto buf4 = at::as_strided(buf6, {1, 3, 3, 16}, {144, 48, 16, 1}); // alias 2023-03-31T06:22:41.3820803Z auto buf5 = at::as_strided(buf6, {1, 3, 3, 16}, {144, 48, 16, 1}, 144); // alias 2023-03-31T06:22:41.3820933Z auto buf10 = at::empty_strided({2, 3, 3, 16}, {144, 48, 16, 1}, at::ScalarType::Double); 2023-03-31T06:22:41.3821057Z auto buf8 = at::as_strided(buf10, {1, 3, 3, 16}, {144, 48, 16, 1}); // alias 2023-03-31T06:22:41.3821184Z auto buf9 = at::as_strided(buf10, {1, 3, 3, 16}, {144, 48, 16, 1}, 144); // alias 2023-03-31T06:22:41.3821324Z auto buf7 = at::empty_strided({2, 3, 3, 16}, {144, 1, 48, 3}, at::ScalarType::Float); 2023-03-31T06:22:41.3821467Z auto buf11 = at::empty_strided({2, 3, 3, 16}, {144, 1, 48, 3}, at::ScalarType::Double); 2023-03-31T06:22:41.3821876Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/rs/crsey3nirashi7hwqhg2cv5r6p2shzpqxmblse4pyu56crvutqfe.so"); 2023-03-31T06:22:41.3822272Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(buf6.data_ptr()), (double*)(buf10.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf1.data_ptr()), (float*)(buf2.data_ptr()), (float*)(buf4.data_ptr()), (float*)(buf5.data_ptr()), (double*)(buf8.data_ptr()), (double*)(buf9.data_ptr()), (float*)(buf7.data_ptr()), (double*)(buf11.data_ptr())); 2023-03-31T06:22:41.3822349Z arg0_1.reset(); 2023-03-31T06:22:41.3822514Z buf10.reset(); 2023-03-31T06:22:41.3822589Z buf4.reset(); 2023-03-31T06:22:41.3822646Z buf5.reset(); 2023-03-31T06:22:41.3822710Z buf6.reset(); 2023-03-31T06:22:41.3822778Z buf8.reset(); 2023-03-31T06:22:41.3822845Z buf9.reset(); 2023-03-31T06:22:41.3822962Z return std::vector({buf3, buf7, buf11}); 2023-03-31T06:22:41.3823054Z } 2023-03-31T06:22:41.3823139Z ''' 2023-03-31T06:22:41.3823189Z ) 2023-03-31T06:22:41.3823195Z 2023-03-31T06:22:41.3823269Z module = load_inline( 2023-03-31T06:22:41.3823547Z name='inline_extension_civyqtdijcdyjwq5kpyey7z6e26ncnbsnrlcog6v7xnkulxyraut', 2023-03-31T06:22:41.3823635Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3823746Z functions=['call_7'], 2023-03-31T06:22:41.3824124Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3824275Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3824990Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3825000Z 2023-03-31T06:22:41.3825076Z def _wrap_func(f): 2023-03-31T06:22:41.3825135Z def g(args): 2023-03-31T06:22:41.3825206Z return f(args) 2023-03-31T06:22:41.3825274Z return g 2023-03-31T06:22:41.3825364Z call = _wrap_func(module.call_7) 2023-03-31T06:22:41.3825369Z 2023-03-31T06:22:41.3825374Z 2023-03-31T06:22:41.3825464Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3825577Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3825701Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3825956Z arg0_1 = rand_strided({1, 3, 3, 16}, {144, 1, 48, 3}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3826128Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.3826135Z 2023-03-31T06:22:41.3826139Z 2023-03-31T06:22:41.3826216Z if __name__ == "__main__": 2023-03-31T06:22:41.3826289Z import argparse 2023-03-31T06:22:41.3826414Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3826422Z 2023-03-31T06:22:41.3826530Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3826840Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3827170Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3827441Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3827524Z args = parser.parse_args() 2023-03-31T06:22:41.3827544Z 2023-03-31T06:22:41.3827623Z if args.benchmark_kernels: 2023-03-31T06:22:41.3827811Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3827880Z else: 2023-03-31T06:22:41.3827993Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3828088Z benchmark_compiled_module() 2023-03-31T06:22:41.3828096Z 2023-03-31T06:22:41.3828164Z if p: 2023-03-31T06:22:41.3828314Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3828427Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3828566Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3828571Z 2023-03-31T06:22:41.3828852Z [2023-03-31 05:31:51,748] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 7 2023-03-31T06:22:41.3828931Z PASSED [ 27%] 2023-03-31T06:22:41.3829321Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear1_cpu [2023-03-31 05:31:51,932] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 8 2023-03-31T06:22:41.3829729Z [2023-03-31 05:31:54,784] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/gr/cgrqquu7xlam6jmzbjnubzslu33a6qdmvrjaz66aowpq5yhogmea.py 2023-03-31T06:22:41.3829946Z [2023-03-31 05:31:54,785] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3829990Z 2023-03-31T06:22:41.3830088Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3830161Z import torch 2023-03-31T06:22:41.3830222Z import math 2023-03-31T06:22:41.3830294Z import random 2023-03-31T06:22:41.3830361Z import os 2023-03-31T06:22:41.3830435Z import tempfile 2023-03-31T06:22:41.3830552Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3830558Z 2023-03-31T06:22:41.3830678Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3830806Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3831316Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3831321Z 2023-03-31T06:22:41.3831403Z aten = torch.ops.aten 2023-03-31T06:22:41.3831610Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3831706Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3831711Z 2023-03-31T06:22:41.3831716Z 2023-03-31T06:22:41.3831866Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3832086Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3832196Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3832259Z { 2023-03-31T06:22:41.3832347Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3832411Z { 2023-03-31T06:22:41.3832476Z { 2023-03-31T06:22:41.3832557Z #pragma omp for 2023-03-31T06:22:41.3832713Z for(long i0=static_cast(0); i0(4); i0+=static_cast(1)) 2023-03-31T06:22:41.3832813Z { 2023-03-31T06:22:41.3833010Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3833139Z auto tmp1 = decltype(tmp0)(1)/(decltype(tmp0)(1) + tmp0.neg().exp()); 2023-03-31T06:22:41.3833262Z tmp1.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3833328Z } 2023-03-31T06:22:41.3833434Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3833587Z for(long i0=static_cast(32); i0(32); i0+=static_cast(1)) 2023-03-31T06:22:41.3833655Z { 2023-03-31T06:22:41.3833774Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3833996Z auto tmp1 = decltype(tmp0)(1) / (decltype(tmp0)(1) + std::exp(-tmp0)); 2023-03-31T06:22:41.3834096Z in_out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3834163Z } 2023-03-31T06:22:41.3834230Z } 2023-03-31T06:22:41.3834295Z } 2023-03-31T06:22:41.3834357Z } 2023-03-31T06:22:41.3834438Z ''') 2023-03-31T06:22:41.3834444Z 2023-03-31T06:22:41.3834536Z async_compile.wait(globals()) 2023-03-31T06:22:41.3834598Z del async_compile 2023-03-31T06:22:41.3834722Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3834794Z wrapper = ( 2023-03-31T06:22:41.3834875Z ''' 2023-03-31T06:22:41.3834956Z #include 2023-03-31T06:22:41.3835032Z #include 2023-03-31T06:22:41.3835038Z 2023-03-31T06:22:41.3835117Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3835164Z 2023-03-31T06:22:41.3835245Z template 2023-03-31T06:22:41.3835369Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3835455Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3835581Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3835678Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3835806Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3835888Z return kernel_cpp; 2023-03-31T06:22:41.3835938Z } 2023-03-31T06:22:41.3836066Z std::vector call_8(std::vector args) { 2023-03-31T06:22:41.3836185Z at::Tensor primals_1, primals_2, primals_3; 2023-03-31T06:22:41.3836267Z primals_1 = args[0]; 2023-03-31T06:22:41.3836346Z primals_2 = args[1]; 2023-03-31T06:22:41.3836451Z primals_3 = args[2]; 2023-03-31T06:22:41.3836587Z auto buf0 = at::empty_strided({2, 16}, {16, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3836733Z at::addmm_out(buf0, primals_2, primals_3, at::as_strided(primals_1, {8, 16}, {1, 8}), 1, 1); 2023-03-31T06:22:41.3836811Z primals_1.reset(); 2023-03-31T06:22:41.3836888Z primals_2.reset(); 2023-03-31T06:22:41.3836989Z auto buf1 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.3837279Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/f6/cf6iuhj5rkdeikc3mirsy6kvurifmh7wonkse6eydkdiligznncm.so"); 2023-03-31T06:22:41.3837387Z kernel_cpp_0((float*)(buf1.data_ptr())); 2023-03-31T06:22:41.3837515Z return std::vector({buf1, primals_3, buf1}); 2023-03-31T06:22:41.3837578Z } 2023-03-31T06:22:41.3837647Z ''' 2023-03-31T06:22:41.3837707Z ) 2023-03-31T06:22:41.3837713Z 2023-03-31T06:22:41.3837792Z module = load_inline( 2023-03-31T06:22:41.3838064Z name='inline_extension_coz75fdm6qsbrcn6i3rloeaniqjtfx6cjnsbe5cvi6ontfs43n7q', 2023-03-31T06:22:41.3838154Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3838267Z functions=['call_8'], 2023-03-31T06:22:41.3838652Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3838797Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3839533Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3839578Z 2023-03-31T06:22:41.3839642Z def _wrap_func(f): 2023-03-31T06:22:41.3839714Z def g(args): 2023-03-31T06:22:41.3839790Z return f(args) 2023-03-31T06:22:41.3839861Z return g 2023-03-31T06:22:41.3839954Z call = _wrap_func(module.call_8) 2023-03-31T06:22:41.3839959Z 2023-03-31T06:22:41.3839964Z 2023-03-31T06:22:41.3840059Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3840180Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3840296Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3840511Z primals_1 = rand_strided({16, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3840719Z primals_2 = rand_strided({16, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3840931Z primals_3 = rand_strided({2, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3841075Z print_performance(lambda: call([primals_1, primals_2, primals_3])) 2023-03-31T06:22:41.3841080Z 2023-03-31T06:22:41.3841085Z 2023-03-31T06:22:41.3841163Z if __name__ == "__main__": 2023-03-31T06:22:41.3841239Z import argparse 2023-03-31T06:22:41.3841370Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3841378Z 2023-03-31T06:22:41.3841475Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3841825Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3842164Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3842447Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3842543Z args = parser.parse_args() 2023-03-31T06:22:41.3842549Z 2023-03-31T06:22:41.3842639Z if args.benchmark_kernels: 2023-03-31T06:22:41.3842830Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3842896Z else: 2023-03-31T06:22:41.3842996Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3843285Z benchmark_compiled_module() 2023-03-31T06:22:41.3843366Z 2023-03-31T06:22:41.3843437Z if p: 2023-03-31T06:22:41.3843588Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3843686Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3843824Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3843829Z 2023-03-31T06:22:41.3844120Z [2023-03-31 05:31:54,785] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 8 2023-03-31T06:22:41.3844198Z PASSED [ 33%] 2023-03-31T06:22:41.3844592Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear2_cpu [2023-03-31 05:31:55,001] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 9 2023-03-31T06:22:41.3844893Z [2023-03-31 05:31:55,055] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.bool not supported by store 2023-03-31T06:22:41.3845292Z [2023-03-31 05:31:57,778] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/cw/ccw32mdw4fgjquy2jik4htw35pmd7u63kyi7k6tpyizqotkyggwp.py 2023-03-31T06:22:41.3845517Z [2023-03-31 05:31:57,778] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3845523Z 2023-03-31T06:22:41.3845621Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3845694Z import torch 2023-03-31T06:22:41.3845765Z import math 2023-03-31T06:22:41.3845838Z import random 2023-03-31T06:22:41.3845906Z import os 2023-03-31T06:22:41.3845967Z import tempfile 2023-03-31T06:22:41.3846085Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3846135Z 2023-03-31T06:22:41.3846295Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3846421Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3846553Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3846560Z 2023-03-31T06:22:41.3846641Z aten = torch.ops.aten 2023-03-31T06:22:41.3846935Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3847021Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3847040Z 2023-03-31T06:22:41.3847045Z 2023-03-31T06:22:41.3847180Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3847401Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3847508Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3847571Z { 2023-03-31T06:22:41.3847673Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3847740Z { 2023-03-31T06:22:41.3847804Z { 2023-03-31T06:22:41.3847874Z #pragma omp for 2023-03-31T06:22:41.3848029Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3848094Z { 2023-03-31T06:22:41.3848263Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3848398Z auto tmp1 = at::vec::clamp_min(tmp0, decltype(tmp0)(0)); 2023-03-31T06:22:41.3848524Z tmp1.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3848590Z } 2023-03-31T06:22:41.3848724Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3848879Z for(long i0=static_cast(16); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3848944Z { 2023-03-31T06:22:41.3849062Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3849158Z auto tmp1 = tmp0 * (tmp0>0); 2023-03-31T06:22:41.3849274Z in_out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3849342Z } 2023-03-31T06:22:41.3849405Z } 2023-03-31T06:22:41.3849456Z } 2023-03-31T06:22:41.3849517Z } 2023-03-31T06:22:41.3849601Z ''') 2023-03-31T06:22:41.3849606Z 2023-03-31T06:22:41.3849611Z 2023-03-31T06:22:41.3849751Z kernel_cpp_1 = async_compile.cpp(''' 2023-03-31T06:22:41.3849973Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3850109Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3850172Z { 2023-03-31T06:22:41.3850262Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3850326Z { 2023-03-31T06:22:41.3850391Z { 2023-03-31T06:22:41.3850472Z #pragma omp for 2023-03-31T06:22:41.3850625Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3850690Z { 2023-03-31T06:22:41.3850857Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3850979Z auto tmp1 = at::vec::clamp_min(tmp0, decltype(tmp0)(0)); 2023-03-31T06:22:41.3851102Z tmp1.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3851168Z } 2023-03-31T06:22:41.3851274Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3851430Z for(long i0=static_cast(16); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3851496Z { 2023-03-31T06:22:41.3851616Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3851712Z auto tmp1 = tmp0 * (tmp0>0); 2023-03-31T06:22:41.3851814Z in_out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3851880Z } 2023-03-31T06:22:41.3851945Z } 2023-03-31T06:22:41.3852009Z } 2023-03-31T06:22:41.3852072Z } 2023-03-31T06:22:41.3852154Z ''') 2023-03-31T06:22:41.3852192Z 2023-03-31T06:22:41.3852225Z 2023-03-31T06:22:41.3852368Z kernel_cpp_2 = async_compile.cpp(''' 2023-03-31T06:22:41.3852577Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3852682Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.3852744Z { 2023-03-31T06:22:41.3852846Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3852912Z { 2023-03-31T06:22:41.3852977Z { 2023-03-31T06:22:41.3853058Z #pragma omp for 2023-03-31T06:22:41.3853201Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3853268Z { 2023-03-31T06:22:41.3853430Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3853567Z auto tmp1 = at::vec::clamp_min(tmp0, decltype(tmp0)(0)); 2023-03-31T06:22:41.3853693Z tmp1.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3853760Z } 2023-03-31T06:22:41.3853863Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3854004Z for(long i0=static_cast(16); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3854070Z { 2023-03-31T06:22:41.3854188Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3854285Z auto tmp1 = tmp0 * (tmp0>0); 2023-03-31T06:22:41.3854400Z in_out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3854466Z } 2023-03-31T06:22:41.3854559Z } 2023-03-31T06:22:41.3854611Z } 2023-03-31T06:22:41.3854673Z } 2023-03-31T06:22:41.3854754Z ''') 2023-03-31T06:22:41.3854759Z 2023-03-31T06:22:41.3854764Z 2023-03-31T06:22:41.3854905Z kernel_cpp_3 = async_compile.cpp(''' 2023-03-31T06:22:41.3855127Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3855235Z extern "C" void kernel(float* in_out_ptr0, 2023-03-31T06:22:41.3855321Z bool* out_ptr0) 2023-03-31T06:22:41.3855383Z { 2023-03-31T06:22:41.3855470Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3855533Z { 2023-03-31T06:22:41.3855598Z { 2023-03-31T06:22:41.3855681Z #pragma omp for 2023-03-31T06:22:41.3855833Z for(long i0=static_cast(0); i0(16); i0+=static_cast(1)) 2023-03-31T06:22:41.3855938Z { 2023-03-31T06:22:41.3856058Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3856144Z auto tmp1 = tmp0 * (tmp0>0); 2023-03-31T06:22:41.3856249Z auto tmp2 = static_cast(0); 2023-03-31T06:22:41.3856341Z auto tmp3 = tmp1 <= tmp2; 2023-03-31T06:22:41.3856454Z in_out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.3856563Z out_ptr0[static_cast(i0)] = tmp3; 2023-03-31T06:22:41.3856634Z } 2023-03-31T06:22:41.3856704Z } 2023-03-31T06:22:41.3856754Z } 2023-03-31T06:22:41.3856816Z } 2023-03-31T06:22:41.3856898Z ''') 2023-03-31T06:22:41.3856903Z 2023-03-31T06:22:41.3856997Z async_compile.wait(globals()) 2023-03-31T06:22:41.3857073Z del async_compile 2023-03-31T06:22:41.3857197Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3857257Z wrapper = ( 2023-03-31T06:22:41.3857339Z ''' 2023-03-31T06:22:41.3857416Z #include 2023-03-31T06:22:41.3857493Z #include 2023-03-31T06:22:41.3857499Z 2023-03-31T06:22:41.3857592Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3857599Z 2023-03-31T06:22:41.3857693Z template 2023-03-31T06:22:41.3857819Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3857906Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3858018Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3858115Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3858296Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3858406Z return kernel_cpp; 2023-03-31T06:22:41.3858469Z } 2023-03-31T06:22:41.3858599Z std::vector call_9(std::vector args) { 2023-03-31T06:22:41.3858784Z at::Tensor primals_1, primals_2, primals_3, primals_4, primals_5, primals_6, primals_7, primals_8, primals_9; 2023-03-31T06:22:41.3858855Z primals_1 = args[0]; 2023-03-31T06:22:41.3858935Z primals_2 = args[1]; 2023-03-31T06:22:41.3859012Z primals_3 = args[2]; 2023-03-31T06:22:41.3859091Z primals_4 = args[3]; 2023-03-31T06:22:41.3859171Z primals_5 = args[4]; 2023-03-31T06:22:41.3859248Z primals_6 = args[5]; 2023-03-31T06:22:41.3859325Z primals_7 = args[6]; 2023-03-31T06:22:41.3859389Z primals_8 = args[7]; 2023-03-31T06:22:41.3859466Z primals_9 = args[8]; 2023-03-31T06:22:41.3859601Z auto buf0 = at::empty_strided({2, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3859758Z at::addmm_out(buf0, primals_2, primals_9, at::as_strided(primals_1, {8, 8}, {1, 8}), 1, 1); 2023-03-31T06:22:41.3859839Z primals_1.reset(); 2023-03-31T06:22:41.3859915Z primals_2.reset(); 2023-03-31T06:22:41.3860019Z auto buf1 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.3860292Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/kh/ckh5frtcjtisz3cjw7mavjxclf6752i3ziqvuxxsvjhpbcmdk4wd.so"); 2023-03-31T06:22:41.3860406Z kernel_cpp_0((float*)(buf1.data_ptr())); 2023-03-31T06:22:41.3860542Z auto buf2 = at::empty_strided({2, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3860722Z at::addmm_out(buf2, primals_4, buf1, at::as_strided(primals_3, {8, 8}, {1, 8}), 1, 1); 2023-03-31T06:22:41.3860802Z primals_4.reset(); 2023-03-31T06:22:41.3860903Z auto buf3 = buf2; buf2.reset(); // reuse 2023-03-31T06:22:41.3861185Z static auto kernel_cpp_1 = load_cpp_kernel("/tmp/torchinductor_jenkins/kh/ckh5frtcjtisz3cjw7mavjxclf6752i3ziqvuxxsvjhpbcmdk4wd.so"); 2023-03-31T06:22:41.3861291Z kernel_cpp_1((float*)(buf3.data_ptr())); 2023-03-31T06:22:41.3861425Z auto buf4 = at::empty_strided({2, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3861559Z at::addmm_out(buf4, primals_6, buf3, at::as_strided(primals_5, {8, 8}, {1, 8}), 1, 1); 2023-03-31T06:22:41.3861636Z primals_6.reset(); 2023-03-31T06:22:41.3861737Z auto buf5 = buf4; buf4.reset(); // reuse 2023-03-31T06:22:41.3862049Z static auto kernel_cpp_2 = load_cpp_kernel("/tmp/torchinductor_jenkins/kh/ckh5frtcjtisz3cjw7mavjxclf6752i3ziqvuxxsvjhpbcmdk4wd.so"); 2023-03-31T06:22:41.3862155Z kernel_cpp_2((float*)(buf5.data_ptr())); 2023-03-31T06:22:41.3862288Z auto buf6 = at::empty_strided({2, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3862524Z at::addmm_out(buf6, primals_8, buf5, at::as_strided(primals_7, {8, 8}, {1, 8}), 1, 1); 2023-03-31T06:22:41.3862605Z primals_8.reset(); 2023-03-31T06:22:41.3862695Z auto buf7 = buf6; buf6.reset(); // reuse 2023-03-31T06:22:41.3862828Z auto buf8 = at::empty_strided({2, 8}, {8, 1}, at::ScalarType::Bool); 2023-03-31T06:22:41.3863119Z static auto kernel_cpp_3 = load_cpp_kernel("/tmp/torchinductor_jenkins/np/cnps4srqi4bdmkr43wywmihymdnevy4b7rc5qsbkibv3mzcrqtfz.so"); 2023-03-31T06:22:41.3863253Z kernel_cpp_3((float*)(buf7.data_ptr()), (bool*)(buf8.data_ptr())); 2023-03-31T06:22:41.3863501Z return std::vector({buf7, primals_9, buf1, buf3, buf5, buf8, at::as_strided(primals_7, {8, 8}, {8, 1}), at::as_strided(primals_5, {8, 8}, {8, 1}), at::as_strided(primals_3, {8, 8}, {8, 1})}); 2023-03-31T06:22:41.3863565Z } 2023-03-31T06:22:41.3863653Z ''' 2023-03-31T06:22:41.3863716Z ) 2023-03-31T06:22:41.3863721Z 2023-03-31T06:22:41.3863800Z module = load_inline( 2023-03-31T06:22:41.3864058Z name='inline_extension_ckf6xyw636pwapjtzj7f7c4sm6e4owkz5kvm7gzhargfkytpx3rf', 2023-03-31T06:22:41.3864145Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3864315Z functions=['call_9'], 2023-03-31T06:22:41.3864728Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3864883Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3865599Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3865608Z 2023-03-31T06:22:41.3865684Z def _wrap_func(f): 2023-03-31T06:22:41.3865756Z def g(args): 2023-03-31T06:22:41.3865831Z return f(args) 2023-03-31T06:22:41.3865886Z return g 2023-03-31T06:22:41.3865979Z call = _wrap_func(module.call_9) 2023-03-31T06:22:41.3865985Z 2023-03-31T06:22:41.3865993Z 2023-03-31T06:22:41.3866091Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3866208Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3866338Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3866553Z primals_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3866761Z primals_2 = rand_strided({8, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3866957Z primals_3 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3867195Z primals_4 = rand_strided({8, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3867403Z primals_5 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3867605Z primals_6 = rand_strided({8, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3867811Z primals_7 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3868017Z primals_8 = rand_strided({8, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3868225Z primals_9 = rand_strided({2, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3868439Z print_performance(lambda: call([primals_1, primals_2, primals_3, primals_4, primals_5, primals_6, primals_7, primals_8, primals_9])) 2023-03-31T06:22:41.3868446Z 2023-03-31T06:22:41.3868481Z 2023-03-31T06:22:41.3868562Z if __name__ == "__main__": 2023-03-31T06:22:41.3868627Z import argparse 2023-03-31T06:22:41.3868759Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3868767Z 2023-03-31T06:22:41.3868879Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3869191Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3869534Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3869811Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3869905Z args = parser.parse_args() 2023-03-31T06:22:41.3869910Z 2023-03-31T06:22:41.3870000Z if args.benchmark_kernels: 2023-03-31T06:22:41.3870189Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3870247Z else: 2023-03-31T06:22:41.3870358Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3870453Z benchmark_compiled_module() 2023-03-31T06:22:41.3870459Z 2023-03-31T06:22:41.3870529Z if p: 2023-03-31T06:22:41.3870680Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3870777Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3870914Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3870920Z 2023-03-31T06:22:41.3871215Z [2023-03-31 05:31:57,779] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 9 2023-03-31T06:22:41.3871317Z PASSED [ 38%] 2023-03-31T06:22:41.3871720Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear_packed_cpu [2023-03-31 05:31:57,828] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 10 2023-03-31T06:22:41.3872133Z [2023-03-31 05:32:20,526] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zh/czhqzhwyrqbkgovfiukgwf3al3w2hilfghftjslemzlwmsgfl23i.py 2023-03-31T06:22:41.3872360Z [2023-03-31 05:32:20,526] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3872365Z 2023-03-31T06:22:41.3872462Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3872537Z import torch 2023-03-31T06:22:41.3872608Z import math 2023-03-31T06:22:41.3872679Z import random 2023-03-31T06:22:41.3872734Z import os 2023-03-31T06:22:41.3872809Z import tempfile 2023-03-31T06:22:41.3872929Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3872936Z 2023-03-31T06:22:41.3873055Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3873180Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3873312Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3873318Z 2023-03-31T06:22:41.3873398Z aten = torch.ops.aten 2023-03-31T06:22:41.3873526Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3873623Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3873628Z 2023-03-31T06:22:41.3873632Z 2023-03-31T06:22:41.3873806Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3874030Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3874137Z extern "C" void kernel(float* in_out_ptr0, 2023-03-31T06:22:41.3874231Z const float* in_ptr0) 2023-03-31T06:22:41.3874294Z { 2023-03-31T06:22:41.3874396Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3874449Z { 2023-03-31T06:22:41.3874516Z { 2023-03-31T06:22:41.3874599Z #pragma omp for 2023-03-31T06:22:41.3874752Z for(long i0=static_cast(0); i0(6); i0+=static_cast(1)) 2023-03-31T06:22:41.3874818Z { 2023-03-31T06:22:41.3874971Z for(long i1=static_cast(0); i1(3); i1+=static_cast(1)) 2023-03-31T06:22:41.3875070Z { 2023-03-31T06:22:41.3875232Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast((8*i1) + (30*i0))); 2023-03-31T06:22:41.3875396Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i1)); 2023-03-31T06:22:41.3875491Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3875623Z tmp2.store(in_out_ptr0 + static_cast((8*i1) + (30*i0))); 2023-03-31T06:22:41.3875691Z } 2023-03-31T06:22:41.3875793Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3875949Z for(long i1=static_cast(24); i1(30); i1+=static_cast(1)) 2023-03-31T06:22:41.3876017Z { 2023-03-31T06:22:41.3876134Z auto tmp0 = in_out_ptr0[static_cast(i1 + (30*i0))]; 2023-03-31T06:22:41.3876251Z auto tmp1 = in_ptr0[static_cast(i1)]; 2023-03-31T06:22:41.3876347Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3876470Z in_out_ptr0[static_cast(i1 + (30*i0))] = tmp2; 2023-03-31T06:22:41.3876540Z } 2023-03-31T06:22:41.3876606Z } 2023-03-31T06:22:41.3876674Z } 2023-03-31T06:22:41.3876725Z } 2023-03-31T06:22:41.3876787Z } 2023-03-31T06:22:41.3876871Z ''') 2023-03-31T06:22:41.3876877Z 2023-03-31T06:22:41.3876970Z async_compile.wait(globals()) 2023-03-31T06:22:41.3877044Z del async_compile 2023-03-31T06:22:41.3877244Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3877344Z wrapper = ( 2023-03-31T06:22:41.3877413Z ''' 2023-03-31T06:22:41.3877491Z #include 2023-03-31T06:22:41.3877574Z #include 2023-03-31T06:22:41.3877579Z 2023-03-31T06:22:41.3877672Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3877677Z 2023-03-31T06:22:41.3877772Z template 2023-03-31T06:22:41.3877897Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3877986Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3878100Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3878200Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3878326Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3878407Z return kernel_cpp; 2023-03-31T06:22:41.3878469Z } 2023-03-31T06:22:41.3878595Z std::vector call_10(std::vector args) { 2023-03-31T06:22:41.3878699Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.3878765Z arg0_1 = args[0]; 2023-03-31T06:22:41.3878840Z arg1_1 = args[1]; 2023-03-31T06:22:41.3878911Z arg2_1 = args[2]; 2023-03-31T06:22:41.3878983Z arg3_1 = args[3]; 2023-03-31T06:22:41.3879045Z 2023-03-31T06:22:41.3879134Z static auto op_mkl_linear = 2023-03-31T06:22:41.3879226Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3879300Z .findSchemaOrThrow( 2023-03-31T06:22:41.3879383Z "mkl::_mkl_linear", 2023-03-31T06:22:41.3879450Z "") 2023-03-31T06:22:41.3879519Z .typed< 2023-03-31T06:22:41.3879595Z at::Tensor( 2023-03-31T06:22:41.3879721Z const at::Tensor& self, 2023-03-31T06:22:41.3879830Z const at::Tensor& mkl_weight_t, 2023-03-31T06:22:41.3879927Z const at::Tensor& origin_weight_t, 2023-03-31T06:22:41.3880044Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3880156Z const int64_t prepack_batch_size)>(); 2023-03-31T06:22:41.3880224Z 2023-03-31T06:22:41.3880369Z auto buf0 = op_mkl_linear.call(arg3_1, arg2_1, arg0_1, at::Tensor(), 6); 2023-03-31T06:22:41.3880443Z arg0_1.reset(); 2023-03-31T06:22:41.3880514Z arg2_1.reset(); 2023-03-31T06:22:41.3880571Z arg3_1.reset(); 2023-03-31T06:22:41.3880674Z auto buf1 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.3880975Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/r7/cr7fmzwq4gqazcuyplnlejcnsx5ndsqdmrjthq4efegzwnnqjity.so"); 2023-03-31T06:22:41.3881145Z kernel_cpp_0((float*)(buf1.data_ptr()), (float*)(arg1_1.data_ptr())); 2023-03-31T06:22:41.3881218Z arg1_1.reset(); 2023-03-31T06:22:41.3881323Z return std::vector({buf1}); 2023-03-31T06:22:41.3881385Z } 2023-03-31T06:22:41.3881468Z ''' 2023-03-31T06:22:41.3881518Z ) 2023-03-31T06:22:41.3881523Z 2023-03-31T06:22:41.3881600Z module = load_inline( 2023-03-31T06:22:41.3881867Z name='inline_extension_coy5vyvt5cn5h365wstlaa2opoquu7s764a3c7mclr2hrn25vbez', 2023-03-31T06:22:41.3881955Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3882071Z functions=['call_10'], 2023-03-31T06:22:41.3882455Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3882611Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3883633Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3883643Z 2023-03-31T06:22:41.3883711Z def _wrap_func(f): 2023-03-31T06:22:41.3883785Z def g(args): 2023-03-31T06:22:41.3883860Z return f(args) 2023-03-31T06:22:41.3883930Z return g 2023-03-31T06:22:41.3884108Z call = _wrap_func(module.call_10) 2023-03-31T06:22:41.3884149Z 2023-03-31T06:22:41.3884153Z 2023-03-31T06:22:41.3884250Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3884370Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3884501Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3884708Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3884917Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3885130Z arg2_1 = rand_strided({1982689, 1}, {1, 0}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3885345Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3885480Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.3885486Z 2023-03-31T06:22:41.3885491Z 2023-03-31T06:22:41.3885570Z if __name__ == "__main__": 2023-03-31T06:22:41.3885651Z import argparse 2023-03-31T06:22:41.3885785Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3885791Z 2023-03-31T06:22:41.3885889Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3886202Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3886539Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3886872Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3886969Z args = parser.parse_args() 2023-03-31T06:22:41.3886974Z 2023-03-31T06:22:41.3887066Z if args.benchmark_kernels: 2023-03-31T06:22:41.3887255Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3887324Z else: 2023-03-31T06:22:41.3887440Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3887525Z benchmark_compiled_module() 2023-03-31T06:22:41.3887530Z 2023-03-31T06:22:41.3887604Z if p: 2023-03-31T06:22:41.3887754Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3887852Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3887990Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3888039Z 2023-03-31T06:22:41.3888323Z [2023-03-31 05:32:20,527] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 10 2023-03-31T06:22:41.3888596Z [2023-03-31 05:32:20,583] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 11 2023-03-31T06:22:41.3889005Z [2023-03-31 05:32:48,797] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ch/cchoerzi52vphuurzedpq7ep2rjkdcjwhgdjejbqiapcgnf23wlb.py 2023-03-31T06:22:41.3889213Z [2023-03-31 05:32:48,797] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3889234Z 2023-03-31T06:22:41.3889319Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3889391Z import torch 2023-03-31T06:22:41.3889462Z import math 2023-03-31T06:22:41.3889535Z import random 2023-03-31T06:22:41.3889603Z import os 2023-03-31T06:22:41.3889678Z import tempfile 2023-03-31T06:22:41.3889783Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3889803Z 2023-03-31T06:22:41.3889909Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3890034Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3890171Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3890177Z 2023-03-31T06:22:41.3890258Z aten = torch.ops.aten 2023-03-31T06:22:41.3890397Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3890492Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3890497Z 2023-03-31T06:22:41.3890590Z async_compile.wait(globals()) 2023-03-31T06:22:41.3890685Z del async_compile 2023-03-31T06:22:41.3890837Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3890913Z wrapper = ( 2023-03-31T06:22:41.3891000Z ''' 2023-03-31T06:22:41.3891077Z #include 2023-03-31T06:22:41.3891154Z #include 2023-03-31T06:22:41.3891159Z 2023-03-31T06:22:41.3891253Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3891258Z 2023-03-31T06:22:41.3891341Z template 2023-03-31T06:22:41.3891464Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3891549Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3891675Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3891775Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3891902Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3891981Z return kernel_cpp; 2023-03-31T06:22:41.3892030Z } 2023-03-31T06:22:41.3892161Z std::vector call_11(std::vector args) { 2023-03-31T06:22:41.3892269Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.3892344Z arg0_1 = args[0]; 2023-03-31T06:22:41.3892417Z arg1_1 = args[1]; 2023-03-31T06:22:41.3892490Z arg2_1 = args[2]; 2023-03-31T06:22:41.3892560Z arg3_1 = args[3]; 2023-03-31T06:22:41.3892610Z 2023-03-31T06:22:41.3892710Z static auto op_linear_pointwise = 2023-03-31T06:22:41.3892804Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3892895Z .findSchemaOrThrow( 2023-03-31T06:22:41.3892995Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.3893062Z "") 2023-03-31T06:22:41.3893167Z .typed< 2023-03-31T06:22:41.3893231Z at::Tensor( 2023-03-31T06:22:41.3893328Z const at::Tensor& input_t, 2023-03-31T06:22:41.3893429Z const at::Tensor& weight_t, 2023-03-31T06:22:41.3893550Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3893640Z c10::string_view attr, 2023-03-31T06:22:41.3893763Z torch::List> scalars, 2023-03-31T06:22:41.3893888Z c10::optional algorithm)>(); 2023-03-31T06:22:41.3893952Z 2023-03-31T06:22:41.3894157Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "none", {-1}, ""); 2023-03-31T06:22:41.3894231Z arg1_1.reset(); 2023-03-31T06:22:41.3894303Z arg2_1.reset(); 2023-03-31T06:22:41.3894407Z arg3_1.reset(); 2023-03-31T06:22:41.3894513Z return std::vector({buf0}); 2023-03-31T06:22:41.3894577Z } 2023-03-31T06:22:41.3894658Z ''' 2023-03-31T06:22:41.3894709Z ) 2023-03-31T06:22:41.3894715Z 2023-03-31T06:22:41.3894793Z module = load_inline( 2023-03-31T06:22:41.3895063Z name='inline_extension_catqtbhl34ib2xzczcyululidzezckswj5v5b33bqss523xgmvq2', 2023-03-31T06:22:41.3895153Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3895271Z functions=['call_11'], 2023-03-31T06:22:41.3895659Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3895817Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3896528Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3896538Z 2023-03-31T06:22:41.3896616Z def _wrap_func(f): 2023-03-31T06:22:41.3896675Z def g(args): 2023-03-31T06:22:41.3896749Z return f(args) 2023-03-31T06:22:41.3896818Z return g 2023-03-31T06:22:41.3896912Z call = _wrap_func(module.call_11) 2023-03-31T06:22:41.3896918Z 2023-03-31T06:22:41.3896923Z 2023-03-31T06:22:41.3897018Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3897170Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3897335Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3897537Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3897744Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3897953Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3898175Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3898311Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.3898317Z 2023-03-31T06:22:41.3898321Z 2023-03-31T06:22:41.3898401Z if __name__ == "__main__": 2023-03-31T06:22:41.3898478Z import argparse 2023-03-31T06:22:41.3898609Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3898614Z 2023-03-31T06:22:41.3898711Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3899025Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3899368Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3899646Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3899742Z args = parser.parse_args() 2023-03-31T06:22:41.3899747Z 2023-03-31T06:22:41.3899838Z if args.benchmark_kernels: 2023-03-31T06:22:41.3900075Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3900145Z else: 2023-03-31T06:22:41.3900258Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3900340Z benchmark_compiled_module() 2023-03-31T06:22:41.3900346Z 2023-03-31T06:22:41.3900413Z if p: 2023-03-31T06:22:41.3900567Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3900666Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3900802Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3900808Z 2023-03-31T06:22:41.3901089Z [2023-03-31 05:32:48,797] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 11 2023-03-31T06:22:41.3901394Z [2023-03-31 05:32:48,927] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 12 2023-03-31T06:22:41.3901796Z [2023-03-31 05:33:21,400] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/d4/cd4jnpbh3vmzm75nlzt37ijfeysslnzo4xbd2gnt5eusjh3udmcw.py 2023-03-31T06:22:41.3902003Z [2023-03-31 05:33:21,401] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3902021Z 2023-03-31T06:22:41.3902105Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3902177Z import torch 2023-03-31T06:22:41.3902250Z import math 2023-03-31T06:22:41.3902325Z import random 2023-03-31T06:22:41.3902481Z import os 2023-03-31T06:22:41.3902560Z import tempfile 2023-03-31T06:22:41.3902667Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3902684Z 2023-03-31T06:22:41.3902791Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3902914Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3903049Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3903055Z 2023-03-31T06:22:41.3903138Z aten = torch.ops.aten 2023-03-31T06:22:41.3903279Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3903375Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3903380Z 2023-03-31T06:22:41.3903472Z async_compile.wait(globals()) 2023-03-31T06:22:41.3903533Z del async_compile 2023-03-31T06:22:41.3903653Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3903725Z wrapper = ( 2023-03-31T06:22:41.3903810Z ''' 2023-03-31T06:22:41.3903922Z #include 2023-03-31T06:22:41.3904023Z #include 2023-03-31T06:22:41.3904029Z 2023-03-31T06:22:41.3904122Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3904127Z 2023-03-31T06:22:41.3904206Z template 2023-03-31T06:22:41.3904329Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3904414Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3904542Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3904638Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3904765Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3904845Z return kernel_cpp; 2023-03-31T06:22:41.3904895Z } 2023-03-31T06:22:41.3905023Z std::vector call_12(std::vector args) { 2023-03-31T06:22:41.3905117Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.3905190Z arg0_1 = args[0]; 2023-03-31T06:22:41.3905263Z arg1_1 = args[1]; 2023-03-31T06:22:41.3905336Z arg2_1 = args[2]; 2023-03-31T06:22:41.3905400Z 2023-03-31T06:22:41.3905476Z static auto op_mkl_linear = 2023-03-31T06:22:41.3905570Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3905658Z .findSchemaOrThrow( 2023-03-31T06:22:41.3905740Z "mkl::_mkl_linear", 2023-03-31T06:22:41.3905806Z "") 2023-03-31T06:22:41.3905875Z .typed< 2023-03-31T06:22:41.3905949Z at::Tensor( 2023-03-31T06:22:41.3906031Z const at::Tensor& self, 2023-03-31T06:22:41.3906138Z const at::Tensor& mkl_weight_t, 2023-03-31T06:22:41.3906280Z const at::Tensor& origin_weight_t, 2023-03-31T06:22:41.3906398Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3906511Z const int64_t prepack_batch_size)>(); 2023-03-31T06:22:41.3906578Z 2023-03-31T06:22:41.3906720Z auto buf0 = op_mkl_linear.call(arg2_1, arg1_1, arg0_1, at::Tensor(), 6); 2023-03-31T06:22:41.3906795Z arg0_1.reset(); 2023-03-31T06:22:41.3906856Z arg1_1.reset(); 2023-03-31T06:22:41.3906926Z arg2_1.reset(); 2023-03-31T06:22:41.3907032Z return std::vector({buf0}); 2023-03-31T06:22:41.3907097Z } 2023-03-31T06:22:41.3907179Z ''' 2023-03-31T06:22:41.3907240Z ) 2023-03-31T06:22:41.3907245Z 2023-03-31T06:22:41.3907310Z module = load_inline( 2023-03-31T06:22:41.3907571Z name='inline_extension_c5cuogvyd2kccsc246od5x6b5rx3h6ap75jrua54e4csrtipsul4', 2023-03-31T06:22:41.3907700Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3907817Z functions=['call_12'], 2023-03-31T06:22:41.3908202Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3908359Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3909088Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3909096Z 2023-03-31T06:22:41.3909172Z def _wrap_func(f): 2023-03-31T06:22:41.3909243Z def g(args): 2023-03-31T06:22:41.3909307Z return f(args) 2023-03-31T06:22:41.3909378Z return g 2023-03-31T06:22:41.3909473Z call = _wrap_func(module.call_12) 2023-03-31T06:22:41.3909478Z 2023-03-31T06:22:41.3909483Z 2023-03-31T06:22:41.3909578Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3909699Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3909826Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3910036Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3910250Z arg1_1 = rand_strided({1982689, 1}, {1, 0}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3910497Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3910652Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.3910658Z 2023-03-31T06:22:41.3910662Z 2023-03-31T06:22:41.3910740Z if __name__ == "__main__": 2023-03-31T06:22:41.3910816Z import argparse 2023-03-31T06:22:41.3910945Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3910953Z 2023-03-31T06:22:41.3911063Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3911376Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3911712Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3911986Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3912070Z args = parser.parse_args() 2023-03-31T06:22:41.3912076Z 2023-03-31T06:22:41.3912166Z if args.benchmark_kernels: 2023-03-31T06:22:41.3912354Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3912421Z else: 2023-03-31T06:22:41.3912533Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3912628Z benchmark_compiled_module() 2023-03-31T06:22:41.3912636Z 2023-03-31T06:22:41.3912704Z if p: 2023-03-31T06:22:41.3912854Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3912979Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3913119Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3913125Z 2023-03-31T06:22:41.3913404Z [2023-03-31 05:33:21,401] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 12 2023-03-31T06:22:41.3913674Z [2023-03-31 05:33:21,459] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 13 2023-03-31T06:22:41.3914079Z [2023-03-31 05:33:44,280] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/2s/c2sqdcbb366pmfyhuasi56q3qtxvtbgywl6icieu6juxtrhjmwsk.py 2023-03-31T06:22:41.3914298Z [2023-03-31 05:33:44,280] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3914304Z 2023-03-31T06:22:41.3914433Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3914505Z import torch 2023-03-31T06:22:41.3914564Z import math 2023-03-31T06:22:41.3914636Z import random 2023-03-31T06:22:41.3914704Z import os 2023-03-31T06:22:41.3914781Z import tempfile 2023-03-31T06:22:41.3914899Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3914904Z 2023-03-31T06:22:41.3915023Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3915148Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3915269Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3915289Z 2023-03-31T06:22:41.3915359Z aten = torch.ops.aten 2023-03-31T06:22:41.3915497Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3915591Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3915596Z 2023-03-31T06:22:41.3915687Z async_compile.wait(globals()) 2023-03-31T06:22:41.3915762Z del async_compile 2023-03-31T06:22:41.3915883Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3915958Z wrapper = ( 2023-03-31T06:22:41.3916027Z ''' 2023-03-31T06:22:41.3916103Z #include 2023-03-31T06:22:41.3916179Z #include 2023-03-31T06:22:41.3916187Z 2023-03-31T06:22:41.3916279Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3916284Z 2023-03-31T06:22:41.3916376Z template 2023-03-31T06:22:41.3916502Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3916588Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3916699Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3916828Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3916978Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3917059Z return kernel_cpp; 2023-03-31T06:22:41.3917121Z } 2023-03-31T06:22:41.3917248Z std::vector call_13(std::vector args) { 2023-03-31T06:22:41.3917341Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.3917405Z arg0_1 = args[0]; 2023-03-31T06:22:41.3917477Z arg1_1 = args[1]; 2023-03-31T06:22:41.3917548Z arg2_1 = args[2]; 2023-03-31T06:22:41.3917609Z 2023-03-31T06:22:41.3917711Z static auto op_linear_pointwise = 2023-03-31T06:22:41.3917804Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3917891Z .findSchemaOrThrow( 2023-03-31T06:22:41.3917977Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.3918043Z "") 2023-03-31T06:22:41.3918111Z .typed< 2023-03-31T06:22:41.3918185Z at::Tensor( 2023-03-31T06:22:41.3918284Z const at::Tensor& input_t, 2023-03-31T06:22:41.3918388Z const at::Tensor& weight_t, 2023-03-31T06:22:41.3918505Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3918582Z c10::string_view attr, 2023-03-31T06:22:41.3918702Z torch::List> scalars, 2023-03-31T06:22:41.3918823Z c10::optional algorithm)>(); 2023-03-31T06:22:41.3918892Z 2023-03-31T06:22:41.3919125Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "none", {-1}, ""); 2023-03-31T06:22:41.3919230Z arg1_1.reset(); 2023-03-31T06:22:41.3919302Z arg2_1.reset(); 2023-03-31T06:22:41.3919408Z return std::vector({buf0}); 2023-03-31T06:22:41.3919458Z } 2023-03-31T06:22:41.3919539Z ''' 2023-03-31T06:22:41.3919602Z ) 2023-03-31T06:22:41.3919607Z 2023-03-31T06:22:41.3919685Z module = load_inline( 2023-03-31T06:22:41.3919961Z name='inline_extension_cwhsedfmuud56s4esovm4o44grk5askronrezwlsef3kubre7klv', 2023-03-31T06:22:41.3920049Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3920165Z functions=['call_13'], 2023-03-31T06:22:41.3920531Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3920687Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3921423Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3921431Z 2023-03-31T06:22:41.3921509Z def _wrap_func(f): 2023-03-31T06:22:41.3921584Z def g(args): 2023-03-31T06:22:41.3921658Z return f(args) 2023-03-31T06:22:41.3921729Z return g 2023-03-31T06:22:41.3921825Z call = _wrap_func(module.call_13) 2023-03-31T06:22:41.3921830Z 2023-03-31T06:22:41.3921835Z 2023-03-31T06:22:41.3921931Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3922036Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3922163Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3922376Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3922591Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3922810Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3922938Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.3922943Z 2023-03-31T06:22:41.3922948Z 2023-03-31T06:22:41.3923185Z if __name__ == "__main__": 2023-03-31T06:22:41.3923285Z import argparse 2023-03-31T06:22:41.3923485Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3923525Z 2023-03-31T06:22:41.3923635Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3923952Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3924288Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3924569Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3924664Z args = parser.parse_args() 2023-03-31T06:22:41.3924669Z 2023-03-31T06:22:41.3924760Z if args.benchmark_kernels: 2023-03-31T06:22:41.3924950Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3925005Z else: 2023-03-31T06:22:41.3925119Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3925213Z benchmark_compiled_module() 2023-03-31T06:22:41.3925220Z 2023-03-31T06:22:41.3925296Z if p: 2023-03-31T06:22:41.3925447Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3925544Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3925682Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3925688Z 2023-03-31T06:22:41.3925969Z [2023-03-31 05:33:44,281] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 13 2023-03-31T06:22:41.3926275Z [2023-03-31 05:33:44,325] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 14 2023-03-31T06:22:41.3926668Z [2023-03-31 05:34:08,712] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/k5/ck5c3vjvs5xgvp44amxx7js5jqmawdceb26hfx5ut4ocszhyzn5e.py 2023-03-31T06:22:41.3926889Z [2023-03-31 05:34:08,712] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3926895Z 2023-03-31T06:22:41.3926994Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3927069Z import torch 2023-03-31T06:22:41.3927142Z import math 2023-03-31T06:22:41.3927215Z import random 2023-03-31T06:22:41.3927282Z import os 2023-03-31T06:22:41.3927345Z import tempfile 2023-03-31T06:22:41.3927461Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3927467Z 2023-03-31T06:22:41.3927589Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3927790Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3927925Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3927933Z 2023-03-31T06:22:41.3928015Z aten = torch.ops.aten 2023-03-31T06:22:41.3928154Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3928250Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3928255Z 2023-03-31T06:22:41.3928260Z 2023-03-31T06:22:41.3928388Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3928614Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3928723Z extern "C" void kernel(float* in_out_ptr0, 2023-03-31T06:22:41.3928819Z const float* in_ptr0) 2023-03-31T06:22:41.3928883Z { 2023-03-31T06:22:41.3928985Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3929049Z { 2023-03-31T06:22:41.3929102Z { 2023-03-31T06:22:41.3929184Z #pragma omp for 2023-03-31T06:22:41.3929341Z for(long i0=static_cast(0); i0(2); i0+=static_cast(1)) 2023-03-31T06:22:41.3929410Z { 2023-03-31T06:22:41.3929565Z for(long i1=static_cast(0); i1(3); i1+=static_cast(1)) 2023-03-31T06:22:41.3929633Z { 2023-03-31T06:22:41.3929809Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast((8*i1) + (30*i0))); 2023-03-31T06:22:41.3930000Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i1)); 2023-03-31T06:22:41.3930107Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3930240Z tmp2.store(in_out_ptr0 + static_cast((8*i1) + (30*i0))); 2023-03-31T06:22:41.3930308Z } 2023-03-31T06:22:41.3930408Z #pragma omp simd simdlen(4) 2023-03-31T06:22:41.3930562Z for(long i1=static_cast(24); i1(30); i1+=static_cast(1)) 2023-03-31T06:22:41.3930634Z { 2023-03-31T06:22:41.3930764Z auto tmp0 = in_out_ptr0[static_cast(i1 + (30*i0))]; 2023-03-31T06:22:41.3930884Z auto tmp1 = in_ptr0[static_cast(i1)]; 2023-03-31T06:22:41.3930966Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3931089Z in_out_ptr0[static_cast(i1 + (30*i0))] = tmp2; 2023-03-31T06:22:41.3931156Z } 2023-03-31T06:22:41.3931224Z } 2023-03-31T06:22:41.3931291Z } 2023-03-31T06:22:41.3931357Z } 2023-03-31T06:22:41.3931407Z } 2023-03-31T06:22:41.3931489Z ''') 2023-03-31T06:22:41.3931495Z 2023-03-31T06:22:41.3931589Z async_compile.wait(globals()) 2023-03-31T06:22:41.3931664Z del async_compile 2023-03-31T06:22:41.3931788Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3931862Z wrapper = ( 2023-03-31T06:22:41.3931942Z ''' 2023-03-31T06:22:41.3932007Z #include 2023-03-31T06:22:41.3932084Z #include 2023-03-31T06:22:41.3932090Z 2023-03-31T06:22:41.3932184Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3932189Z 2023-03-31T06:22:41.3932325Z template 2023-03-31T06:22:41.3932453Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3932539Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3932663Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3932761Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3932876Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3932958Z return kernel_cpp; 2023-03-31T06:22:41.3933021Z } 2023-03-31T06:22:41.3933149Z std::vector call_14(std::vector args) { 2023-03-31T06:22:41.3933252Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.3933326Z arg0_1 = args[0]; 2023-03-31T06:22:41.3933399Z arg1_1 = args[1]; 2023-03-31T06:22:41.3933457Z arg2_1 = args[2]; 2023-03-31T06:22:41.3933559Z arg3_1 = args[3]; 2023-03-31T06:22:41.3933621Z 2023-03-31T06:22:41.3933714Z static auto op_mkl_linear = 2023-03-31T06:22:41.3933814Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3933902Z .findSchemaOrThrow( 2023-03-31T06:22:41.3933984Z "mkl::_mkl_linear", 2023-03-31T06:22:41.3934037Z "") 2023-03-31T06:22:41.3934106Z .typed< 2023-03-31T06:22:41.3934184Z at::Tensor( 2023-03-31T06:22:41.3934278Z const at::Tensor& self, 2023-03-31T06:22:41.3934386Z const at::Tensor& mkl_weight_t, 2023-03-31T06:22:41.3934497Z const at::Tensor& origin_weight_t, 2023-03-31T06:22:41.3934616Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3934714Z const int64_t prepack_batch_size)>(); 2023-03-31T06:22:41.3934781Z 2023-03-31T06:22:41.3934922Z auto buf0 = op_mkl_linear.call(arg3_1, arg2_1, arg0_1, at::Tensor(), 2); 2023-03-31T06:22:41.3935000Z arg0_1.reset(); 2023-03-31T06:22:41.3935073Z arg2_1.reset(); 2023-03-31T06:22:41.3935142Z arg3_1.reset(); 2023-03-31T06:22:41.3935250Z auto buf1 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.3935523Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/ij/cijeceh24qzjyl3372gjxjgrw2y5c72urlmcf6ttk5vlxyj4qxsw.so"); 2023-03-31T06:22:41.3935662Z kernel_cpp_0((float*)(buf1.data_ptr()), (float*)(arg1_1.data_ptr())); 2023-03-31T06:22:41.3935734Z arg1_1.reset(); 2023-03-31T06:22:41.3935870Z return std::vector({buf1}); 2023-03-31T06:22:41.3935963Z } 2023-03-31T06:22:41.3936046Z ''' 2023-03-31T06:22:41.3936108Z ) 2023-03-31T06:22:41.3936113Z 2023-03-31T06:22:41.3936178Z module = load_inline( 2023-03-31T06:22:41.3936443Z name='inline_extension_ceaxsbyz2m3hdtof445x4fxc3jku72ry4hsa4tqvnt3jjew75ttb', 2023-03-31T06:22:41.3936532Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3936651Z functions=['call_14'], 2023-03-31T06:22:41.3937036Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3937192Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3937911Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3937919Z 2023-03-31T06:22:41.3937995Z def _wrap_func(f): 2023-03-31T06:22:41.3938066Z def g(args): 2023-03-31T06:22:41.3938129Z return f(args) 2023-03-31T06:22:41.3938196Z return g 2023-03-31T06:22:41.3938290Z call = _wrap_func(module.call_14) 2023-03-31T06:22:41.3938296Z 2023-03-31T06:22:41.3938303Z 2023-03-31T06:22:41.3938398Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3938516Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3938671Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3938885Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3939088Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3939286Z arg2_1 = rand_strided({1982689, 1}, {1, 0}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3939493Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3939629Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.3939635Z 2023-03-31T06:22:41.3939640Z 2023-03-31T06:22:41.3939720Z if __name__ == "__main__": 2023-03-31T06:22:41.3939798Z import argparse 2023-03-31T06:22:41.3939929Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3939961Z 2023-03-31T06:22:41.3940074Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3940387Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3940724Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3940988Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3941083Z args = parser.parse_args() 2023-03-31T06:22:41.3941090Z 2023-03-31T06:22:41.3941179Z if args.benchmark_kernels: 2023-03-31T06:22:41.3941370Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3941438Z else: 2023-03-31T06:22:41.3941550Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3941644Z benchmark_compiled_module() 2023-03-31T06:22:41.3941650Z 2023-03-31T06:22:41.3941709Z if p: 2023-03-31T06:22:41.3941860Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3941958Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3942096Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3942102Z 2023-03-31T06:22:41.3942469Z [2023-03-31 05:34:08,713] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 14 2023-03-31T06:22:41.3942777Z [2023-03-31 05:34:08,764] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 15 2023-03-31T06:22:41.3943219Z [2023-03-31 05:34:31,923] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/2t/c2tr5hmbefbveyiwxjkcqrmsubwkexvncsa44zup5rgzihpilu2i.py 2023-03-31T06:22:41.3943439Z [2023-03-31 05:34:31,923] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3943445Z 2023-03-31T06:22:41.3943545Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3943608Z import torch 2023-03-31T06:22:41.3943680Z import math 2023-03-31T06:22:41.3943753Z import random 2023-03-31T06:22:41.3943823Z import os 2023-03-31T06:22:41.3943899Z import tempfile 2023-03-31T06:22:41.3944019Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3944024Z 2023-03-31T06:22:41.3944144Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3944256Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3944388Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3944394Z 2023-03-31T06:22:41.3944476Z aten = torch.ops.aten 2023-03-31T06:22:41.3944617Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3944712Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3944717Z 2023-03-31T06:22:41.3944808Z async_compile.wait(globals()) 2023-03-31T06:22:41.3944882Z del async_compile 2023-03-31T06:22:41.3945004Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3945068Z wrapper = ( 2023-03-31T06:22:41.3945151Z ''' 2023-03-31T06:22:41.3945226Z #include 2023-03-31T06:22:41.3945304Z #include 2023-03-31T06:22:41.3945309Z 2023-03-31T06:22:41.3945432Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3945438Z 2023-03-31T06:22:41.3945532Z template 2023-03-31T06:22:41.3945653Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3945727Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3945852Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3945950Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3946077Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3946157Z return kernel_cpp; 2023-03-31T06:22:41.3946219Z } 2023-03-31T06:22:41.3946348Z std::vector call_15(std::vector args) { 2023-03-31T06:22:41.3946438Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.3946512Z arg0_1 = args[0]; 2023-03-31T06:22:41.3946612Z arg1_1 = args[1]; 2023-03-31T06:22:41.3946685Z arg2_1 = args[2]; 2023-03-31T06:22:41.3946756Z arg3_1 = args[3]; 2023-03-31T06:22:41.3946819Z 2023-03-31T06:22:41.3946921Z static auto op_linear_pointwise = 2023-03-31T06:22:41.3947003Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3947090Z .findSchemaOrThrow( 2023-03-31T06:22:41.3947189Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.3947256Z "") 2023-03-31T06:22:41.3947324Z .typed< 2023-03-31T06:22:41.3947399Z at::Tensor( 2023-03-31T06:22:41.3947499Z const at::Tensor& input_t, 2023-03-31T06:22:41.3947590Z const at::Tensor& weight_t, 2023-03-31T06:22:41.3947707Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3947796Z c10::string_view attr, 2023-03-31T06:22:41.3947919Z torch::List> scalars, 2023-03-31T06:22:41.3948044Z c10::optional algorithm)>(); 2023-03-31T06:22:41.3948109Z 2023-03-31T06:22:41.3948333Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "none", {-1}, ""); 2023-03-31T06:22:41.3948406Z arg1_1.reset(); 2023-03-31T06:22:41.3948464Z arg2_1.reset(); 2023-03-31T06:22:41.3948534Z arg3_1.reset(); 2023-03-31T06:22:41.3948639Z return std::vector({buf0}); 2023-03-31T06:22:41.3948703Z } 2023-03-31T06:22:41.3948783Z ''' 2023-03-31T06:22:41.3948845Z ) 2023-03-31T06:22:41.3948850Z 2023-03-31T06:22:41.3948949Z module = load_inline( 2023-03-31T06:22:41.3949246Z name='inline_extension_catqtbhl34ib2xzczcyululidzezckswj5v5b33bqss523xgmvq2', 2023-03-31T06:22:41.3949334Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3949451Z functions=['call_15'], 2023-03-31T06:22:41.3949833Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3949991Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3950696Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3950703Z 2023-03-31T06:22:41.3950778Z def _wrap_func(f): 2023-03-31T06:22:41.3950850Z def g(args): 2023-03-31T06:22:41.3950915Z return f(args) 2023-03-31T06:22:41.3950985Z return g 2023-03-31T06:22:41.3951078Z call = _wrap_func(module.call_15) 2023-03-31T06:22:41.3951084Z 2023-03-31T06:22:41.3951089Z 2023-03-31T06:22:41.3951184Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3951303Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3951431Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3951648Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3951888Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3952084Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3952291Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3952423Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.3952429Z 2023-03-31T06:22:41.3952436Z 2023-03-31T06:22:41.3952516Z if __name__ == "__main__": 2023-03-31T06:22:41.3952593Z import argparse 2023-03-31T06:22:41.3952726Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3952732Z 2023-03-31T06:22:41.3952843Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3953154Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3953528Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3953795Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3953887Z args = parser.parse_args() 2023-03-31T06:22:41.3953892Z 2023-03-31T06:22:41.3953981Z if args.benchmark_kernels: 2023-03-31T06:22:41.3954174Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3954244Z else: 2023-03-31T06:22:41.3954360Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3954455Z benchmark_compiled_module() 2023-03-31T06:22:41.3954460Z 2023-03-31T06:22:41.3954516Z if p: 2023-03-31T06:22:41.3954665Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3954761Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3954901Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3954907Z 2023-03-31T06:22:41.3955191Z [2023-03-31 05:34:31,923] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 15 2023-03-31T06:22:41.3955460Z [2023-03-31 05:34:31,967] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 16 2023-03-31T06:22:41.3955919Z [2023-03-31 05:34:57,963] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/rv/crvg57fknfkf2gjtqiucgpaapgdbj33pephoqsuf6ytxk7eojdo5.py 2023-03-31T06:22:41.3956170Z [2023-03-31 05:34:57,964] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3956176Z 2023-03-31T06:22:41.3956274Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3956334Z import torch 2023-03-31T06:22:41.3956405Z import math 2023-03-31T06:22:41.3956477Z import random 2023-03-31T06:22:41.3956545Z import os 2023-03-31T06:22:41.3956620Z import tempfile 2023-03-31T06:22:41.3956740Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3956746Z 2023-03-31T06:22:41.3956868Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3956983Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3957116Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3957122Z 2023-03-31T06:22:41.3957202Z aten = torch.ops.aten 2023-03-31T06:22:41.3957343Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3957438Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3957446Z 2023-03-31T06:22:41.3957539Z async_compile.wait(globals()) 2023-03-31T06:22:41.3957614Z del async_compile 2023-03-31T06:22:41.3957735Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3957794Z wrapper = ( 2023-03-31T06:22:41.3957877Z ''' 2023-03-31T06:22:41.3957953Z #include 2023-03-31T06:22:41.3958030Z #include 2023-03-31T06:22:41.3958036Z 2023-03-31T06:22:41.3958129Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3958134Z 2023-03-31T06:22:41.3958226Z template 2023-03-31T06:22:41.3958383Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3958458Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3958583Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3958681Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3958805Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3958885Z return kernel_cpp; 2023-03-31T06:22:41.3958951Z } 2023-03-31T06:22:41.3959084Z std::vector call_16(std::vector args) { 2023-03-31T06:22:41.3959164Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.3959237Z arg0_1 = args[0]; 2023-03-31T06:22:41.3959309Z arg1_1 = args[1]; 2023-03-31T06:22:41.3959381Z arg2_1 = args[2]; 2023-03-31T06:22:41.3959443Z 2023-03-31T06:22:41.3959531Z static auto op_mkl_linear = 2023-03-31T06:22:41.3959658Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3959732Z .findSchemaOrThrow( 2023-03-31T06:22:41.3959814Z "mkl::_mkl_linear", 2023-03-31T06:22:41.3959883Z "") 2023-03-31T06:22:41.3959953Z .typed< 2023-03-31T06:22:41.3960028Z at::Tensor( 2023-03-31T06:22:41.3960119Z const at::Tensor& self, 2023-03-31T06:22:41.3960223Z const at::Tensor& mkl_weight_t, 2023-03-31T06:22:41.3960317Z const at::Tensor& origin_weight_t, 2023-03-31T06:22:41.3960434Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3960549Z const int64_t prepack_batch_size)>(); 2023-03-31T06:22:41.3960618Z 2023-03-31T06:22:41.3960759Z auto buf0 = op_mkl_linear.call(arg2_1, arg1_1, arg0_1, at::Tensor(), 2); 2023-03-31T06:22:41.3960832Z arg0_1.reset(); 2023-03-31T06:22:41.3960904Z arg1_1.reset(); 2023-03-31T06:22:41.3960960Z arg2_1.reset(); 2023-03-31T06:22:41.3961068Z return std::vector({buf0}); 2023-03-31T06:22:41.3961130Z } 2023-03-31T06:22:41.3961217Z ''' 2023-03-31T06:22:41.3961279Z ) 2023-03-31T06:22:41.3961284Z 2023-03-31T06:22:41.3961363Z module = load_inline( 2023-03-31T06:22:41.3961634Z name='inline_extension_cgep5rbi652fncgjxw6ojbr2naanxkk6rmyltfgth7j6uqtbrgmx', 2023-03-31T06:22:41.3961708Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3961829Z functions=['call_16'], 2023-03-31T06:22:41.3962246Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3962435Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3963377Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3963393Z 2023-03-31T06:22:41.3963474Z def _wrap_func(f): 2023-03-31T06:22:41.3963550Z def g(args): 2023-03-31T06:22:41.3963627Z return f(args) 2023-03-31T06:22:41.3963696Z return g 2023-03-31T06:22:41.3963778Z call = _wrap_func(module.call_16) 2023-03-31T06:22:41.3963784Z 2023-03-31T06:22:41.3963789Z 2023-03-31T06:22:41.3963884Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3964002Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3964133Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3964351Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3964565Z arg1_1 = rand_strided({1982689, 1}, {1, 0}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3964771Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3964898Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.3964906Z 2023-03-31T06:22:41.3964910Z 2023-03-31T06:22:41.3964975Z if __name__ == "__main__": 2023-03-31T06:22:41.3965129Z import argparse 2023-03-31T06:22:41.3965262Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3965268Z 2023-03-31T06:22:41.3965377Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3965689Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3966031Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3966310Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3966405Z args = parser.parse_args() 2023-03-31T06:22:41.3966410Z 2023-03-31T06:22:41.3966500Z if args.benchmark_kernels: 2023-03-31T06:22:41.3966719Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3966788Z else: 2023-03-31T06:22:41.3966904Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3967000Z benchmark_compiled_module() 2023-03-31T06:22:41.3967006Z 2023-03-31T06:22:41.3967075Z if p: 2023-03-31T06:22:41.3967224Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3967321Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3967448Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3967470Z 2023-03-31T06:22:41.3967737Z [2023-03-31 05:34:57,964] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 16 2023-03-31T06:22:41.3968003Z [2023-03-31 05:34:58,031] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 17 2023-03-31T06:22:41.3968412Z [2023-03-31 05:35:25,479] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/qv/cqvbdoxybsozjjeo5qphwnieueyksnkvd3cx6bknybyfsk7gqdpj.py 2023-03-31T06:22:41.3968636Z [2023-03-31 05:35:25,479] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3968642Z 2023-03-31T06:22:41.3968738Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3968811Z import torch 2023-03-31T06:22:41.3968882Z import math 2023-03-31T06:22:41.3968955Z import random 2023-03-31T06:22:41.3969009Z import os 2023-03-31T06:22:41.3969085Z import tempfile 2023-03-31T06:22:41.3969248Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3969295Z 2023-03-31T06:22:41.3969419Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3969546Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3969681Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3969687Z 2023-03-31T06:22:41.3969768Z aten = torch.ops.aten 2023-03-31T06:22:41.3969897Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3969994Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3969999Z 2023-03-31T06:22:41.3970091Z async_compile.wait(globals()) 2023-03-31T06:22:41.3970172Z del async_compile 2023-03-31T06:22:41.3970294Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3970368Z wrapper = ( 2023-03-31T06:22:41.3970453Z ''' 2023-03-31T06:22:41.3970515Z #include 2023-03-31T06:22:41.3970592Z #include 2023-03-31T06:22:41.3970598Z 2023-03-31T06:22:41.3970692Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3970700Z 2023-03-31T06:22:41.3970795Z template 2023-03-31T06:22:41.3970920Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3971008Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3971135Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3971232Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3971344Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3971425Z return kernel_cpp; 2023-03-31T06:22:41.3971493Z } 2023-03-31T06:22:41.3971655Z std::vector call_17(std::vector args) { 2023-03-31T06:22:41.3971750Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.3971825Z arg0_1 = args[0]; 2023-03-31T06:22:41.3971899Z arg1_1 = args[1]; 2023-03-31T06:22:41.3971957Z arg2_1 = args[2]; 2023-03-31T06:22:41.3972020Z 2023-03-31T06:22:41.3972121Z static auto op_linear_pointwise = 2023-03-31T06:22:41.3972215Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.3972306Z .findSchemaOrThrow( 2023-03-31T06:22:41.3972407Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.3972474Z "") 2023-03-31T06:22:41.3972529Z .typed< 2023-03-31T06:22:41.3972604Z at::Tensor( 2023-03-31T06:22:41.3972703Z const at::Tensor& input_t, 2023-03-31T06:22:41.3972807Z const at::Tensor& weight_t, 2023-03-31T06:22:41.3972954Z const c10::optional& bias_opt, 2023-03-31T06:22:41.3973043Z c10::string_view attr, 2023-03-31T06:22:41.3973164Z torch::List> scalars, 2023-03-31T06:22:41.3973274Z c10::optional algorithm)>(); 2023-03-31T06:22:41.3973341Z 2023-03-31T06:22:41.3973576Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "none", {-1}, ""); 2023-03-31T06:22:41.3973650Z arg1_1.reset(); 2023-03-31T06:22:41.3973721Z arg2_1.reset(); 2023-03-31T06:22:41.3973829Z return std::vector({buf0}); 2023-03-31T06:22:41.3973895Z } 2023-03-31T06:22:41.3973963Z ''' 2023-03-31T06:22:41.3974024Z ) 2023-03-31T06:22:41.3974030Z 2023-03-31T06:22:41.3974108Z module = load_inline( 2023-03-31T06:22:41.3974377Z name='inline_extension_cwhsedfmuud56s4esovm4o44grk5askronrezwlsef3kubre7klv', 2023-03-31T06:22:41.3974464Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3974582Z functions=['call_17'], 2023-03-31T06:22:41.3974968Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3975125Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3975856Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3975899Z 2023-03-31T06:22:41.3975964Z def _wrap_func(f): 2023-03-31T06:22:41.3976035Z def g(args): 2023-03-31T06:22:41.3976109Z return f(args) 2023-03-31T06:22:41.3976177Z return g 2023-03-31T06:22:41.3976272Z call = _wrap_func(module.call_17) 2023-03-31T06:22:41.3976278Z 2023-03-31T06:22:41.3976285Z 2023-03-31T06:22:41.3976380Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3976499Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3976616Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3976832Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3977043Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3977249Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.3977381Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.3977387Z 2023-03-31T06:22:41.3977392Z 2023-03-31T06:22:41.3977472Z if __name__ == "__main__": 2023-03-31T06:22:41.3977549Z import argparse 2023-03-31T06:22:41.3977680Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3977685Z 2023-03-31T06:22:41.3977794Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3978093Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3978517Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3978799Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3978893Z args = parser.parse_args() 2023-03-31T06:22:41.3978898Z 2023-03-31T06:22:41.3978989Z if args.benchmark_kernels: 2023-03-31T06:22:41.3979181Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3979250Z else: 2023-03-31T06:22:41.3979365Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3979448Z benchmark_compiled_module() 2023-03-31T06:22:41.3979453Z 2023-03-31T06:22:41.3979522Z if p: 2023-03-31T06:22:41.3979673Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3979802Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3979940Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3979948Z 2023-03-31T06:22:41.3980231Z [2023-03-31 05:35:25,479] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 17 2023-03-31T06:22:41.3980504Z [2023-03-31 05:35:25,528] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 18 2023-03-31T06:22:41.3980920Z [2023-03-31 05:35:59,349] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/mx/cmx7kgiuemryptbornkopg4uoqqlfestvhebrnwhwun5xdy52up6.py 2023-03-31T06:22:41.3981142Z [2023-03-31 05:35:59,349] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3981148Z 2023-03-31T06:22:41.3981233Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3981306Z import torch 2023-03-31T06:22:41.3981376Z import math 2023-03-31T06:22:41.3981451Z import random 2023-03-31T06:22:41.3981519Z import os 2023-03-31T06:22:41.3981593Z import tempfile 2023-03-31T06:22:41.3981712Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3981720Z 2023-03-31T06:22:41.3981825Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3981949Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3982080Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3982086Z 2023-03-31T06:22:41.3982166Z aten = torch.ops.aten 2023-03-31T06:22:41.3982334Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3982547Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3982553Z 2023-03-31T06:22:41.3982558Z 2023-03-31T06:22:41.3982703Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.3982924Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.3983016Z extern "C" void kernel(float* in_out_ptr0, 2023-03-31T06:22:41.3983114Z const float* in_ptr0) 2023-03-31T06:22:41.3983177Z { 2023-03-31T06:22:41.3983279Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.3983345Z { 2023-03-31T06:22:41.3983411Z { 2023-03-31T06:22:41.3983493Z #pragma omp for 2023-03-31T06:22:41.3983636Z for(long i0=static_cast(0); i0(3); i0+=static_cast(1)) 2023-03-31T06:22:41.3983702Z { 2023-03-31T06:22:41.3983868Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3984030Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3984125Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3984248Z tmp2.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.3984315Z } 2023-03-31T06:22:41.3984418Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.3984560Z for(long i0=static_cast(24); i0(30); i0+=static_cast(1)) 2023-03-31T06:22:41.3984626Z { 2023-03-31T06:22:41.3984780Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3984898Z auto tmp1 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.3984988Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.3985102Z in_out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.3985167Z } 2023-03-31T06:22:41.3985220Z } 2023-03-31T06:22:41.3985286Z } 2023-03-31T06:22:41.3985350Z } 2023-03-31T06:22:41.3985434Z ''') 2023-03-31T06:22:41.3985440Z 2023-03-31T06:22:41.3985533Z async_compile.wait(globals()) 2023-03-31T06:22:41.3985608Z del async_compile 2023-03-31T06:22:41.3985732Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3985793Z wrapper = ( 2023-03-31T06:22:41.3985874Z ''' 2023-03-31T06:22:41.3985950Z #include 2023-03-31T06:22:41.3986082Z #include 2023-03-31T06:22:41.3986088Z 2023-03-31T06:22:41.3986181Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3986186Z 2023-03-31T06:22:41.3986283Z template 2023-03-31T06:22:41.3986407Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3986479Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3986604Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3986702Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3986827Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3986909Z return kernel_cpp; 2023-03-31T06:22:41.3986972Z } 2023-03-31T06:22:41.3987098Z std::vector call_18(std::vector args) { 2023-03-31T06:22:41.3987179Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.3987252Z arg0_1 = args[0]; 2023-03-31T06:22:41.3987324Z arg1_1 = args[1]; 2023-03-31T06:22:41.3987395Z arg2_1 = args[2]; 2023-03-31T06:22:41.3987533Z auto buf0 = at::empty_strided({1, 30}, {30, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.3987689Z at::mm_out(buf0, at::as_strided(arg2_1, {1, 10}, {10, 1}), at::as_strided(arg0_1, {10, 30}, {1, 10})); 2023-03-31T06:22:41.3987762Z arg0_1.reset(); 2023-03-31T06:22:41.3987821Z arg2_1.reset(); 2023-03-31T06:22:41.3987953Z auto buf1 = at::as_strided(buf0, {30, }, {1, }); buf0.reset(); // reuse 2023-03-31T06:22:41.3988280Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/xi/cxibs3plvqvfzaqkb4hbpt74pv5pmkgisj6biuoxw76scamxcdxb.so"); 2023-03-31T06:22:41.3988444Z kernel_cpp_0((float*)(buf1.data_ptr()), (float*)(arg1_1.data_ptr())); 2023-03-31T06:22:41.3988519Z arg1_1.reset(); 2023-03-31T06:22:41.3988625Z return std::vector({buf1}); 2023-03-31T06:22:41.3988687Z } 2023-03-31T06:22:41.3988769Z ''' 2023-03-31T06:22:41.3988818Z ) 2023-03-31T06:22:41.3988823Z 2023-03-31T06:22:41.3988899Z module = load_inline( 2023-03-31T06:22:41.3989170Z name='inline_extension_cfidwn352wc2yipgz5bi55halm5vrbduyrap4frk3d3jsknkgcfg', 2023-03-31T06:22:41.3989258Z cpp_sources=[wrapper], 2023-03-31T06:22:41.3989375Z functions=['call_18'], 2023-03-31T06:22:41.3989756Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.3989912Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.3990631Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.3990641Z 2023-03-31T06:22:41.3990718Z def _wrap_func(f): 2023-03-31T06:22:41.3990776Z def g(args): 2023-03-31T06:22:41.3990854Z return f(args) 2023-03-31T06:22:41.3990922Z return g 2023-03-31T06:22:41.3991017Z call = _wrap_func(module.call_18) 2023-03-31T06:22:41.3991022Z 2023-03-31T06:22:41.3991027Z 2023-03-31T06:22:41.3991155Z def benchmark_compiled_module(): 2023-03-31T06:22:41.3991276Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.3991405Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.3991605Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3991811Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3992011Z arg2_1 = rand_strided({10, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.3992137Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.3992142Z 2023-03-31T06:22:41.3992147Z 2023-03-31T06:22:41.3992225Z if __name__ == "__main__": 2023-03-31T06:22:41.3992302Z import argparse 2023-03-31T06:22:41.3992463Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.3992469Z 2023-03-31T06:22:41.3992579Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.3992891Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.3993214Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.3993495Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.3993589Z args = parser.parse_args() 2023-03-31T06:22:41.3993595Z 2023-03-31T06:22:41.3993684Z if args.benchmark_kernels: 2023-03-31T06:22:41.3993872Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.3993939Z else: 2023-03-31T06:22:41.3994052Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.3994149Z benchmark_compiled_module() 2023-03-31T06:22:41.3994154Z 2023-03-31T06:22:41.3994208Z if p: 2023-03-31T06:22:41.3994361Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.3994457Z p.export_chrome_trace(path) 2023-03-31T06:22:41.3994594Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.3994600Z 2023-03-31T06:22:41.3994923Z [2023-03-31 05:35:59,350] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 18 2023-03-31T06:22:41.3995261Z [2023-03-31 05:35:59,422] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 19 2023-03-31T06:22:41.3995880Z [2023-03-31 05:36:27,748] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/lc/clctcsiy2pyfaewklsmnwkm5sbdpi3byhkk6w3y2tjaeqexj6pou.py 2023-03-31T06:22:41.3996138Z [2023-03-31 05:36:27,748] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.3996146Z 2023-03-31T06:22:41.3996231Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.3996335Z import torch 2023-03-31T06:22:41.3996442Z import math 2023-03-31T06:22:41.3996548Z import random 2023-03-31T06:22:41.3996649Z import os 2023-03-31T06:22:41.3996761Z import tempfile 2023-03-31T06:22:41.3996914Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.3996919Z 2023-03-31T06:22:41.3997026Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.3997239Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.3997406Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.3997413Z 2023-03-31T06:22:41.3997524Z aten = torch.ops.aten 2023-03-31T06:22:41.3997693Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.3997820Z async_compile = AsyncCompile() 2023-03-31T06:22:41.3997825Z 2023-03-31T06:22:41.3997952Z async_compile.wait(globals()) 2023-03-31T06:22:41.3998056Z del async_compile 2023-03-31T06:22:41.3998169Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.3998274Z wrapper = ( 2023-03-31T06:22:41.3998410Z ''' 2023-03-31T06:22:41.3998515Z #include 2023-03-31T06:22:41.3998654Z #include 2023-03-31T06:22:41.3998660Z 2023-03-31T06:22:41.3998787Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.3998792Z 2023-03-31T06:22:41.3998920Z template 2023-03-31T06:22:41.3999032Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.3999148Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.3999304Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.3999467Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.3999640Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.3999749Z return kernel_cpp; 2023-03-31T06:22:41.3999844Z } 2023-03-31T06:22:41.3999961Z std::vector call_19(std::vector args) { 2023-03-31T06:22:41.4000083Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.4000216Z arg0_1 = args[0]; 2023-03-31T06:22:41.4000320Z arg1_1 = args[1]; 2023-03-31T06:22:41.4000490Z auto buf0 = at::empty_strided({1, 30}, {30, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.4000676Z at::mm_out(buf0, at::as_strided(arg1_1, {1, 10}, {10, 1}), at::as_strided(arg0_1, {10, 30}, {1, 10})); 2023-03-31T06:22:41.4000799Z arg0_1.reset(); 2023-03-31T06:22:41.4000907Z arg1_1.reset(); 2023-03-31T06:22:41.4001029Z return std::vector({at::as_strided(buf0, {30, }, {1, })}); 2023-03-31T06:22:41.4001123Z } 2023-03-31T06:22:41.4001243Z ''' 2023-03-31T06:22:41.4001337Z ) 2023-03-31T06:22:41.4001342Z 2023-03-31T06:22:41.4001452Z module = load_inline( 2023-03-31T06:22:41.4001785Z name='inline_extension_cxp4be75bbmqgeukctfe5zhqvszmnld5ic57cnjrg7rvsrryf2z5', 2023-03-31T06:22:41.4001901Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4002006Z functions=['call_19'], 2023-03-31T06:22:41.4002439Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4002632Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4003675Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4003717Z 2023-03-31T06:22:41.4003829Z def _wrap_func(f): 2023-03-31T06:22:41.4003930Z def g(args): 2023-03-31T06:22:41.4004039Z return f(args) 2023-03-31T06:22:41.4004144Z return g 2023-03-31T06:22:41.4004226Z call = _wrap_func(module.call_19) 2023-03-31T06:22:41.4004274Z 2023-03-31T06:22:41.4004281Z 2023-03-31T06:22:41.4004364Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4004546Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4004707Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4004960Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.4005194Z arg1_1 = rand_strided({10, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.4005372Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.4005377Z 2023-03-31T06:22:41.4005382Z 2023-03-31T06:22:41.4005499Z if __name__ == "__main__": 2023-03-31T06:22:41.4005609Z import argparse 2023-03-31T06:22:41.4005731Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4005736Z 2023-03-31T06:22:41.4005877Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4006240Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4006610Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4006964Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4007089Z args = parser.parse_args() 2023-03-31T06:22:41.4007094Z 2023-03-31T06:22:41.4007221Z if args.benchmark_kernels: 2023-03-31T06:22:41.4007448Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4007503Z else: 2023-03-31T06:22:41.4007645Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4007775Z benchmark_compiled_module() 2023-03-31T06:22:41.4007783Z 2023-03-31T06:22:41.4007908Z if p: 2023-03-31T06:22:41.4008089Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4008213Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4008419Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4008468Z 2023-03-31T06:22:41.4008783Z [2023-03-31 05:36:27,748] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 19 2023-03-31T06:22:41.4008848Z PASSED [ 44%] 2023-03-31T06:22:41.4009279Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_linear_unary_cpu [2023-03-31 05:36:27,806] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 20 2023-03-31T06:22:41.4009712Z [2023-03-31 05:36:57,263] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/46/c46zbqbikyrgo2okbl43mgwm4epfg6uwkhzul4uql2awknwl7xyk.py 2023-03-31T06:22:41.4009968Z [2023-03-31 05:36:57,264] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4009974Z 2023-03-31T06:22:41.4010123Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4010233Z import torch 2023-03-31T06:22:41.4010333Z import math 2023-03-31T06:22:41.4010435Z import random 2023-03-31T06:22:41.4010492Z import os 2023-03-31T06:22:41.4010598Z import tempfile 2023-03-31T06:22:41.4010744Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4010750Z 2023-03-31T06:22:41.4010901Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4011059Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4011273Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4011279Z 2023-03-31T06:22:41.4011388Z aten = torch.ops.aten 2023-03-31T06:22:41.4011586Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4011669Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4011711Z 2023-03-31T06:22:41.4011858Z async_compile.wait(globals()) 2023-03-31T06:22:41.4011963Z del async_compile 2023-03-31T06:22:41.4012114Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4012225Z wrapper = ( 2023-03-31T06:22:41.4012339Z ''' 2023-03-31T06:22:41.4012464Z #include 2023-03-31T06:22:41.4012529Z #include 2023-03-31T06:22:41.4012535Z 2023-03-31T06:22:41.4012659Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4012664Z 2023-03-31T06:22:41.4012787Z template 2023-03-31T06:22:41.4012943Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4013063Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4013224Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4013354Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4013466Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4013575Z return kernel_cpp; 2023-03-31T06:22:41.4013687Z } 2023-03-31T06:22:41.4013878Z std::vector call_20(std::vector args) { 2023-03-31T06:22:41.4014012Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4014116Z arg0_1 = args[0]; 2023-03-31T06:22:41.4014225Z arg1_1 = args[1]; 2023-03-31T06:22:41.4014284Z arg2_1 = args[2]; 2023-03-31T06:22:41.4014384Z arg3_1 = args[3]; 2023-03-31T06:22:41.4014479Z 2023-03-31T06:22:41.4014612Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4014756Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4014873Z .findSchemaOrThrow( 2023-03-31T06:22:41.4015036Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4015093Z "") 2023-03-31T06:22:41.4015198Z .typed< 2023-03-31T06:22:41.4015302Z at::Tensor( 2023-03-31T06:22:41.4015429Z const at::Tensor& input_t, 2023-03-31T06:22:41.4015563Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4015711Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4015882Z c10::string_view attr, 2023-03-31T06:22:41.4015992Z torch::List> scalars, 2023-03-31T06:22:41.4016142Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4016248Z 2023-03-31T06:22:41.4016505Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.4016644Z arg1_1.reset(); 2023-03-31T06:22:41.4016748Z arg2_1.reset(); 2023-03-31T06:22:41.4016848Z arg3_1.reset(); 2023-03-31T06:22:41.4016984Z return std::vector({buf0}); 2023-03-31T06:22:41.4017036Z } 2023-03-31T06:22:41.4017173Z ''' 2023-03-31T06:22:41.4017267Z ) 2023-03-31T06:22:41.4017272Z 2023-03-31T06:22:41.4017382Z module = load_inline( 2023-03-31T06:22:41.4017676Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.4025172Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4025390Z functions=['call_20'], 2023-03-31T06:22:41.4025789Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4025952Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4026678Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4026693Z 2023-03-31T06:22:41.4026773Z def _wrap_func(f): 2023-03-31T06:22:41.4026848Z def g(args): 2023-03-31T06:22:41.4026925Z return f(args) 2023-03-31T06:22:41.4026996Z return g 2023-03-31T06:22:41.4027094Z call = _wrap_func(module.call_20) 2023-03-31T06:22:41.4027100Z 2023-03-31T06:22:41.4027210Z 2023-03-31T06:22:41.4027338Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4027446Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4027577Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4027795Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4028004Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4028219Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4028439Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4028578Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4028584Z 2023-03-31T06:22:41.4028589Z 2023-03-31T06:22:41.4028669Z if __name__ == "__main__": 2023-03-31T06:22:41.4028750Z import argparse 2023-03-31T06:22:41.4028872Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4028879Z 2023-03-31T06:22:41.4028993Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4029308Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4029648Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4029931Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4030062Z args = parser.parse_args() 2023-03-31T06:22:41.4030068Z 2023-03-31T06:22:41.4030161Z if args.benchmark_kernels: 2023-03-31T06:22:41.4030354Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4030408Z else: 2023-03-31T06:22:41.4030524Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4030621Z benchmark_compiled_module() 2023-03-31T06:22:41.4030629Z 2023-03-31T06:22:41.4030701Z if p: 2023-03-31T06:22:41.4030855Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4030953Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4031091Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4031097Z 2023-03-31T06:22:41.4031382Z [2023-03-31 05:36:57,264] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 20 2023-03-31T06:22:41.4031677Z [2023-03-31 05:36:57,312] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 21 2023-03-31T06:22:41.4032089Z [2023-03-31 05:37:24,855] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ng/cngqxkyv5i2hwmohzqfpidfahcrmcze7np5wqubj65ih7eabtcfs.py 2023-03-31T06:22:41.4032312Z [2023-03-31 05:37:24,856] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4032317Z 2023-03-31T06:22:41.4032420Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4032503Z import torch 2023-03-31T06:22:41.4032575Z import math 2023-03-31T06:22:41.4032649Z import random 2023-03-31T06:22:41.4032718Z import os 2023-03-31T06:22:41.4032779Z import tempfile 2023-03-31T06:22:41.4032898Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4032903Z 2023-03-31T06:22:41.4033023Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4033151Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4033284Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4033290Z 2023-03-31T06:22:41.4033375Z aten = torch.ops.aten 2023-03-31T06:22:41.4033513Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4033594Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4033613Z 2023-03-31T06:22:41.4033691Z async_compile.wait(globals()) 2023-03-31T06:22:41.4033764Z del async_compile 2023-03-31T06:22:41.4033921Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4034028Z wrapper = ( 2023-03-31T06:22:41.4034110Z ''' 2023-03-31T06:22:41.4034189Z #include 2023-03-31T06:22:41.4034250Z #include 2023-03-31T06:22:41.4034268Z 2023-03-31T06:22:41.4034356Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4034361Z 2023-03-31T06:22:41.4034452Z template 2023-03-31T06:22:41.4034576Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4034665Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4034789Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4034890Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4035016Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4035082Z return kernel_cpp; 2023-03-31T06:22:41.4035144Z } 2023-03-31T06:22:41.4035274Z std::vector call_21(std::vector args) { 2023-03-31T06:22:41.4035372Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4035450Z arg0_1 = args[0]; 2023-03-31T06:22:41.4035525Z arg1_1 = args[1]; 2023-03-31T06:22:41.4035597Z arg2_1 = args[2]; 2023-03-31T06:22:41.4035647Z 2023-03-31T06:22:41.4035747Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4035843Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4035933Z .findSchemaOrThrow( 2023-03-31T06:22:41.4036031Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4036100Z "") 2023-03-31T06:22:41.4036168Z .typed< 2023-03-31T06:22:41.4036230Z at::Tensor( 2023-03-31T06:22:41.4036395Z const at::Tensor& input_t, 2023-03-31T06:22:41.4036498Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4036616Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4036705Z c10::string_view attr, 2023-03-31T06:22:41.4036825Z torch::List> scalars, 2023-03-31T06:22:41.4036950Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4037006Z 2023-03-31T06:22:41.4037243Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.4037318Z arg1_1.reset(); 2023-03-31T06:22:41.4037389Z arg2_1.reset(); 2023-03-31T06:22:41.4037499Z return std::vector({buf0}); 2023-03-31T06:22:41.4037564Z } 2023-03-31T06:22:41.4037646Z ''' 2023-03-31T06:22:41.4037741Z ) 2023-03-31T06:22:41.4037761Z 2023-03-31T06:22:41.4037827Z module = load_inline( 2023-03-31T06:22:41.4038102Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.4038189Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4038305Z functions=['call_21'], 2023-03-31T06:22:41.4038688Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4038843Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4039552Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4039561Z 2023-03-31T06:22:41.4039639Z def _wrap_func(f): 2023-03-31T06:22:41.4039698Z def g(args): 2023-03-31T06:22:41.4039774Z return f(args) 2023-03-31T06:22:41.4039842Z return g 2023-03-31T06:22:41.4039940Z call = _wrap_func(module.call_21) 2023-03-31T06:22:41.4039946Z 2023-03-31T06:22:41.4039951Z 2023-03-31T06:22:41.4040045Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4040165Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4040293Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4040534Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4040759Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4040977Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4041103Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4041108Z 2023-03-31T06:22:41.4041112Z 2023-03-31T06:22:41.4041192Z if __name__ == "__main__": 2023-03-31T06:22:41.4041269Z import argparse 2023-03-31T06:22:41.4041398Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4041408Z 2023-03-31T06:22:41.4041522Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4041832Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4042172Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4042439Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4042532Z args = parser.parse_args() 2023-03-31T06:22:41.4042537Z 2023-03-31T06:22:41.4042627Z if args.benchmark_kernels: 2023-03-31T06:22:41.4042816Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4042885Z else: 2023-03-31T06:22:41.4043148Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4043246Z benchmark_compiled_module() 2023-03-31T06:22:41.4043252Z 2023-03-31T06:22:41.4043376Z if p: 2023-03-31T06:22:41.4043531Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4043629Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4043768Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4043773Z 2023-03-31T06:22:41.4044063Z [2023-03-31 05:37:24,856] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 21 2023-03-31T06:22:41.4044333Z [2023-03-31 05:37:24,938] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 22 2023-03-31T06:22:41.4044747Z [2023-03-31 05:37:55,973] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/xa/cxabbrzmqg7hsjlvcwf3mt64hkjgp3ytqh47vmzrjz3xhs6ofv6l.py 2023-03-31T06:22:41.4045012Z [2023-03-31 05:37:55,974] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4045018Z 2023-03-31T06:22:41.4045117Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4045179Z import torch 2023-03-31T06:22:41.4045252Z import math 2023-03-31T06:22:41.4045325Z import random 2023-03-31T06:22:41.4045394Z import os 2023-03-31T06:22:41.4045469Z import tempfile 2023-03-31T06:22:41.4045587Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4045592Z 2023-03-31T06:22:41.4045715Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4045831Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4045967Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4045972Z 2023-03-31T06:22:41.4046054Z aten = torch.ops.aten 2023-03-31T06:22:41.4046192Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4046286Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4046294Z 2023-03-31T06:22:41.4046385Z async_compile.wait(globals()) 2023-03-31T06:22:41.4046459Z del async_compile 2023-03-31T06:22:41.4046580Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4046642Z wrapper = ( 2023-03-31T06:22:41.4046725Z ''' 2023-03-31T06:22:41.4046801Z #include 2023-03-31T06:22:41.4046876Z #include 2023-03-31T06:22:41.4046882Z 2023-03-31T06:22:41.4046975Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4046980Z 2023-03-31T06:22:41.4047074Z template 2023-03-31T06:22:41.4047232Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4047338Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4047463Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4047562Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4047685Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4047766Z return kernel_cpp; 2023-03-31T06:22:41.4047829Z } 2023-03-31T06:22:41.4047959Z std::vector call_22(std::vector args) { 2023-03-31T06:22:41.4048050Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4048125Z arg0_1 = args[0]; 2023-03-31T06:22:41.4048196Z arg1_1 = args[1]; 2023-03-31T06:22:41.4048270Z arg2_1 = args[2]; 2023-03-31T06:22:41.4048341Z arg3_1 = args[3]; 2023-03-31T06:22:41.4048405Z 2023-03-31T06:22:41.4048504Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4048584Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4048675Z .findSchemaOrThrow( 2023-03-31T06:22:41.4048777Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4048848Z "") 2023-03-31T06:22:41.4048918Z .typed< 2023-03-31T06:22:41.4048994Z at::Tensor( 2023-03-31T06:22:41.4049094Z const at::Tensor& input_t, 2023-03-31T06:22:41.4049181Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4049298Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4049392Z c10::string_view attr, 2023-03-31T06:22:41.4049511Z torch::List> scalars, 2023-03-31T06:22:41.4049664Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4049732Z 2023-03-31T06:22:41.4049955Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.4050029Z arg1_1.reset(); 2023-03-31T06:22:41.4050089Z arg2_1.reset(); 2023-03-31T06:22:41.4050158Z arg3_1.reset(); 2023-03-31T06:22:41.4050264Z return std::vector({buf0}); 2023-03-31T06:22:41.4050330Z } 2023-03-31T06:22:41.4050410Z ''' 2023-03-31T06:22:41.4050474Z ) 2023-03-31T06:22:41.4050480Z 2023-03-31T06:22:41.4050545Z module = load_inline( 2023-03-31T06:22:41.4050811Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.4050897Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4051012Z functions=['call_22'], 2023-03-31T06:22:41.4051425Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4051587Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4052303Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4052312Z 2023-03-31T06:22:41.4052388Z def _wrap_func(f): 2023-03-31T06:22:41.4052459Z def g(args): 2023-03-31T06:22:41.4052521Z return f(args) 2023-03-31T06:22:41.4052589Z return g 2023-03-31T06:22:41.4052683Z call = _wrap_func(module.call_22) 2023-03-31T06:22:41.4052689Z 2023-03-31T06:22:41.4052693Z 2023-03-31T06:22:41.4052791Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4052910Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4053039Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4053256Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4053463Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4053657Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4053890Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4054055Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4054061Z 2023-03-31T06:22:41.4054066Z 2023-03-31T06:22:41.4054146Z if __name__ == "__main__": 2023-03-31T06:22:41.4054222Z import argparse 2023-03-31T06:22:41.4054352Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4054361Z 2023-03-31T06:22:41.4054473Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4054786Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4055125Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4055389Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4055482Z args = parser.parse_args() 2023-03-31T06:22:41.4055490Z 2023-03-31T06:22:41.4055581Z if args.benchmark_kernels: 2023-03-31T06:22:41.4055773Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4055840Z else: 2023-03-31T06:22:41.4055952Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4056047Z benchmark_compiled_module() 2023-03-31T06:22:41.4056053Z 2023-03-31T06:22:41.4056111Z if p: 2023-03-31T06:22:41.4056262Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4056358Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4056526Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4056532Z 2023-03-31T06:22:41.4056815Z [2023-03-31 05:37:55,974] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 22 2023-03-31T06:22:41.4057083Z [2023-03-31 05:37:56,176] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 23 2023-03-31T06:22:41.4057494Z [2023-03-31 05:38:30,514] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ua/cuajdcxkgvwbovfapry6lwfpccrubirrekgfjv6q4ok6hazsfatw.py 2023-03-31T06:22:41.4057713Z [2023-03-31 05:38:30,514] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4057719Z 2023-03-31T06:22:41.4057817Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4057905Z import torch 2023-03-31T06:22:41.4057977Z import math 2023-03-31T06:22:41.4058050Z import random 2023-03-31T06:22:41.4058118Z import os 2023-03-31T06:22:41.4058193Z import tempfile 2023-03-31T06:22:41.4058313Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4058318Z 2023-03-31T06:22:41.4058437Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4058551Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4058684Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4058690Z 2023-03-31T06:22:41.4058774Z aten = torch.ops.aten 2023-03-31T06:22:41.4058915Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4059010Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4059015Z 2023-03-31T06:22:41.4059105Z async_compile.wait(globals()) 2023-03-31T06:22:41.4059180Z del async_compile 2023-03-31T06:22:41.4059301Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4059363Z wrapper = ( 2023-03-31T06:22:41.4059446Z ''' 2023-03-31T06:22:41.4059521Z #include 2023-03-31T06:22:41.4059596Z #include 2023-03-31T06:22:41.4059602Z 2023-03-31T06:22:41.4059697Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4059702Z 2023-03-31T06:22:41.4059795Z template 2023-03-31T06:22:41.4059917Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4059990Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4060115Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4060244Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4060406Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4060486Z return kernel_cpp; 2023-03-31T06:22:41.4060549Z } 2023-03-31T06:22:41.4060677Z std::vector call_23(std::vector args) { 2023-03-31T06:22:41.4060758Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4060831Z arg0_1 = args[0]; 2023-03-31T06:22:41.4060907Z arg1_1 = args[1]; 2023-03-31T06:22:41.4060980Z arg2_1 = args[2]; 2023-03-31T06:22:41.4061044Z 2023-03-31T06:22:41.4061142Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4061239Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4061315Z .findSchemaOrThrow( 2023-03-31T06:22:41.4061413Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4061481Z "") 2023-03-31T06:22:41.4061550Z .typed< 2023-03-31T06:22:41.4061626Z at::Tensor( 2023-03-31T06:22:41.4061722Z const at::Tensor& input_t, 2023-03-31T06:22:41.4061824Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4061931Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4062021Z c10::string_view attr, 2023-03-31T06:22:41.4062141Z torch::List> scalars, 2023-03-31T06:22:41.4062262Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4062330Z 2023-03-31T06:22:41.4062680Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.4062755Z arg1_1.reset(); 2023-03-31T06:22:41.4062868Z arg2_1.reset(); 2023-03-31T06:22:41.4062963Z return std::vector({buf0}); 2023-03-31T06:22:41.4063027Z } 2023-03-31T06:22:41.4063111Z ''' 2023-03-31T06:22:41.4063175Z ) 2023-03-31T06:22:41.4063181Z 2023-03-31T06:22:41.4063259Z module = load_inline( 2023-03-31T06:22:41.4063530Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.4063619Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4063725Z functions=['call_23'], 2023-03-31T06:22:41.4064107Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4064263Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4064972Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4065009Z 2023-03-31T06:22:41.4065088Z def _wrap_func(f): 2023-03-31T06:22:41.4065160Z def g(args): 2023-03-31T06:22:41.4065236Z return f(args) 2023-03-31T06:22:41.4065305Z return g 2023-03-31T06:22:41.4065390Z call = _wrap_func(module.call_23) 2023-03-31T06:22:41.4065412Z 2023-03-31T06:22:41.4065417Z 2023-03-31T06:22:41.4065501Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4065621Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4065749Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4065967Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4066183Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4066396Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4066525Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4066530Z 2023-03-31T06:22:41.4066535Z 2023-03-31T06:22:41.4066616Z if __name__ == "__main__": 2023-03-31T06:22:41.4066679Z import argparse 2023-03-31T06:22:41.4066810Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4066815Z 2023-03-31T06:22:41.4066957Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4067295Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4067632Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4067912Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4068008Z args = parser.parse_args() 2023-03-31T06:22:41.4068013Z 2023-03-31T06:22:41.4068107Z if args.benchmark_kernels: 2023-03-31T06:22:41.4068284Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4068351Z else: 2023-03-31T06:22:41.4068465Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4068563Z benchmark_compiled_module() 2023-03-31T06:22:41.4068569Z 2023-03-31T06:22:41.4068641Z if p: 2023-03-31T06:22:41.4068794Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4068893Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4069033Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4069038Z 2023-03-31T06:22:41.4069304Z [2023-03-31 05:38:30,515] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 23 2023-03-31T06:22:41.4069573Z [2023-03-31 05:38:30,645] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 24 2023-03-31T06:22:41.4070009Z [2023-03-31 05:38:58,067] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/l5/cl5flatyvbisxxhkhh4ezho2jobqsjvtxvt6i6d5xxqsow4prubl.py 2023-03-31T06:22:41.4070232Z [2023-03-31 05:38:58,067] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4070237Z 2023-03-31T06:22:41.4070337Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4070413Z import torch 2023-03-31T06:22:41.4070487Z import math 2023-03-31T06:22:41.4070562Z import random 2023-03-31T06:22:41.4070616Z import os 2023-03-31T06:22:41.4070692Z import tempfile 2023-03-31T06:22:41.4070812Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4070817Z 2023-03-31T06:22:41.4070937Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4071064Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4071230Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4071236Z 2023-03-31T06:22:41.4071319Z aten = torch.ops.aten 2023-03-31T06:22:41.4071459Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4071541Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4071546Z 2023-03-31T06:22:41.4071639Z async_compile.wait(globals()) 2023-03-31T06:22:41.4071715Z del async_compile 2023-03-31T06:22:41.4071837Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4071916Z wrapper = ( 2023-03-31T06:22:41.4072002Z ''' 2023-03-31T06:22:41.4072083Z #include 2023-03-31T06:22:41.4072146Z #include 2023-03-31T06:22:41.4072151Z 2023-03-31T06:22:41.4072241Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4072246Z 2023-03-31T06:22:41.4072339Z template 2023-03-31T06:22:41.4072464Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4072552Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4072677Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4072776Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4072889Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4072969Z return kernel_cpp; 2023-03-31T06:22:41.4073033Z } 2023-03-31T06:22:41.4073164Z std::vector call_24(std::vector args) { 2023-03-31T06:22:41.4073269Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4073343Z arg0_1 = args[0]; 2023-03-31T06:22:41.4073449Z arg1_1 = args[1]; 2023-03-31T06:22:41.4073547Z arg2_1 = args[2]; 2023-03-31T06:22:41.4073620Z arg3_1 = args[3]; 2023-03-31T06:22:41.4073682Z 2023-03-31T06:22:41.4073782Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4073877Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4073964Z .findSchemaOrThrow( 2023-03-31T06:22:41.4074063Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4074120Z "") 2023-03-31T06:22:41.4074188Z .typed< 2023-03-31T06:22:41.4074264Z at::Tensor( 2023-03-31T06:22:41.4074364Z const at::Tensor& input_t, 2023-03-31T06:22:41.4074465Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4074582Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4074671Z c10::string_view attr, 2023-03-31T06:22:41.4074779Z torch::List> scalars, 2023-03-31T06:22:41.4074904Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4074974Z 2023-03-31T06:22:41.4075201Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4075275Z arg1_1.reset(); 2023-03-31T06:22:41.4075346Z arg2_1.reset(); 2023-03-31T06:22:41.4075417Z arg3_1.reset(); 2023-03-31T06:22:41.4075521Z return std::vector({buf0}); 2023-03-31T06:22:41.4075574Z } 2023-03-31T06:22:41.4075656Z ''' 2023-03-31T06:22:41.4075719Z ) 2023-03-31T06:22:41.4075724Z 2023-03-31T06:22:41.4075801Z module = load_inline( 2023-03-31T06:22:41.4076103Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.4076192Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4076308Z functions=['call_24'], 2023-03-31T06:22:41.4076670Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4076828Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4077537Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4077573Z 2023-03-31T06:22:41.4077650Z def _wrap_func(f): 2023-03-31T06:22:41.4077723Z def g(args): 2023-03-31T06:22:41.4077801Z return f(args) 2023-03-31T06:22:41.4077870Z return g 2023-03-31T06:22:41.4077964Z call = _wrap_func(module.call_24) 2023-03-31T06:22:41.4077970Z 2023-03-31T06:22:41.4077975Z 2023-03-31T06:22:41.4078070Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4078174Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4078302Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4078520Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4078729Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4078938Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4079153Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4079290Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4079295Z 2023-03-31T06:22:41.4079300Z 2023-03-31T06:22:41.4079380Z if __name__ == "__main__": 2023-03-31T06:22:41.4079444Z import argparse 2023-03-31T06:22:41.4079574Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4079579Z 2023-03-31T06:22:41.4079690Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4080054Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4080422Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4080699Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4080792Z args = parser.parse_args() 2023-03-31T06:22:41.4080799Z 2023-03-31T06:22:41.4080889Z if args.benchmark_kernels: 2023-03-31T06:22:41.4081065Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4081132Z else: 2023-03-31T06:22:41.4081246Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4081342Z benchmark_compiled_module() 2023-03-31T06:22:41.4081348Z 2023-03-31T06:22:41.4081416Z if p: 2023-03-31T06:22:41.4081566Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4081662Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4081803Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4081810Z 2023-03-31T06:22:41.4082077Z [2023-03-31 05:38:58,068] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 24 2023-03-31T06:22:41.4082340Z [2023-03-31 05:38:58,173] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 25 2023-03-31T06:22:41.4082749Z [2023-03-31 05:39:28,372] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zb/czbbcnmirtayklc5iulppqj6z3e7rbq7jsyubcjuin4gip66lxbl.py 2023-03-31T06:22:41.4083157Z [2023-03-31 05:39:28,372] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4083168Z 2023-03-31T06:22:41.4083316Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4083432Z import torch 2023-03-31T06:22:41.4083508Z import math 2023-03-31T06:22:41.4083583Z import random 2023-03-31T06:22:41.4083637Z import os 2023-03-31T06:22:41.4083712Z import tempfile 2023-03-31T06:22:41.4083835Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4083843Z 2023-03-31T06:22:41.4083964Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4084091Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4084225Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4084231Z 2023-03-31T06:22:41.4084313Z aten = torch.ops.aten 2023-03-31T06:22:41.4084548Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4084632Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4084637Z 2023-03-31T06:22:41.4084732Z async_compile.wait(globals()) 2023-03-31T06:22:41.4084808Z del async_compile 2023-03-31T06:22:41.4084930Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4085006Z wrapper = ( 2023-03-31T06:22:41.4085098Z ''' 2023-03-31T06:22:41.4085176Z #include 2023-03-31T06:22:41.4085239Z #include 2023-03-31T06:22:41.4085244Z 2023-03-31T06:22:41.4085348Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4085355Z 2023-03-31T06:22:41.4085451Z template 2023-03-31T06:22:41.4085577Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4085666Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4085792Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4085890Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4086006Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4086085Z return kernel_cpp; 2023-03-31T06:22:41.4086148Z } 2023-03-31T06:22:41.4086278Z std::vector call_25(std::vector args) { 2023-03-31T06:22:41.4086375Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4086449Z arg0_1 = args[0]; 2023-03-31T06:22:41.4086521Z arg1_1 = args[1]; 2023-03-31T06:22:41.4086580Z arg2_1 = args[2]; 2023-03-31T06:22:41.4086643Z 2023-03-31T06:22:41.4086740Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4086872Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4087039Z .findSchemaOrThrow( 2023-03-31T06:22:41.4087139Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4087207Z "") 2023-03-31T06:22:41.4087263Z .typed< 2023-03-31T06:22:41.4087340Z at::Tensor( 2023-03-31T06:22:41.4087440Z const at::Tensor& input_t, 2023-03-31T06:22:41.4087543Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4087665Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4087756Z c10::string_view attr, 2023-03-31T06:22:41.4087882Z torch::List> scalars, 2023-03-31T06:22:41.4088005Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4088059Z 2023-03-31T06:22:41.4088306Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4088381Z arg1_1.reset(); 2023-03-31T06:22:41.4088458Z arg2_1.reset(); 2023-03-31T06:22:41.4088568Z return std::vector({buf0}); 2023-03-31T06:22:41.4088631Z } 2023-03-31T06:22:41.4088712Z ''' 2023-03-31T06:22:41.4088761Z ) 2023-03-31T06:22:41.4088766Z 2023-03-31T06:22:41.4088844Z module = load_inline( 2023-03-31T06:22:41.4089106Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.4089195Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4089311Z functions=['call_25'], 2023-03-31T06:22:41.4089732Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4089890Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4090594Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4090603Z 2023-03-31T06:22:41.4090679Z def _wrap_func(f): 2023-03-31T06:22:41.4090738Z def g(args): 2023-03-31T06:22:41.4090814Z return f(args) 2023-03-31T06:22:41.4090881Z return g 2023-03-31T06:22:41.4090975Z call = _wrap_func(module.call_25) 2023-03-31T06:22:41.4091009Z 2023-03-31T06:22:41.4091014Z 2023-03-31T06:22:41.4091109Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4091228Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4091358Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4091560Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4091770Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4091988Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4092117Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4092122Z 2023-03-31T06:22:41.4092127Z 2023-03-31T06:22:41.4092206Z if __name__ == "__main__": 2023-03-31T06:22:41.4092283Z import argparse 2023-03-31T06:22:41.4092413Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4092418Z 2023-03-31T06:22:41.4092534Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4092847Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4093173Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4093452Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4093545Z args = parser.parse_args() 2023-03-31T06:22:41.4093578Z 2023-03-31T06:22:41.4093694Z if args.benchmark_kernels: 2023-03-31T06:22:41.4093886Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4093954Z else: 2023-03-31T06:22:41.4094069Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4094165Z benchmark_compiled_module() 2023-03-31T06:22:41.4094171Z 2023-03-31T06:22:41.4094227Z if p: 2023-03-31T06:22:41.4094380Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4094477Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4094616Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4094622Z 2023-03-31T06:22:41.4094902Z [2023-03-31 05:39:28,372] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 25 2023-03-31T06:22:41.4095166Z [2023-03-31 05:39:28,424] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 26 2023-03-31T06:22:41.4095571Z [2023-03-31 05:39:49,797] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/57/c57kpxirncl7mjsvyfyjn27plkjeuugpgvtooqhfnzzuivpry2te.py 2023-03-31T06:22:41.4095793Z [2023-03-31 05:39:49,797] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4095799Z 2023-03-31T06:22:41.4095895Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4095955Z import torch 2023-03-31T06:22:41.4096029Z import math 2023-03-31T06:22:41.4096103Z import random 2023-03-31T06:22:41.4096170Z import os 2023-03-31T06:22:41.4096243Z import tempfile 2023-03-31T06:22:41.4096392Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4096398Z 2023-03-31T06:22:41.4096519Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4096632Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4096765Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4096770Z 2023-03-31T06:22:41.4096854Z aten = torch.ops.aten 2023-03-31T06:22:41.4097001Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4097098Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4097103Z 2023-03-31T06:22:41.4097194Z async_compile.wait(globals()) 2023-03-31T06:22:41.4097269Z del async_compile 2023-03-31T06:22:41.4097377Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4097449Z wrapper = ( 2023-03-31T06:22:41.4097563Z ''' 2023-03-31T06:22:41.4097640Z #include 2023-03-31T06:22:41.4097717Z #include 2023-03-31T06:22:41.4097722Z 2023-03-31T06:22:41.4097817Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4097822Z 2023-03-31T06:22:41.4097916Z template 2023-03-31T06:22:41.4098026Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4098113Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4098237Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4098333Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4098462Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4098544Z return kernel_cpp; 2023-03-31T06:22:41.4098607Z } 2023-03-31T06:22:41.4098722Z std::vector call_26(std::vector args) { 2023-03-31T06:22:41.4098825Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4098898Z arg0_1 = args[0]; 2023-03-31T06:22:41.4098973Z arg1_1 = args[1]; 2023-03-31T06:22:41.4099044Z arg2_1 = args[2]; 2023-03-31T06:22:41.4099115Z arg3_1 = args[3]; 2023-03-31T06:22:41.4099177Z 2023-03-31T06:22:41.4099265Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4099360Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4099448Z .findSchemaOrThrow( 2023-03-31T06:22:41.4099545Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4099612Z "") 2023-03-31T06:22:41.4099680Z .typed< 2023-03-31T06:22:41.4099757Z at::Tensor( 2023-03-31T06:22:41.4099873Z const at::Tensor& input_t, 2023-03-31T06:22:41.4100001Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4100121Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4100210Z c10::string_view attr, 2023-03-31T06:22:41.4100330Z torch::List> scalars, 2023-03-31T06:22:41.4100452Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4100521Z 2023-03-31T06:22:41.4100749Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4100811Z arg1_1.reset(); 2023-03-31T06:22:41.4100883Z arg2_1.reset(); 2023-03-31T06:22:41.4100953Z arg3_1.reset(); 2023-03-31T06:22:41.4101058Z return std::vector({buf0}); 2023-03-31T06:22:41.4101121Z } 2023-03-31T06:22:41.4101202Z ''' 2023-03-31T06:22:41.4101250Z ) 2023-03-31T06:22:41.4101269Z 2023-03-31T06:22:41.4101333Z module = load_inline( 2023-03-31T06:22:41.4101608Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.4101698Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4101817Z functions=['call_26'], 2023-03-31T06:22:41.4102198Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4102354Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4103228Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4103237Z 2023-03-31T06:22:41.4103316Z def _wrap_func(f): 2023-03-31T06:22:41.4103375Z def g(args): 2023-03-31T06:22:41.4103455Z return f(args) 2023-03-31T06:22:41.4103529Z return g 2023-03-31T06:22:41.4103625Z call = _wrap_func(module.call_26) 2023-03-31T06:22:41.4103630Z 2023-03-31T06:22:41.4103635Z 2023-03-31T06:22:41.4103732Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4103851Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4103981Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4104234Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4104431Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4104642Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4104851Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4104985Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4104991Z 2023-03-31T06:22:41.4104996Z 2023-03-31T06:22:41.4105080Z if __name__ == "__main__": 2023-03-31T06:22:41.4105160Z import argparse 2023-03-31T06:22:41.4105291Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4105297Z 2023-03-31T06:22:41.4105409Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4105718Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4106046Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4106327Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4106422Z args = parser.parse_args() 2023-03-31T06:22:41.4106427Z 2023-03-31T06:22:41.4106518Z if args.benchmark_kernels: 2023-03-31T06:22:41.4106708Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4106776Z else: 2023-03-31T06:22:41.4106917Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4107051Z benchmark_compiled_module() 2023-03-31T06:22:41.4107057Z 2023-03-31T06:22:41.4107112Z if p: 2023-03-31T06:22:41.4107262Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4107360Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4107500Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4107508Z 2023-03-31T06:22:41.4107790Z [2023-03-31 05:39:49,798] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 26 2023-03-31T06:22:41.4108059Z [2023-03-31 05:39:49,846] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 27 2023-03-31T06:22:41.4108472Z [2023-03-31 05:40:15,050] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ta/ctaxy5nucuhlthjczikris7vqvighpuknyvtnwcrgns23kxmcfyj.py 2023-03-31T06:22:41.4108699Z [2023-03-31 05:40:15,050] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4108707Z 2023-03-31T06:22:41.4108806Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4108865Z import torch 2023-03-31T06:22:41.4108937Z import math 2023-03-31T06:22:41.4109012Z import random 2023-03-31T06:22:41.4109081Z import os 2023-03-31T06:22:41.4109157Z import tempfile 2023-03-31T06:22:41.4109276Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4109285Z 2023-03-31T06:22:41.4109404Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4109546Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4109683Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4109689Z 2023-03-31T06:22:41.4109770Z aten = torch.ops.aten 2023-03-31T06:22:41.4109910Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4110006Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4110011Z 2023-03-31T06:22:41.4110107Z async_compile.wait(globals()) 2023-03-31T06:22:41.4110184Z del async_compile 2023-03-31T06:22:41.4110293Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4110367Z wrapper = ( 2023-03-31T06:22:41.4110451Z ''' 2023-03-31T06:22:41.4110527Z #include 2023-03-31T06:22:41.4110605Z #include 2023-03-31T06:22:41.4110610Z 2023-03-31T06:22:41.4110704Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4110738Z 2023-03-31T06:22:41.4110834Z template 2023-03-31T06:22:41.4110945Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4111034Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4111161Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4111257Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4111384Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4111464Z return kernel_cpp; 2023-03-31T06:22:41.4111527Z } 2023-03-31T06:22:41.4111645Z std::vector call_27(std::vector args) { 2023-03-31T06:22:41.4111739Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4111812Z arg0_1 = args[0]; 2023-03-31T06:22:41.4111886Z arg1_1 = args[1]; 2023-03-31T06:22:41.4111958Z arg2_1 = args[2]; 2023-03-31T06:22:41.4112020Z 2023-03-31T06:22:41.4112118Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4112199Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4112288Z .findSchemaOrThrow( 2023-03-31T06:22:41.4112386Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4112454Z "") 2023-03-31T06:22:41.4112526Z .typed< 2023-03-31T06:22:41.4112602Z at::Tensor( 2023-03-31T06:22:41.4112698Z const at::Tensor& input_t, 2023-03-31T06:22:41.4112785Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4112902Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4112993Z c10::string_view attr, 2023-03-31T06:22:41.4113145Z torch::List> scalars, 2023-03-31T06:22:41.4113302Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4113371Z 2023-03-31T06:22:41.4113613Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4113686Z arg1_1.reset(); 2023-03-31T06:22:41.4113745Z arg2_1.reset(); 2023-03-31T06:22:41.4113852Z return std::vector({buf0}); 2023-03-31T06:22:41.4113920Z } 2023-03-31T06:22:41.4114000Z ''' 2023-03-31T06:22:41.4114061Z ) 2023-03-31T06:22:41.4114069Z 2023-03-31T06:22:41.4114147Z module = load_inline( 2023-03-31T06:22:41.4114411Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.4114486Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4114601Z functions=['call_27'], 2023-03-31T06:22:41.4114986Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4115145Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4115856Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4115865Z 2023-03-31T06:22:41.4115948Z def _wrap_func(f): 2023-03-31T06:22:41.4116051Z def g(args): 2023-03-31T06:22:41.4116128Z return f(args) 2023-03-31T06:22:41.4116195Z return g 2023-03-31T06:22:41.4116278Z call = _wrap_func(module.call_27) 2023-03-31T06:22:41.4116283Z 2023-03-31T06:22:41.4116288Z 2023-03-31T06:22:41.4116382Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4116502Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4116633Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4116847Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4117056Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4117262Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4117418Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4117424Z 2023-03-31T06:22:41.4117428Z 2023-03-31T06:22:41.4117495Z if __name__ == "__main__": 2023-03-31T06:22:41.4117574Z import argparse 2023-03-31T06:22:41.4117705Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4117710Z 2023-03-31T06:22:41.4117820Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4118129Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4118473Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4118752Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4118844Z args = parser.parse_args() 2023-03-31T06:22:41.4118849Z 2023-03-31T06:22:41.4118941Z if args.benchmark_kernels: 2023-03-31T06:22:41.4119120Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4119188Z else: 2023-03-31T06:22:41.4119303Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4119397Z benchmark_compiled_module() 2023-03-31T06:22:41.4119403Z 2023-03-31T06:22:41.4119472Z if p: 2023-03-31T06:22:41.4119620Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4119716Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4119870Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4119902Z 2023-03-31T06:22:41.4120184Z [2023-03-31 05:40:15,050] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 27 2023-03-31T06:22:41.4120451Z [2023-03-31 05:40:15,220] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 28 2023-03-31T06:22:41.4120855Z [2023-03-31 05:40:39,646] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/2w/c2w7xyvhjjtbylckwtudlvjxtnju2rrxna3xdsl4u2epn425slkz.py 2023-03-31T06:22:41.4121081Z [2023-03-31 05:40:39,646] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4121087Z 2023-03-31T06:22:41.4121187Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4121260Z import torch 2023-03-31T06:22:41.4121331Z import math 2023-03-31T06:22:41.4121403Z import random 2023-03-31T06:22:41.4121458Z import os 2023-03-31T06:22:41.4121532Z import tempfile 2023-03-31T06:22:41.4121652Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4121660Z 2023-03-31T06:22:41.4121779Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4121906Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4122041Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4122047Z 2023-03-31T06:22:41.4122129Z aten = torch.ops.aten 2023-03-31T06:22:41.4122257Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4122353Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4122359Z 2023-03-31T06:22:41.4122483Z async_compile.wait(globals()) 2023-03-31T06:22:41.4122561Z del async_compile 2023-03-31T06:22:41.4122683Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4122757Z wrapper = ( 2023-03-31T06:22:41.4122840Z ''' 2023-03-31T06:22:41.4122902Z #include 2023-03-31T06:22:41.4122979Z #include 2023-03-31T06:22:41.4122984Z 2023-03-31T06:22:41.4123273Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4123280Z 2023-03-31T06:22:41.4123376Z template 2023-03-31T06:22:41.4123504Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4123595Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4123726Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4123825Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4123936Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4124081Z return kernel_cpp; 2023-03-31T06:22:41.4124145Z } 2023-03-31T06:22:41.4124276Z std::vector call_28(std::vector args) { 2023-03-31T06:22:41.4124382Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4124457Z arg0_1 = args[0]; 2023-03-31T06:22:41.4124531Z arg1_1 = args[1]; 2023-03-31T06:22:41.4124591Z arg2_1 = args[2]; 2023-03-31T06:22:41.4124663Z arg3_1 = args[3]; 2023-03-31T06:22:41.4124728Z 2023-03-31T06:22:41.4124831Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4124929Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4125019Z .findSchemaOrThrow( 2023-03-31T06:22:41.4125105Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4125174Z "") 2023-03-31T06:22:41.4125244Z .typed< 2023-03-31T06:22:41.4125321Z at::Tensor( 2023-03-31T06:22:41.4125422Z const at::Tensor& input_t, 2023-03-31T06:22:41.4125525Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4125645Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4125735Z c10::string_view attr, 2023-03-31T06:22:41.4125842Z torch::List> scalars, 2023-03-31T06:22:41.4125964Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4126030Z 2023-03-31T06:22:41.4126257Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.4126372Z arg1_1.reset(); 2023-03-31T06:22:41.4126482Z arg2_1.reset(); 2023-03-31T06:22:41.4126552Z arg3_1.reset(); 2023-03-31T06:22:41.4126645Z return std::vector({buf0}); 2023-03-31T06:22:41.4126707Z } 2023-03-31T06:22:41.4126794Z ''' 2023-03-31T06:22:41.4126860Z ) 2023-03-31T06:22:41.4126866Z 2023-03-31T06:22:41.4126947Z module = load_inline( 2023-03-31T06:22:41.4127213Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.4127304Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4127407Z functions=['call_28'], 2023-03-31T06:22:41.4127790Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4127947Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4128657Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4128665Z 2023-03-31T06:22:41.4128743Z def _wrap_func(f): 2023-03-31T06:22:41.4128816Z def g(args): 2023-03-31T06:22:41.4128894Z return f(args) 2023-03-31T06:22:41.4128968Z return g 2023-03-31T06:22:41.4129050Z call = _wrap_func(module.call_28) 2023-03-31T06:22:41.4129070Z 2023-03-31T06:22:41.4129074Z 2023-03-31T06:22:41.4129156Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4129311Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4129446Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4129661Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4129870Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4130081Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4130298Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4130433Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4130439Z 2023-03-31T06:22:41.4130445Z 2023-03-31T06:22:41.4130510Z if __name__ == "__main__": 2023-03-31T06:22:41.4130627Z import argparse 2023-03-31T06:22:41.4130758Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4130763Z 2023-03-31T06:22:41.4130877Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4131190Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4131528Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4131807Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4131902Z args = parser.parse_args() 2023-03-31T06:22:41.4131907Z 2023-03-31T06:22:41.4131996Z if args.benchmark_kernels: 2023-03-31T06:22:41.4132171Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4132239Z else: 2023-03-31T06:22:41.4132351Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4132449Z benchmark_compiled_module() 2023-03-31T06:22:41.4132454Z 2023-03-31T06:22:41.4132526Z if p: 2023-03-31T06:22:41.4132677Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4132773Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4132897Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4132902Z 2023-03-31T06:22:41.4133209Z [2023-03-31 05:40:39,646] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 28 2023-03-31T06:22:41.4133508Z [2023-03-31 05:40:39,697] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 29 2023-03-31T06:22:41.4133915Z [2023-03-31 05:41:01,693] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/vk/cvklvlxpabonrqeq3cf4uyysikss3tqeo57lkubfldxrktgi4uhc.py 2023-03-31T06:22:41.4134134Z [2023-03-31 05:41:01,693] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4134142Z 2023-03-31T06:22:41.4134239Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4134312Z import torch 2023-03-31T06:22:41.4134384Z import math 2023-03-31T06:22:41.4134457Z import random 2023-03-31T06:22:41.4134511Z import os 2023-03-31T06:22:41.4134585Z import tempfile 2023-03-31T06:22:41.4134704Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4134710Z 2023-03-31T06:22:41.4134828Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4134955Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4135092Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4135098Z 2023-03-31T06:22:41.4135177Z aten = torch.ops.aten 2023-03-31T06:22:41.4135303Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4135399Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4135404Z 2023-03-31T06:22:41.4135495Z async_compile.wait(globals()) 2023-03-31T06:22:41.4135572Z del async_compile 2023-03-31T06:22:41.4135695Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4135767Z wrapper = ( 2023-03-31T06:22:41.4135877Z ''' 2023-03-31T06:22:41.4135941Z #include 2023-03-31T06:22:41.4136017Z #include 2023-03-31T06:22:41.4136023Z 2023-03-31T06:22:41.4136116Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4136121Z 2023-03-31T06:22:41.4136213Z template 2023-03-31T06:22:41.4136338Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4136426Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4136553Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4136651Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4136764Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4136842Z return kernel_cpp; 2023-03-31T06:22:41.4136904Z } 2023-03-31T06:22:41.4137032Z std::vector call_29(std::vector args) { 2023-03-31T06:22:41.4137157Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4137232Z arg0_1 = args[0]; 2023-03-31T06:22:41.4137303Z arg1_1 = args[1]; 2023-03-31T06:22:41.4137364Z arg2_1 = args[2]; 2023-03-31T06:22:41.4137427Z 2023-03-31T06:22:41.4137527Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4137620Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4137709Z .findSchemaOrThrow( 2023-03-31T06:22:41.4137806Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4137861Z "") 2023-03-31T06:22:41.4137931Z .typed< 2023-03-31T06:22:41.4138009Z at::Tensor( 2023-03-31T06:22:41.4138108Z const at::Tensor& input_t, 2023-03-31T06:22:41.4138210Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4138330Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4138420Z c10::string_view attr, 2023-03-31T06:22:41.4138542Z torch::List> scalars, 2023-03-31T06:22:41.4138652Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4138718Z 2023-03-31T06:22:41.4138962Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.4139036Z arg1_1.reset(); 2023-03-31T06:22:41.4139109Z arg2_1.reset(); 2023-03-31T06:22:41.4139214Z return std::vector({buf0}); 2023-03-31T06:22:41.4139278Z } 2023-03-31T06:22:41.4139345Z ''' 2023-03-31T06:22:41.4139406Z ) 2023-03-31T06:22:41.4139411Z 2023-03-31T06:22:41.4139561Z module = load_inline( 2023-03-31T06:22:41.4139856Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.4139944Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4140061Z functions=['call_29'], 2023-03-31T06:22:41.4140445Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4140607Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4141306Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4141328Z 2023-03-31T06:22:41.4141391Z def _wrap_func(f): 2023-03-31T06:22:41.4141467Z def g(args): 2023-03-31T06:22:41.4141547Z return f(args) 2023-03-31T06:22:41.4141616Z return g 2023-03-31T06:22:41.4141712Z call = _wrap_func(module.call_29) 2023-03-31T06:22:41.4141718Z 2023-03-31T06:22:41.4141723Z 2023-03-31T06:22:41.4141820Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4141942Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4142058Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4142273Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4142614Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4142842Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4142970Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4142975Z 2023-03-31T06:22:41.4142980Z 2023-03-31T06:22:41.4143059Z if __name__ == "__main__": 2023-03-31T06:22:41.4143144Z import argparse 2023-03-31T06:22:41.4143278Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4143284Z 2023-03-31T06:22:41.4143397Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4143697Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4144234Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4144746Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4144917Z args = parser.parse_args() 2023-03-31T06:22:41.4144926Z 2023-03-31T06:22:41.4145072Z if args.benchmark_kernels: 2023-03-31T06:22:41.4145285Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4145357Z else: 2023-03-31T06:22:41.4145473Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4145559Z benchmark_compiled_module() 2023-03-31T06:22:41.4145565Z 2023-03-31T06:22:41.4145635Z if p: 2023-03-31T06:22:41.4145787Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4145884Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4146023Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4146032Z 2023-03-31T06:22:41.4146318Z [2023-03-31 05:41:01,694] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 29 2023-03-31T06:22:41.4146588Z [2023-03-31 05:41:01,757] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 30 2023-03-31T06:22:41.4146997Z [2023-03-31 05:41:23,482] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/dd/cdddeaj4ep7elqyounnb46agfzyaki45inlagr2epiyjgh42y24p.py 2023-03-31T06:22:41.4147263Z [2023-03-31 05:41:23,483] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4147295Z 2023-03-31T06:22:41.4147381Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4147454Z import torch 2023-03-31T06:22:41.4147524Z import math 2023-03-31T06:22:41.4147601Z import random 2023-03-31T06:22:41.4147672Z import os 2023-03-31T06:22:41.4147749Z import tempfile 2023-03-31T06:22:41.4147854Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4147875Z 2023-03-31T06:22:41.4147980Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4148109Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4148243Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4148248Z 2023-03-31T06:22:41.4148331Z aten = torch.ops.aten 2023-03-31T06:22:41.4148473Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4148570Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4148575Z 2023-03-31T06:22:41.4148670Z async_compile.wait(globals()) 2023-03-31T06:22:41.4148734Z del async_compile 2023-03-31T06:22:41.4148858Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4148934Z wrapper = ( 2023-03-31T06:22:41.4149018Z ''' 2023-03-31T06:22:41.4149096Z #include 2023-03-31T06:22:41.4149172Z #include 2023-03-31T06:22:41.4149178Z 2023-03-31T06:22:41.4149275Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4149283Z 2023-03-31T06:22:41.4149377Z template 2023-03-31T06:22:41.4149489Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4149606Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4149734Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4149833Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4149960Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4150039Z return kernel_cpp; 2023-03-31T06:22:41.4150104Z } 2023-03-31T06:22:41.4150222Z std::vector call_30(std::vector args) { 2023-03-31T06:22:41.4150328Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4150403Z arg0_1 = args[0]; 2023-03-31T06:22:41.4150478Z arg1_1 = args[1]; 2023-03-31T06:22:41.4150549Z arg2_1 = args[2]; 2023-03-31T06:22:41.4150620Z arg3_1 = args[3]; 2023-03-31T06:22:41.4150683Z 2023-03-31T06:22:41.4150768Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4150895Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4150982Z .findSchemaOrThrow( 2023-03-31T06:22:41.4151080Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4151150Z "") 2023-03-31T06:22:41.4151219Z .typed< 2023-03-31T06:22:41.4151281Z at::Tensor( 2023-03-31T06:22:41.4151380Z const at::Tensor& input_t, 2023-03-31T06:22:41.4151480Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4151596Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4151687Z c10::string_view attr, 2023-03-31T06:22:41.4151811Z torch::List> scalars, 2023-03-31T06:22:41.4151933Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4151998Z 2023-03-31T06:22:41.4152208Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.4152281Z arg1_1.reset(); 2023-03-31T06:22:41.4152354Z arg2_1.reset(); 2023-03-31T06:22:41.4152425Z arg3_1.reset(); 2023-03-31T06:22:41.4152533Z return std::vector({buf0}); 2023-03-31T06:22:41.4152597Z } 2023-03-31T06:22:41.4152679Z ''' 2023-03-31T06:22:41.4152728Z ) 2023-03-31T06:22:41.4152733Z 2023-03-31T06:22:41.4152811Z module = load_inline( 2023-03-31T06:22:41.4153078Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.4153164Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4153280Z functions=['call_30'], 2023-03-31T06:22:41.4153704Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4153902Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4154605Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4154617Z 2023-03-31T06:22:41.4154695Z def _wrap_func(f): 2023-03-31T06:22:41.4154754Z def g(args): 2023-03-31T06:22:41.4154827Z return f(args) 2023-03-31T06:22:41.4154894Z return g 2023-03-31T06:22:41.4154991Z call = _wrap_func(module.call_30) 2023-03-31T06:22:41.4154997Z 2023-03-31T06:22:41.4155002Z 2023-03-31T06:22:41.4155099Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4155222Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4155352Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4155554Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4155764Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4155974Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4156187Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4156361Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4156367Z 2023-03-31T06:22:41.4156372Z 2023-03-31T06:22:41.4156452Z if __name__ == "__main__": 2023-03-31T06:22:41.4156529Z import argparse 2023-03-31T06:22:41.4156665Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4156671Z 2023-03-31T06:22:41.4156783Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4157082Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4157424Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4157707Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4157836Z args = parser.parse_args() 2023-03-31T06:22:41.4157841Z 2023-03-31T06:22:41.4157933Z if args.benchmark_kernels: 2023-03-31T06:22:41.4158123Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4158191Z else: 2023-03-31T06:22:41.4158306Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4158389Z benchmark_compiled_module() 2023-03-31T06:22:41.4158395Z 2023-03-31T06:22:41.4158464Z if p: 2023-03-31T06:22:41.4158615Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4158715Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4158855Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4158860Z 2023-03-31T06:22:41.4159139Z [2023-03-31 05:41:23,483] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 30 2023-03-31T06:22:41.4159411Z [2023-03-31 05:41:23,530] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 31 2023-03-31T06:22:41.4159811Z [2023-03-31 05:41:47,324] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/r7/cr7dzpb7mrzbrtm6rv4tcj36shl33xdofx2m2yv62oj4rxdtr25i.py 2023-03-31T06:22:41.4160030Z [2023-03-31 05:41:47,324] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4160037Z 2023-03-31T06:22:41.4160121Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4160194Z import torch 2023-03-31T06:22:41.4160294Z import math 2023-03-31T06:22:41.4160394Z import random 2023-03-31T06:22:41.4160465Z import os 2023-03-31T06:22:41.4160541Z import tempfile 2023-03-31T06:22:41.4160647Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4160665Z 2023-03-31T06:22:41.4160772Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4160898Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4161033Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4161038Z 2023-03-31T06:22:41.4161119Z aten = torch.ops.aten 2023-03-31T06:22:41.4161261Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4161357Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4161363Z 2023-03-31T06:22:41.4161458Z async_compile.wait(globals()) 2023-03-31T06:22:41.4161521Z del async_compile 2023-03-31T06:22:41.4161641Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4161714Z wrapper = ( 2023-03-31T06:22:41.4161799Z ''' 2023-03-31T06:22:41.4161879Z #include 2023-03-31T06:22:41.4161959Z #include 2023-03-31T06:22:41.4161964Z 2023-03-31T06:22:41.4162057Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4162062Z 2023-03-31T06:22:41.4162153Z template 2023-03-31T06:22:41.4162265Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4162351Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4162476Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4162574Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4162732Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4162813Z return kernel_cpp; 2023-03-31T06:22:41.4162879Z } 2023-03-31T06:22:41.4162996Z std::vector call_31(std::vector args) { 2023-03-31T06:22:41.4163300Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4163413Z arg0_1 = args[0]; 2023-03-31T06:22:41.4163489Z arg1_1 = args[1]; 2023-03-31T06:22:41.4163565Z arg2_1 = args[2]; 2023-03-31T06:22:41.4163635Z 2023-03-31T06:22:41.4163722Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4163826Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4163918Z .findSchemaOrThrow( 2023-03-31T06:22:41.4164023Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4164094Z "") 2023-03-31T06:22:41.4164166Z .typed< 2023-03-31T06:22:41.4164317Z at::Tensor( 2023-03-31T06:22:41.4164403Z const at::Tensor& input_t, 2023-03-31T06:22:41.4164507Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4164627Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4164718Z c10::string_view attr, 2023-03-31T06:22:41.4164841Z torch::List> scalars, 2023-03-31T06:22:41.4164966Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4165034Z 2023-03-31T06:22:41.4165281Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.4165343Z arg1_1.reset(); 2023-03-31T06:22:41.4165417Z arg2_1.reset(); 2023-03-31T06:22:41.4165524Z return std::vector({buf0}); 2023-03-31T06:22:41.4165588Z } 2023-03-31T06:22:41.4165670Z ''' 2023-03-31T06:22:41.4165731Z ) 2023-03-31T06:22:41.4165737Z 2023-03-31T06:22:41.4165818Z module = load_inline( 2023-03-31T06:22:41.4166070Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.4166157Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4166275Z functions=['call_31'], 2023-03-31T06:22:41.4166661Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4166822Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4167582Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4167623Z 2023-03-31T06:22:41.4167701Z def _wrap_func(f): 2023-03-31T06:22:41.4167773Z def g(args): 2023-03-31T06:22:41.4167850Z return f(args) 2023-03-31T06:22:41.4167906Z return g 2023-03-31T06:22:41.4168000Z call = _wrap_func(module.call_31) 2023-03-31T06:22:41.4168006Z 2023-03-31T06:22:41.4168012Z 2023-03-31T06:22:41.4168108Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4168227Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4168353Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4168573Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4168789Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4169001Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4169116Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4169121Z 2023-03-31T06:22:41.4169141Z 2023-03-31T06:22:41.4169207Z if __name__ == "__main__": 2023-03-31T06:22:41.4169286Z import argparse 2023-03-31T06:22:41.4169421Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4169427Z 2023-03-31T06:22:41.4169538Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4169888Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4170230Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4170514Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4170608Z args = parser.parse_args() 2023-03-31T06:22:41.4170614Z 2023-03-31T06:22:41.4170691Z if args.benchmark_kernels: 2023-03-31T06:22:41.4170881Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4170949Z else: 2023-03-31T06:22:41.4171063Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4171191Z benchmark_compiled_module() 2023-03-31T06:22:41.4171197Z 2023-03-31T06:22:41.4171265Z if p: 2023-03-31T06:22:41.4171417Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4171500Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4171638Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4171643Z 2023-03-31T06:22:41.4171925Z [2023-03-31 05:41:47,324] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 31 2023-03-31T06:22:41.4172195Z [2023-03-31 05:41:47,402] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 32 2023-03-31T06:22:41.4172605Z [2023-03-31 05:42:10,696] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/d3/cd3ivig5m7g4ckodzanercpphtfpqsnkkhmz65phupy3wx7w4nxm.py 2023-03-31T06:22:41.4172825Z [2023-03-31 05:42:10,696] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4172833Z 2023-03-31T06:22:41.4172930Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4173002Z import torch 2023-03-31T06:22:41.4173073Z import math 2023-03-31T06:22:41.4173135Z import random 2023-03-31T06:22:41.4173203Z import os 2023-03-31T06:22:41.4173276Z import tempfile 2023-03-31T06:22:41.4173396Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4173402Z 2023-03-31T06:22:41.4173520Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4173647Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4173828Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4173858Z 2023-03-31T06:22:41.4173928Z aten = torch.ops.aten 2023-03-31T06:22:41.4174074Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4174170Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4174175Z 2023-03-31T06:22:41.4174266Z async_compile.wait(globals()) 2023-03-31T06:22:41.4174344Z del async_compile 2023-03-31T06:22:41.4174466Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4174538Z wrapper = ( 2023-03-31T06:22:41.4174607Z ''' 2023-03-31T06:22:41.4174688Z #include 2023-03-31T06:22:41.4174766Z #include 2023-03-31T06:22:41.4174771Z 2023-03-31T06:22:41.4174864Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4174869Z 2023-03-31T06:22:41.4174962Z template 2023-03-31T06:22:41.4175086Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4175174Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4175303Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4175389Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4175517Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4175598Z return kernel_cpp; 2023-03-31T06:22:41.4175663Z } 2023-03-31T06:22:41.4175793Z std::vector call_32(std::vector args) { 2023-03-31T06:22:41.4175898Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4175973Z arg0_1 = args[0]; 2023-03-31T06:22:41.4176033Z arg1_1 = args[1]; 2023-03-31T06:22:41.4176138Z arg2_1 = args[2]; 2023-03-31T06:22:41.4176210Z arg3_1 = args[3]; 2023-03-31T06:22:41.4176273Z 2023-03-31T06:22:41.4176371Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4176465Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4176554Z .findSchemaOrThrow( 2023-03-31T06:22:41.4176640Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4176710Z "") 2023-03-31T06:22:41.4176779Z .typed< 2023-03-31T06:22:41.4176854Z at::Tensor( 2023-03-31T06:22:41.4176953Z const at::Tensor& input_t, 2023-03-31T06:22:41.4177053Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4177170Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4177246Z c10::string_view attr, 2023-03-31T06:22:41.4177414Z torch::List> scalars, 2023-03-31T06:22:41.4177539Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4177610Z 2023-03-31T06:22:41.4177843Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardswish", {-1}, ""); 2023-03-31T06:22:41.4177916Z arg1_1.reset(); 2023-03-31T06:22:41.4177987Z arg2_1.reset(); 2023-03-31T06:22:41.4178048Z arg3_1.reset(); 2023-03-31T06:22:41.4178152Z return std::vector({buf0}); 2023-03-31T06:22:41.4178215Z } 2023-03-31T06:22:41.4178299Z ''' 2023-03-31T06:22:41.4178362Z ) 2023-03-31T06:22:41.4178367Z 2023-03-31T06:22:41.4178447Z module = load_inline( 2023-03-31T06:22:41.4178724Z name='inline_extension_cudwzcjmrkzefzcijxyxj2xexarvj4uoxq3zkd6yjvof2g3iyxm6', 2023-03-31T06:22:41.4178797Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4178917Z functions=['call_32'], 2023-03-31T06:22:41.4179298Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4179458Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4180168Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4180207Z 2023-03-31T06:22:41.4180314Z def _wrap_func(f): 2023-03-31T06:22:41.4180385Z def g(args): 2023-03-31T06:22:41.4180461Z return f(args) 2023-03-31T06:22:41.4180527Z return g 2023-03-31T06:22:41.4180611Z call = _wrap_func(module.call_32) 2023-03-31T06:22:41.4180616Z 2023-03-31T06:22:41.4180621Z 2023-03-31T06:22:41.4180716Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4180834Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4180964Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4181182Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4181393Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4181602Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4181820Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4181943Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4181950Z 2023-03-31T06:22:41.4181969Z 2023-03-31T06:22:41.4182035Z if __name__ == "__main__": 2023-03-31T06:22:41.4182113Z import argparse 2023-03-31T06:22:41.4182245Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4182250Z 2023-03-31T06:22:41.4182362Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4182777Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4183159Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4183441Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4183536Z args = parser.parse_args() 2023-03-31T06:22:41.4183542Z 2023-03-31T06:22:41.4183622Z if args.benchmark_kernels: 2023-03-31T06:22:41.4183815Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4183886Z else: 2023-03-31T06:22:41.4184000Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4184098Z benchmark_compiled_module() 2023-03-31T06:22:41.4184104Z 2023-03-31T06:22:41.4184174Z if p: 2023-03-31T06:22:41.4184324Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4184442Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4184584Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4184590Z 2023-03-31T06:22:41.4184870Z [2023-03-31 05:42:10,697] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 32 2023-03-31T06:22:41.4185138Z [2023-03-31 05:42:10,754] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 33 2023-03-31T06:22:41.4185548Z [2023-03-31 05:42:40,595] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/go/cgoxw7akyolth64sj7or24dtgdpnub2fzvbxz3ixy7da7ggslh76.py 2023-03-31T06:22:41.4185772Z [2023-03-31 05:42:40,595] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4185778Z 2023-03-31T06:22:41.4185879Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4185953Z import torch 2023-03-31T06:22:41.4186025Z import math 2023-03-31T06:22:41.4186087Z import random 2023-03-31T06:22:41.4186155Z import os 2023-03-31T06:22:41.4186231Z import tempfile 2023-03-31T06:22:41.4186350Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4186358Z 2023-03-31T06:22:41.4186479Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4186606Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4186740Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4186746Z 2023-03-31T06:22:41.4186813Z aten = torch.ops.aten 2023-03-31T06:22:41.4187003Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4187128Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4187134Z 2023-03-31T06:22:41.4187224Z async_compile.wait(globals()) 2023-03-31T06:22:41.4187301Z del async_compile 2023-03-31T06:22:41.4187422Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4187497Z wrapper = ( 2023-03-31T06:22:41.4187567Z ''' 2023-03-31T06:22:41.4187645Z #include 2023-03-31T06:22:41.4187723Z #include 2023-03-31T06:22:41.4187727Z 2023-03-31T06:22:41.4187822Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4187827Z 2023-03-31T06:22:41.4187924Z template 2023-03-31T06:22:41.4188049Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4188134Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4188259Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4188343Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4188477Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4188561Z return kernel_cpp; 2023-03-31T06:22:41.4188625Z } 2023-03-31T06:22:41.4188756Z std::vector call_33(std::vector args) { 2023-03-31T06:22:41.4188852Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4188928Z arg0_1 = args[0]; 2023-03-31T06:22:41.4188987Z arg1_1 = args[1]; 2023-03-31T06:22:41.4189058Z arg2_1 = args[2]; 2023-03-31T06:22:41.4189123Z 2023-03-31T06:22:41.4189221Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4189314Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4189433Z .findSchemaOrThrow( 2023-03-31T06:22:41.4189532Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4189586Z "") 2023-03-31T06:22:41.4189654Z .typed< 2023-03-31T06:22:41.4189731Z at::Tensor( 2023-03-31T06:22:41.4189830Z const at::Tensor& input_t, 2023-03-31T06:22:41.4189932Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4190052Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4190144Z c10::string_view attr, 2023-03-31T06:22:41.4190252Z torch::List> scalars, 2023-03-31T06:22:41.4190373Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4190438Z 2023-03-31T06:22:41.4190684Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardswish", {-1}, ""); 2023-03-31T06:22:41.4190850Z arg1_1.reset(); 2023-03-31T06:22:41.4190923Z arg2_1.reset(); 2023-03-31T06:22:41.4191033Z return std::vector({buf0}); 2023-03-31T06:22:41.4191084Z } 2023-03-31T06:22:41.4191165Z ''' 2023-03-31T06:22:41.4191229Z ) 2023-03-31T06:22:41.4191234Z 2023-03-31T06:22:41.4191314Z module = load_inline( 2023-03-31T06:22:41.4191580Z name='inline_extension_cnvk2g3e2rdsnmm65pobm4q7xt5awnshzz37lydqy4gckkvhk5qf', 2023-03-31T06:22:41.4191669Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4191789Z functions=['call_33'], 2023-03-31T06:22:41.4192176Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4192319Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4193033Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4193042Z 2023-03-31T06:22:41.4193122Z def _wrap_func(f): 2023-03-31T06:22:41.4193194Z def g(args): 2023-03-31T06:22:41.4193270Z return f(args) 2023-03-31T06:22:41.4193339Z return g 2023-03-31T06:22:41.4193434Z call = _wrap_func(module.call_33) 2023-03-31T06:22:41.4193440Z 2023-03-31T06:22:41.4193475Z 2023-03-31T06:22:41.4193598Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4193716Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4193831Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4194052Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4194262Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4194484Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4194614Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4194620Z 2023-03-31T06:22:41.4194624Z 2023-03-31T06:22:41.4194704Z if __name__ == "__main__": 2023-03-31T06:22:41.4194785Z import argparse 2023-03-31T06:22:41.4194916Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4194922Z 2023-03-31T06:22:41.4195019Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4195332Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4195669Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4195951Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4196047Z args = parser.parse_args() 2023-03-31T06:22:41.4196052Z 2023-03-31T06:22:41.4196143Z if args.benchmark_kernels: 2023-03-31T06:22:41.4196362Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4196431Z else: 2023-03-31T06:22:41.4196530Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4196625Z benchmark_compiled_module() 2023-03-31T06:22:41.4196630Z 2023-03-31T06:22:41.4196699Z if p: 2023-03-31T06:22:41.4196851Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4196950Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4197087Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4197094Z 2023-03-31T06:22:41.4197373Z [2023-03-31 05:42:40,596] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 33 2023-03-31T06:22:41.4197669Z [2023-03-31 05:42:40,668] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 34 2023-03-31T06:22:41.4198076Z [2023-03-31 05:43:03,035] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/bx/cbxdm5xjwzukb75hxelyewc4wgob3c4nmdjrrj3yqiwtai5zodmw.py 2023-03-31T06:22:41.4198285Z [2023-03-31 05:43:03,035] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4198291Z 2023-03-31T06:22:41.4198387Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4198459Z import torch 2023-03-31T06:22:41.4198531Z import math 2023-03-31T06:22:41.4198605Z import random 2023-03-31T06:22:41.4198675Z import os 2023-03-31T06:22:41.4198749Z import tempfile 2023-03-31T06:22:41.4198856Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4198862Z 2023-03-31T06:22:41.4198981Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4199105Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4199242Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4199247Z 2023-03-31T06:22:41.4199328Z aten = torch.ops.aten 2023-03-31T06:22:41.4199470Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4199566Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4199571Z 2023-03-31T06:22:41.4199662Z async_compile.wait(globals()) 2023-03-31T06:22:41.4199723Z del async_compile 2023-03-31T06:22:41.4199844Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4199917Z wrapper = ( 2023-03-31T06:22:41.4199999Z ''' 2023-03-31T06:22:41.4200105Z #include 2023-03-31T06:22:41.4200221Z #include 2023-03-31T06:22:41.4200226Z 2023-03-31T06:22:41.4200318Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4200323Z 2023-03-31T06:22:41.4200403Z template 2023-03-31T06:22:41.4200526Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4200613Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4200740Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4200839Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4200966Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4201045Z return kernel_cpp; 2023-03-31T06:22:41.4201095Z } 2023-03-31T06:22:41.4201224Z std::vector call_34(std::vector args) { 2023-03-31T06:22:41.4201329Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4201404Z arg0_1 = args[0]; 2023-03-31T06:22:41.4201478Z arg1_1 = args[1]; 2023-03-31T06:22:41.4201553Z arg2_1 = args[2]; 2023-03-31T06:22:41.4201628Z arg3_1 = args[3]; 2023-03-31T06:22:41.4201678Z 2023-03-31T06:22:41.4201777Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4201871Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4201959Z .findSchemaOrThrow( 2023-03-31T06:22:41.4202057Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4202124Z "") 2023-03-31T06:22:41.4202194Z .typed< 2023-03-31T06:22:41.4202256Z at::Tensor( 2023-03-31T06:22:41.4202354Z const at::Tensor& input_t, 2023-03-31T06:22:41.4202485Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4202604Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4202692Z c10::string_view attr, 2023-03-31T06:22:41.4202814Z torch::List> scalars, 2023-03-31T06:22:41.4202935Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4202990Z 2023-03-31T06:22:41.4203400Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardswish", {-1}, ""); 2023-03-31T06:22:41.4203477Z arg1_1.reset(); 2023-03-31T06:22:41.4203550Z arg2_1.reset(); 2023-03-31T06:22:41.4203621Z arg3_1.reset(); 2023-03-31T06:22:41.4203728Z return std::vector({buf0}); 2023-03-31T06:22:41.4203793Z } 2023-03-31T06:22:41.4203860Z ''' 2023-03-31T06:22:41.4203988Z ) 2023-03-31T06:22:41.4203994Z 2023-03-31T06:22:41.4204075Z module = load_inline( 2023-03-31T06:22:41.4204357Z name='inline_extension_cudwzcjmrkzefzcijxyxj2xexarvj4uoxq3zkd6yjvof2g3iyxm6', 2023-03-31T06:22:41.4204446Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4204564Z functions=['call_34'], 2023-03-31T06:22:41.4204949Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4205107Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4205803Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4205830Z 2023-03-31T06:22:41.4205893Z def _wrap_func(f): 2023-03-31T06:22:41.4205968Z def g(args): 2023-03-31T06:22:41.4206045Z return f(args) 2023-03-31T06:22:41.4206114Z return g 2023-03-31T06:22:41.4206212Z call = _wrap_func(module.call_34) 2023-03-31T06:22:41.4206218Z 2023-03-31T06:22:41.4206222Z 2023-03-31T06:22:41.4206320Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4206440Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4206556Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4206813Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4207058Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4207266Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4207472Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4207607Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4207615Z 2023-03-31T06:22:41.4207620Z 2023-03-31T06:22:41.4207699Z if __name__ == "__main__": 2023-03-31T06:22:41.4207778Z import argparse 2023-03-31T06:22:41.4207909Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4207915Z 2023-03-31T06:22:41.4208011Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4208323Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4208664Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4208944Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4209038Z args = parser.parse_args() 2023-03-31T06:22:41.4209043Z 2023-03-31T06:22:41.4209132Z if args.benchmark_kernels: 2023-03-31T06:22:41.4209323Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4209390Z else: 2023-03-31T06:22:41.4209492Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4209626Z benchmark_compiled_module() 2023-03-31T06:22:41.4209632Z 2023-03-31T06:22:41.4209703Z if p: 2023-03-31T06:22:41.4209853Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4209947Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4210086Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4210094Z 2023-03-31T06:22:41.4210376Z [2023-03-31 05:43:03,036] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 34 2023-03-31T06:22:41.4210642Z [2023-03-31 05:43:03,082] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 35 2023-03-31T06:22:41.4211040Z [2023-03-31 05:43:25,550] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/x7/cx7e6k57saft5orvr2rqr2qrgffxcrsyn2zu53ahjmsenj7yykib.py 2023-03-31T06:22:41.4211280Z [2023-03-31 05:43:25,550] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4211285Z 2023-03-31T06:22:41.4211386Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4211458Z import torch 2023-03-31T06:22:41.4211529Z import math 2023-03-31T06:22:41.4211602Z import random 2023-03-31T06:22:41.4211670Z import os 2023-03-31T06:22:41.4211743Z import tempfile 2023-03-31T06:22:41.4211849Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4211857Z 2023-03-31T06:22:41.4211980Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4212106Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4212237Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4212243Z 2023-03-31T06:22:41.4212323Z aten = torch.ops.aten 2023-03-31T06:22:41.4212462Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4212559Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4212565Z 2023-03-31T06:22:41.4212656Z async_compile.wait(globals()) 2023-03-31T06:22:41.4212720Z del async_compile 2023-03-31T06:22:41.4212843Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4212916Z wrapper = ( 2023-03-31T06:22:41.4212998Z ''' 2023-03-31T06:22:41.4213074Z #include 2023-03-31T06:22:41.4213150Z #include 2023-03-31T06:22:41.4213155Z 2023-03-31T06:22:41.4213247Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4213252Z 2023-03-31T06:22:41.4213365Z template 2023-03-31T06:22:41.4213514Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4213600Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4213726Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4213821Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4213946Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4214028Z return kernel_cpp; 2023-03-31T06:22:41.4214078Z } 2023-03-31T06:22:41.4214207Z std::vector call_35(std::vector args) { 2023-03-31T06:22:41.4214302Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4214377Z arg0_1 = args[0]; 2023-03-31T06:22:41.4214449Z arg1_1 = args[1]; 2023-03-31T06:22:41.4214521Z arg2_1 = args[2]; 2023-03-31T06:22:41.4214582Z 2023-03-31T06:22:41.4214668Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4214762Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4214854Z .findSchemaOrThrow( 2023-03-31T06:22:41.4214957Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4215025Z "") 2023-03-31T06:22:41.4215095Z .typed< 2023-03-31T06:22:41.4215171Z at::Tensor( 2023-03-31T06:22:41.4215258Z const at::Tensor& input_t, 2023-03-31T06:22:41.4215360Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4215479Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4215568Z c10::string_view attr, 2023-03-31T06:22:41.4215716Z torch::List> scalars, 2023-03-31T06:22:41.4215839Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4215905Z 2023-03-31T06:22:41.4216151Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardswish", {-1}, ""); 2023-03-31T06:22:41.4216212Z arg1_1.reset(); 2023-03-31T06:22:41.4216283Z arg2_1.reset(); 2023-03-31T06:22:41.4216391Z return std::vector({buf0}); 2023-03-31T06:22:41.4216457Z } 2023-03-31T06:22:41.4216536Z ''' 2023-03-31T06:22:41.4216598Z ) 2023-03-31T06:22:41.4216603Z 2023-03-31T06:22:41.4216681Z module = load_inline( 2023-03-31T06:22:41.4216933Z name='inline_extension_cnvk2g3e2rdsnmm65pobm4q7xt5awnshzz37lydqy4gckkvhk5qf', 2023-03-31T06:22:41.4217019Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4217135Z functions=['call_35'], 2023-03-31T06:22:41.4217555Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4217713Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4218430Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4218439Z 2023-03-31T06:22:41.4218516Z def _wrap_func(f): 2023-03-31T06:22:41.4218588Z def g(args): 2023-03-31T06:22:41.4218650Z return f(args) 2023-03-31T06:22:41.4218718Z return g 2023-03-31T06:22:41.4218814Z call = _wrap_func(module.call_35) 2023-03-31T06:22:41.4218819Z 2023-03-31T06:22:41.4218826Z 2023-03-31T06:22:41.4218922Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4219039Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4219169Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4219384Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4219594Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4219787Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4219949Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4219980Z 2023-03-31T06:22:41.4219985Z 2023-03-31T06:22:41.4220067Z if __name__ == "__main__": 2023-03-31T06:22:41.4220145Z import argparse 2023-03-31T06:22:41.4220278Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4220283Z 2023-03-31T06:22:41.4220395Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4220708Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4221050Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4221329Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4221409Z args = parser.parse_args() 2023-03-31T06:22:41.4221415Z 2023-03-31T06:22:41.4221505Z if args.benchmark_kernels: 2023-03-31T06:22:41.4221698Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4221768Z else: 2023-03-31T06:22:41.4221883Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4221980Z benchmark_compiled_module() 2023-03-31T06:22:41.4221985Z 2023-03-31T06:22:41.4222056Z if p: 2023-03-31T06:22:41.4222207Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4222293Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4222534Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4222578Z 2023-03-31T06:22:41.4222870Z [2023-03-31 05:43:25,550] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 35 2023-03-31T06:22:41.4223141Z [2023-03-31 05:43:25,604] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 36 2023-03-31T06:22:41.4223552Z [2023-03-31 05:43:48,078] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/cw/ccwxitwoj5yrfrrzx7mwaorya7h2abvy3ayypwcqonosnmwzbvev.py 2023-03-31T06:22:41.4223776Z [2023-03-31 05:43:48,078] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4223782Z 2023-03-31T06:22:41.4223884Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4223958Z import torch 2023-03-31T06:22:41.4224017Z import math 2023-03-31T06:22:41.4224132Z import random 2023-03-31T06:22:41.4224202Z import os 2023-03-31T06:22:41.4224278Z import tempfile 2023-03-31T06:22:41.4224399Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4224407Z 2023-03-31T06:22:41.4224527Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4224653Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4224774Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4224798Z 2023-03-31T06:22:41.4224869Z aten = torch.ops.aten 2023-03-31T06:22:41.4225011Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4225112Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4225117Z 2023-03-31T06:22:41.4225210Z async_compile.wait(globals()) 2023-03-31T06:22:41.4225287Z del async_compile 2023-03-31T06:22:41.4225408Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4225481Z wrapper = ( 2023-03-31T06:22:41.4225552Z ''' 2023-03-31T06:22:41.4225629Z #include 2023-03-31T06:22:41.4225710Z #include 2023-03-31T06:22:41.4225716Z 2023-03-31T06:22:41.4225812Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4225817Z 2023-03-31T06:22:41.4225913Z template 2023-03-31T06:22:41.4226037Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4226124Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4226236Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4226334Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4226492Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4226601Z return kernel_cpp; 2023-03-31T06:22:41.4226665Z } 2023-03-31T06:22:41.4226795Z std::vector call_37(std::vector args) { 2023-03-31T06:22:41.4226900Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4226962Z arg0_1 = args[0]; 2023-03-31T06:22:41.4227035Z arg1_1 = args[1]; 2023-03-31T06:22:41.4227109Z arg2_1 = args[2]; 2023-03-31T06:22:41.4227184Z arg3_1 = args[3]; 2023-03-31T06:22:41.4227248Z 2023-03-31T06:22:41.4227348Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4227444Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4227518Z .findSchemaOrThrow( 2023-03-31T06:22:41.4227615Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4227682Z "") 2023-03-31T06:22:41.4227750Z .typed< 2023-03-31T06:22:41.4227825Z at::Tensor( 2023-03-31T06:22:41.4227923Z const at::Tensor& input_t, 2023-03-31T06:22:41.4228027Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4228134Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4228223Z c10::string_view attr, 2023-03-31T06:22:41.4228347Z torch::List> scalars, 2023-03-31T06:22:41.4228470Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4228537Z 2023-03-31T06:22:41.4228701Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4228774Z arg1_1.reset(); 2023-03-31T06:22:41.4228871Z arg2_1.reset(); 2023-03-31T06:22:41.4228929Z arg3_1.reset(); 2023-03-31T06:22:41.4229038Z return std::vector({buf0}); 2023-03-31T06:22:41.4229100Z } 2023-03-31T06:22:41.4229181Z ''' 2023-03-31T06:22:41.4229242Z ) 2023-03-31T06:22:41.4229248Z 2023-03-31T06:22:41.4229324Z module = load_inline( 2023-03-31T06:22:41.4229597Z name='inline_extension_c2hs5opcafpw4goqfmhdn4sfm5mn5hcfq3zgnku6z7hzzbefhon2', 2023-03-31T06:22:41.4229674Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4229789Z functions=['call_37'], 2023-03-31T06:22:41.4230171Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4230327Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4231069Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4231078Z 2023-03-31T06:22:41.4231155Z def _wrap_func(f): 2023-03-31T06:22:41.4231226Z def g(args): 2023-03-31T06:22:41.4231300Z return f(args) 2023-03-31T06:22:41.4231355Z return g 2023-03-31T06:22:41.4231453Z call = _wrap_func(module.call_37) 2023-03-31T06:22:41.4231461Z 2023-03-31T06:22:41.4231466Z 2023-03-31T06:22:41.4231561Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4231681Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4231810Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4232027Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4232236Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4232447Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4232649Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4232784Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4232790Z 2023-03-31T06:22:41.4232795Z 2023-03-31T06:22:41.4232872Z if __name__ == "__main__": 2023-03-31T06:22:41.4232976Z import argparse 2023-03-31T06:22:41.4233135Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4233140Z 2023-03-31T06:22:41.4233251Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4233562Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4233899Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4234183Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4234264Z args = parser.parse_args() 2023-03-31T06:22:41.4234269Z 2023-03-31T06:22:41.4234357Z if args.benchmark_kernels: 2023-03-31T06:22:41.4234546Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4234614Z else: 2023-03-31T06:22:41.4234730Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4234828Z benchmark_compiled_module() 2023-03-31T06:22:41.4234834Z 2023-03-31T06:22:41.4234902Z if p: 2023-03-31T06:22:41.4235050Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4235134Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4235273Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4235281Z 2023-03-31T06:22:41.4235561Z [2023-03-31 05:43:48,078] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 36 2023-03-31T06:22:41.4235869Z [2023-03-31 05:43:48,118] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 37 2023-03-31T06:22:41.4236273Z [2023-03-31 05:44:10,131] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/4l/c4lwowm4qv3f5wvp4c35oske65vnhrrftowatthgchotvsw2bm4a.py 2023-03-31T06:22:41.4236497Z [2023-03-31 05:44:10,131] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4236505Z 2023-03-31T06:22:41.4236603Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4236676Z import torch 2023-03-31T06:22:41.4236733Z import math 2023-03-31T06:22:41.4236807Z import random 2023-03-31T06:22:41.4236875Z import os 2023-03-31T06:22:41.4236950Z import tempfile 2023-03-31T06:22:41.4237068Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4237104Z 2023-03-31T06:22:41.4237224Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4237349Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4237472Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4237490Z 2023-03-31T06:22:41.4237557Z aten = torch.ops.aten 2023-03-31T06:22:41.4237697Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4237793Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4237798Z 2023-03-31T06:22:41.4237888Z async_compile.wait(globals()) 2023-03-31T06:22:41.4237966Z del async_compile 2023-03-31T06:22:41.4238091Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4238164Z wrapper = ( 2023-03-31T06:22:41.4238233Z ''' 2023-03-31T06:22:41.4238309Z #include 2023-03-31T06:22:41.4238384Z #include 2023-03-31T06:22:41.4238389Z 2023-03-31T06:22:41.4238483Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4238489Z 2023-03-31T06:22:41.4238584Z template 2023-03-31T06:22:41.4238707Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4238793Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4238907Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4239006Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4239130Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4239209Z return kernel_cpp; 2023-03-31T06:22:41.4239272Z } 2023-03-31T06:22:41.4239439Z std::vector call_39(std::vector args) { 2023-03-31T06:22:41.4239558Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4239619Z arg0_1 = args[0]; 2023-03-31T06:22:41.4239691Z arg1_1 = args[1]; 2023-03-31T06:22:41.4239762Z arg2_1 = args[2]; 2023-03-31T06:22:41.4239823Z 2023-03-31T06:22:41.4239921Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4240013Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4240104Z .findSchemaOrThrow( 2023-03-31T06:22:41.4240188Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4240256Z "") 2023-03-31T06:22:41.4240323Z .typed< 2023-03-31T06:22:41.4240404Z at::Tensor( 2023-03-31T06:22:41.4240502Z const at::Tensor& input_t, 2023-03-31T06:22:41.4240603Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4240720Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4240796Z c10::string_view attr, 2023-03-31T06:22:41.4240918Z torch::List> scalars, 2023-03-31T06:22:41.4241044Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4241111Z 2023-03-31T06:22:41.4241281Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4241354Z arg1_1.reset(); 2023-03-31T06:22:41.4241426Z arg2_1.reset(); 2023-03-31T06:22:41.4241536Z return std::vector({buf0}); 2023-03-31T06:22:41.4241586Z } 2023-03-31T06:22:41.4241666Z ''' 2023-03-31T06:22:41.4241729Z ) 2023-03-31T06:22:41.4241734Z 2023-03-31T06:22:41.4241875Z module = load_inline( 2023-03-31T06:22:41.4242143Z name='inline_extension_cabai5zl24epsf4tb5et443eooog5gqm5vfqz6torsz4rh3eskkg', 2023-03-31T06:22:41.4242232Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4242352Z functions=['call_39'], 2023-03-31T06:22:41.4242722Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4242881Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4243999Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4244088Z 2023-03-31T06:22:41.4244168Z def _wrap_func(f): 2023-03-31T06:22:41.4244245Z def g(args): 2023-03-31T06:22:41.4244321Z return f(args) 2023-03-31T06:22:41.4244391Z return g 2023-03-31T06:22:41.4244487Z call = _wrap_func(module.call_39) 2023-03-31T06:22:41.4244493Z 2023-03-31T06:22:41.4244498Z 2023-03-31T06:22:41.4244595Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4244699Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4244831Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4245054Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4245270Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4245488Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4245621Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4245626Z 2023-03-31T06:22:41.4245631Z 2023-03-31T06:22:41.4245712Z if __name__ == "__main__": 2023-03-31T06:22:41.4245795Z import argparse 2023-03-31T06:22:41.4245914Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4245935Z 2023-03-31T06:22:41.4246036Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4246352Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4246731Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4247058Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4247155Z args = parser.parse_args() 2023-03-31T06:22:41.4247160Z 2023-03-31T06:22:41.4247251Z if args.benchmark_kernels: 2023-03-31T06:22:41.4247444Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4247513Z else: 2023-03-31T06:22:41.4247614Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4247711Z benchmark_compiled_module() 2023-03-31T06:22:41.4247716Z 2023-03-31T06:22:41.4247787Z if p: 2023-03-31T06:22:41.4247939Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4248036Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4248178Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4248185Z 2023-03-31T06:22:41.4248468Z [2023-03-31 05:44:10,131] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 37 2023-03-31T06:22:41.4248736Z [2023-03-31 05:44:10,194] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 38 2023-03-31T06:22:41.4249130Z [2023-03-31 05:44:33,814] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/cd/ccdpdmj74t3minj3dt27ogw6zowlnpir6vox7hzzuiwpnt442ihq.py 2023-03-31T06:22:41.4249393Z [2023-03-31 05:44:33,814] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4249400Z 2023-03-31T06:22:41.4249500Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4249573Z import torch 2023-03-31T06:22:41.4249644Z import math 2023-03-31T06:22:41.4249716Z import random 2023-03-31T06:22:41.4249784Z import os 2023-03-31T06:22:41.4249845Z import tempfile 2023-03-31T06:22:41.4249965Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4249973Z 2023-03-31T06:22:41.4250093Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4250219Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4250351Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4250357Z 2023-03-31T06:22:41.4250438Z aten = torch.ops.aten 2023-03-31T06:22:41.4250576Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4250699Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4250704Z 2023-03-31T06:22:41.4250781Z async_compile.wait(globals()) 2023-03-31T06:22:41.4250859Z del async_compile 2023-03-31T06:22:41.4250980Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4251054Z wrapper = ( 2023-03-31T06:22:41.4251138Z ''' 2023-03-31T06:22:41.4251214Z #include 2023-03-31T06:22:41.4251291Z #include 2023-03-31T06:22:41.4251296Z 2023-03-31T06:22:41.4251377Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4251398Z 2023-03-31T06:22:41.4251480Z template 2023-03-31T06:22:41.4251603Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4251689Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4251815Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4251912Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4252038Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4252121Z return kernel_cpp; 2023-03-31T06:22:41.4252172Z } 2023-03-31T06:22:41.4252303Z std::vector call_41(std::vector args) { 2023-03-31T06:22:41.4252406Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4252481Z arg0_1 = args[0]; 2023-03-31T06:22:41.4252553Z arg1_1 = args[1]; 2023-03-31T06:22:41.4252624Z arg2_1 = args[2]; 2023-03-31T06:22:41.4252682Z arg3_1 = args[3]; 2023-03-31T06:22:41.4252744Z 2023-03-31T06:22:41.4252843Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4252964Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4253081Z .findSchemaOrThrow( 2023-03-31T06:22:41.4253181Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4253248Z "") 2023-03-31T06:22:41.4253302Z .typed< 2023-03-31T06:22:41.4253377Z at::Tensor( 2023-03-31T06:22:41.4253477Z const at::Tensor& input_t, 2023-03-31T06:22:41.4253580Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4253697Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4253790Z c10::string_view attr, 2023-03-31T06:22:41.4253915Z torch::List> scalars, 2023-03-31T06:22:41.4254035Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4254088Z 2023-03-31T06:22:41.4254249Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4254324Z arg1_1.reset(); 2023-03-31T06:22:41.4254399Z arg2_1.reset(); 2023-03-31T06:22:41.4254469Z arg3_1.reset(); 2023-03-31T06:22:41.4254573Z return std::vector({buf0}); 2023-03-31T06:22:41.4254637Z } 2023-03-31T06:22:41.4254707Z ''' 2023-03-31T06:22:41.4254769Z ) 2023-03-31T06:22:41.4254775Z 2023-03-31T06:22:41.4254853Z module = load_inline( 2023-03-31T06:22:41.4255128Z name='inline_extension_c2hs5opcafpw4goqfmhdn4sfm5mn5hcfq3zgnku6z7hzzbefhon2', 2023-03-31T06:22:41.4255217Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4255332Z functions=['call_41'], 2023-03-31T06:22:41.4255743Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4255885Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4256591Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4256614Z 2023-03-31T06:22:41.4256677Z def _wrap_func(f): 2023-03-31T06:22:41.4256749Z def g(args): 2023-03-31T06:22:41.4256823Z return f(args) 2023-03-31T06:22:41.4256924Z return g 2023-03-31T06:22:41.4257018Z call = _wrap_func(module.call_41) 2023-03-31T06:22:41.4257024Z 2023-03-31T06:22:41.4257028Z 2023-03-31T06:22:41.4257124Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4257246Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4257361Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4257578Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4257784Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4257997Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4258206Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4258341Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4258347Z 2023-03-31T06:22:41.4258352Z 2023-03-31T06:22:41.4258435Z if __name__ == "__main__": 2023-03-31T06:22:41.4258516Z import argparse 2023-03-31T06:22:41.4258635Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4258658Z 2023-03-31T06:22:41.4258757Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4259066Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4259400Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4259707Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4259829Z args = parser.parse_args() 2023-03-31T06:22:41.4259835Z 2023-03-31T06:22:41.4259924Z if args.benchmark_kernels: 2023-03-31T06:22:41.4260116Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4260184Z else: 2023-03-31T06:22:41.4260283Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4260383Z benchmark_compiled_module() 2023-03-31T06:22:41.4260389Z 2023-03-31T06:22:41.4260457Z if p: 2023-03-31T06:22:41.4260610Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4260708Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4260846Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4260852Z 2023-03-31T06:22:41.4261134Z [2023-03-31 05:44:33,815] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 38 2023-03-31T06:22:41.4261406Z [2023-03-31 05:44:33,899] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 39 2023-03-31T06:22:41.4261793Z [2023-03-31 05:44:57,966] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/re/creojtn5vlc6jt6s6dxbe4kpu7nntcs7ug4ixwym3ljemz43tign.py 2023-03-31T06:22:41.4262018Z [2023-03-31 05:44:57,966] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4262026Z 2023-03-31T06:22:41.4262124Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4262198Z import torch 2023-03-31T06:22:41.4262302Z import math 2023-03-31T06:22:41.4262472Z import random 2023-03-31T06:22:41.4262548Z import os 2023-03-31T06:22:41.4262609Z import tempfile 2023-03-31T06:22:41.4262729Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4262734Z 2023-03-31T06:22:41.4262853Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4262985Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4263121Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4263127Z 2023-03-31T06:22:41.4263209Z aten = torch.ops.aten 2023-03-31T06:22:41.4263349Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4263445Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4263451Z 2023-03-31T06:22:41.4263529Z async_compile.wait(globals()) 2023-03-31T06:22:41.4263640Z del async_compile 2023-03-31T06:22:41.4263764Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4263836Z wrapper = ( 2023-03-31T06:22:41.4263926Z ''' 2023-03-31T06:22:41.4264004Z #include 2023-03-31T06:22:41.4264082Z #include 2023-03-31T06:22:41.4264089Z 2023-03-31T06:22:41.4264169Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4264188Z 2023-03-31T06:22:41.4264269Z template 2023-03-31T06:22:41.4264393Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4264482Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4264611Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4264707Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4264895Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4279900Z return kernel_cpp; 2023-03-31T06:22:41.4280027Z } 2023-03-31T06:22:41.4280196Z std::vector call_43(std::vector args) { 2023-03-31T06:22:41.4280306Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4280384Z arg0_1 = args[0]; 2023-03-31T06:22:41.4280459Z arg1_1 = args[1]; 2023-03-31T06:22:41.4280536Z arg2_1 = args[2]; 2023-03-31T06:22:41.4280586Z 2023-03-31T06:22:41.4280692Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4280791Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4280882Z .findSchemaOrThrow( 2023-03-31T06:22:41.4280984Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4281055Z "") 2023-03-31T06:22:41.4281230Z .typed< 2023-03-31T06:22:41.4281357Z at::Tensor( 2023-03-31T06:22:41.4281461Z const at::Tensor& input_t, 2023-03-31T06:22:41.4281563Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4281684Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4281779Z c10::string_view attr, 2023-03-31T06:22:41.4281908Z torch::List> scalars, 2023-03-31T06:22:41.4282035Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4282104Z 2023-03-31T06:22:41.4282271Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4282346Z arg1_1.reset(); 2023-03-31T06:22:41.4282419Z arg2_1.reset(); 2023-03-31T06:22:41.4282526Z return std::vector({buf0}); 2023-03-31T06:22:41.4282591Z } 2023-03-31T06:22:41.4282720Z ''' 2023-03-31T06:22:41.4282784Z ) 2023-03-31T06:22:41.4282795Z 2023-03-31T06:22:41.4282863Z module = load_inline( 2023-03-31T06:22:41.4283298Z name='inline_extension_cabai5zl24epsf4tb5et443eooog5gqm5vfqz6torsz4rh3eskkg', 2023-03-31T06:22:41.4283387Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4283509Z functions=['call_43'], 2023-03-31T06:22:41.4283896Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4284056Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4284846Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4284854Z 2023-03-31T06:22:41.4284937Z def _wrap_func(f): 2023-03-31T06:22:41.4285013Z def g(args): 2023-03-31T06:22:41.4285076Z return f(args) 2023-03-31T06:22:41.4285146Z return g 2023-03-31T06:22:41.4285244Z call = _wrap_func(module.call_43) 2023-03-31T06:22:41.4285250Z 2023-03-31T06:22:41.4285255Z 2023-03-31T06:22:41.4285352Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4285473Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4285651Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4285868Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4286081Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4286275Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4286404Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4286409Z 2023-03-31T06:22:41.4286414Z 2023-03-31T06:22:41.4286495Z if __name__ == "__main__": 2023-03-31T06:22:41.4286576Z import argparse 2023-03-31T06:22:41.4286707Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4286713Z 2023-03-31T06:22:41.4286825Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4287142Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4287485Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4287762Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4287843Z args = parser.parse_args() 2023-03-31T06:22:41.4287848Z 2023-03-31T06:22:41.4287937Z if args.benchmark_kernels: 2023-03-31T06:22:41.4288130Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4288199Z else: 2023-03-31T06:22:41.4288356Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4288487Z benchmark_compiled_module() 2023-03-31T06:22:41.4288493Z 2023-03-31T06:22:41.4288560Z if p: 2023-03-31T06:22:41.4288697Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4288793Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4288930Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4288939Z 2023-03-31T06:22:41.4289223Z [2023-03-31 05:44:57,966] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 39 2023-03-31T06:22:41.4289492Z [2023-03-31 05:44:58,045] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 40 2023-03-31T06:22:41.4289900Z [2023-03-31 05:45:19,491] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/pu/cpu2ahxvujrjwo3bipi7ddl3df4ovu2he6cl5dkw6y4qd2fctch6.py 2023-03-31T06:22:41.4290122Z [2023-03-31 05:45:19,491] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4290131Z 2023-03-31T06:22:41.4290229Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4290302Z import torch 2023-03-31T06:22:41.4290360Z import math 2023-03-31T06:22:41.4290434Z import random 2023-03-31T06:22:41.4290505Z import os 2023-03-31T06:22:41.4290581Z import tempfile 2023-03-31T06:22:41.4290700Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4290708Z 2023-03-31T06:22:41.4290830Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4290984Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4291106Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4291111Z 2023-03-31T06:22:41.4291192Z aten = torch.ops.aten 2023-03-31T06:22:41.4291330Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4291425Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4291430Z 2023-03-31T06:22:41.4291527Z async_compile.wait(globals()) 2023-03-31T06:22:41.4291603Z del async_compile 2023-03-31T06:22:41.4291724Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4291796Z wrapper = ( 2023-03-31T06:22:41.4291867Z ''' 2023-03-31T06:22:41.4291943Z #include 2023-03-31T06:22:41.4292020Z #include 2023-03-31T06:22:41.4292025Z 2023-03-31T06:22:41.4292118Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4292153Z 2023-03-31T06:22:41.4292246Z template 2023-03-31T06:22:41.4292370Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4292460Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4292573Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4292672Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4292797Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4292876Z return kernel_cpp; 2023-03-31T06:22:41.4292938Z } 2023-03-31T06:22:41.4293070Z std::vector call_45(std::vector args) { 2023-03-31T06:22:41.4293176Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4293237Z arg0_1 = args[0]; 2023-03-31T06:22:41.4293310Z arg1_1 = args[1]; 2023-03-31T06:22:41.4293381Z arg2_1 = args[2]; 2023-03-31T06:22:41.4293452Z arg3_1 = args[3]; 2023-03-31T06:22:41.4293514Z 2023-03-31T06:22:41.4293612Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4293708Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4293782Z .findSchemaOrThrow( 2023-03-31T06:22:41.4293881Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4293949Z "") 2023-03-31T06:22:41.4294018Z .typed< 2023-03-31T06:22:41.4294093Z at::Tensor( 2023-03-31T06:22:41.4294190Z const at::Tensor& input_t, 2023-03-31T06:22:41.4294290Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4294395Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4294509Z c10::string_view attr, 2023-03-31T06:22:41.4294657Z torch::List> scalars, 2023-03-31T06:22:41.4294779Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4294845Z 2023-03-31T06:22:41.4295106Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4295181Z arg1_1.reset(); 2023-03-31T06:22:41.4295240Z arg2_1.reset(); 2023-03-31T06:22:41.4295309Z arg3_1.reset(); 2023-03-31T06:22:41.4295416Z return std::vector({buf0}); 2023-03-31T06:22:41.4295481Z } 2023-03-31T06:22:41.4295562Z ''' 2023-03-31T06:22:41.4295623Z ) 2023-03-31T06:22:41.4295629Z 2023-03-31T06:22:41.4295705Z module = load_inline( 2023-03-31T06:22:41.4295962Z name='inline_extension_cow6quriqr7bswivrijfisjdrcn3dcra3pq6fb32ucqf4e4yfsyt', 2023-03-31T06:22:41.4296048Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4296163Z functions=['call_45'], 2023-03-31T06:22:41.4296545Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4296702Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4297433Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4297443Z 2023-03-31T06:22:41.4297520Z def _wrap_func(f): 2023-03-31T06:22:41.4297591Z def g(args): 2023-03-31T06:22:41.4297665Z return f(args) 2023-03-31T06:22:41.4297720Z return g 2023-03-31T06:22:41.4297813Z call = _wrap_func(module.call_45) 2023-03-31T06:22:41.4297819Z 2023-03-31T06:22:41.4297824Z 2023-03-31T06:22:41.4297921Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4298041Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4298169Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4298386Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4298591Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4298831Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4299036Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4299171Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4299177Z 2023-03-31T06:22:41.4299182Z 2023-03-31T06:22:41.4299262Z if __name__ == "__main__": 2023-03-31T06:22:41.4299338Z import argparse 2023-03-31T06:22:41.4299471Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4299477Z 2023-03-31T06:22:41.4299589Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4299907Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4300246Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4300523Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4300603Z args = parser.parse_args() 2023-03-31T06:22:41.4300608Z 2023-03-31T06:22:41.4300703Z if args.benchmark_kernels: 2023-03-31T06:22:41.4300893Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4300963Z else: 2023-03-31T06:22:41.4301075Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4301171Z benchmark_compiled_module() 2023-03-31T06:22:41.4301177Z 2023-03-31T06:22:41.4301272Z if p: 2023-03-31T06:22:41.4301434Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4301530Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4301667Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4301673Z 2023-03-31T06:22:41.4301954Z [2023-03-31 05:45:19,492] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 40 2023-03-31T06:22:41.4302224Z [2023-03-31 05:45:19,533] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 41 2023-03-31T06:22:41.4302740Z [2023-03-31 05:45:40,366] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/e3/ce3ked7pfgylqfjbbbt7xevonuwhzkcvi7oeealw24o2rbzhftwk.py 2023-03-31T06:22:41.4302967Z [2023-03-31 05:45:40,367] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4302973Z 2023-03-31T06:22:41.4303071Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4303149Z import torch 2023-03-31T06:22:41.4303209Z import math 2023-03-31T06:22:41.4303284Z import random 2023-03-31T06:22:41.4303353Z import os 2023-03-31T06:22:41.4303429Z import tempfile 2023-03-31T06:22:41.4303547Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4303553Z 2023-03-31T06:22:41.4303675Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4303802Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4303926Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4303946Z 2023-03-31T06:22:41.4304051Z aten = torch.ops.aten 2023-03-31T06:22:41.4304193Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4304288Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4304294Z 2023-03-31T06:22:41.4304386Z async_compile.wait(globals()) 2023-03-31T06:22:41.4304462Z del async_compile 2023-03-31T06:22:41.4304583Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4304658Z wrapper = ( 2023-03-31T06:22:41.4304730Z ''' 2023-03-31T06:22:41.4304806Z #include 2023-03-31T06:22:41.4304883Z #include 2023-03-31T06:22:41.4304888Z 2023-03-31T06:22:41.4304981Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4304987Z 2023-03-31T06:22:41.4305079Z template 2023-03-31T06:22:41.4305203Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4305363Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4305474Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4305572Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4305701Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4305782Z return kernel_cpp; 2023-03-31T06:22:41.4305847Z } 2023-03-31T06:22:41.4305977Z std::vector call_47(std::vector args) { 2023-03-31T06:22:41.4306072Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4306134Z arg0_1 = args[0]; 2023-03-31T06:22:41.4306211Z arg1_1 = args[1]; 2023-03-31T06:22:41.4306288Z arg2_1 = args[2]; 2023-03-31T06:22:41.4306354Z 2023-03-31T06:22:41.4306453Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4306548Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4306637Z .findSchemaOrThrow( 2023-03-31T06:22:41.4306723Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4306793Z "") 2023-03-31T06:22:41.4306863Z .typed< 2023-03-31T06:22:41.4306939Z at::Tensor( 2023-03-31T06:22:41.4307039Z const at::Tensor& input_t, 2023-03-31T06:22:41.4307142Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4307261Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4307337Z c10::string_view attr, 2023-03-31T06:22:41.4307458Z torch::List> scalars, 2023-03-31T06:22:41.4307580Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4307674Z 2023-03-31T06:22:41.4307983Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4308057Z arg1_1.reset(); 2023-03-31T06:22:41.4308126Z arg2_1.reset(); 2023-03-31T06:22:41.4308232Z return std::vector({buf0}); 2023-03-31T06:22:41.4308283Z } 2023-03-31T06:22:41.4308365Z ''' 2023-03-31T06:22:41.4308429Z ) 2023-03-31T06:22:41.4308435Z 2023-03-31T06:22:41.4308514Z module = load_inline( 2023-03-31T06:22:41.4308784Z name='inline_extension_cgzzjygjdtyaeq5ee6eazo4mwcm6w5k4l5xxnkhipg3d5ktdbcnm', 2023-03-31T06:22:41.4308872Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4308987Z functions=['call_47'], 2023-03-31T06:22:41.4309356Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4309510Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4310218Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4310227Z 2023-03-31T06:22:41.4310305Z def _wrap_func(f): 2023-03-31T06:22:41.4310377Z def g(args): 2023-03-31T06:22:41.4310453Z return f(args) 2023-03-31T06:22:41.4310520Z return g 2023-03-31T06:22:41.4310643Z call = _wrap_func(module.call_47) 2023-03-31T06:22:41.4310650Z 2023-03-31T06:22:41.4310655Z 2023-03-31T06:22:41.4310750Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4310855Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4310984Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4311201Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4311415Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4311631Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4311759Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4311764Z 2023-03-31T06:22:41.4311769Z 2023-03-31T06:22:41.4311875Z if __name__ == "__main__": 2023-03-31T06:22:41.4311952Z import argparse 2023-03-31T06:22:41.4312069Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4312074Z 2023-03-31T06:22:41.4312187Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4312496Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4312832Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4313114Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4313209Z args = parser.parse_args() 2023-03-31T06:22:41.4313215Z 2023-03-31T06:22:41.4313304Z if args.benchmark_kernels: 2023-03-31T06:22:41.4313494Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4313548Z else: 2023-03-31T06:22:41.4313662Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4313757Z benchmark_compiled_module() 2023-03-31T06:22:41.4313762Z 2023-03-31T06:22:41.4313833Z if p: 2023-03-31T06:22:41.4313982Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4314078Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4314215Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4314220Z 2023-03-31T06:22:41.4314527Z [2023-03-31 05:45:40,367] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 41 2023-03-31T06:22:41.4314807Z [2023-03-31 05:45:40,432] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 42 2023-03-31T06:22:41.4315210Z [2023-03-31 05:46:04,370] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/tf/ctfnvex4v3c6au2trmwcj7fec2c7eciopzowk6lmjrikckovcsfv.py 2023-03-31T06:22:41.4315434Z [2023-03-31 05:46:04,370] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4315439Z 2023-03-31T06:22:41.4315537Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4315611Z import torch 2023-03-31T06:22:41.4315683Z import math 2023-03-31T06:22:41.4315757Z import random 2023-03-31T06:22:41.4315826Z import os 2023-03-31T06:22:41.4315887Z import tempfile 2023-03-31T06:22:41.4316005Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4316010Z 2023-03-31T06:22:41.4316131Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4316260Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4316396Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4316401Z 2023-03-31T06:22:41.4316483Z aten = torch.ops.aten 2023-03-31T06:22:41.4316624Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4316721Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4316727Z 2023-03-31T06:22:41.4316807Z async_compile.wait(globals()) 2023-03-31T06:22:41.4316882Z del async_compile 2023-03-31T06:22:41.4317004Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4317104Z wrapper = ( 2023-03-31T06:22:41.4317187Z ''' 2023-03-31T06:22:41.4317263Z #include 2023-03-31T06:22:41.4317339Z #include 2023-03-31T06:22:41.4317344Z 2023-03-31T06:22:41.4317423Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4317428Z 2023-03-31T06:22:41.4317520Z template 2023-03-31T06:22:41.4317646Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4317735Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4317862Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4317960Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4318086Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4318152Z return kernel_cpp; 2023-03-31T06:22:41.4318214Z } 2023-03-31T06:22:41.4318376Z std::vector call_49(std::vector args) { 2023-03-31T06:22:41.4318479Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4318554Z arg0_1 = args[0]; 2023-03-31T06:22:41.4318630Z arg1_1 = args[1]; 2023-03-31T06:22:41.4318702Z arg2_1 = args[2]; 2023-03-31T06:22:41.4318761Z arg3_1 = args[3]; 2023-03-31T06:22:41.4318824Z 2023-03-31T06:22:41.4318926Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4319022Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4319113Z .findSchemaOrThrow( 2023-03-31T06:22:41.4319216Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4319287Z "") 2023-03-31T06:22:41.4319342Z .typed< 2023-03-31T06:22:41.4319419Z at::Tensor( 2023-03-31T06:22:41.4319519Z const at::Tensor& input_t, 2023-03-31T06:22:41.4319621Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4319742Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4319835Z c10::string_view attr, 2023-03-31T06:22:41.4319959Z torch::List> scalars, 2023-03-31T06:22:41.4320086Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4320139Z 2023-03-31T06:22:41.4320398Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4320473Z arg1_1.reset(); 2023-03-31T06:22:41.4320547Z arg2_1.reset(); 2023-03-31T06:22:41.4320619Z arg3_1.reset(); 2023-03-31T06:22:41.4320756Z return std::vector({buf0}); 2023-03-31T06:22:41.4320847Z } 2023-03-31T06:22:41.4320915Z ''' 2023-03-31T06:22:41.4320977Z ) 2023-03-31T06:22:41.4320982Z 2023-03-31T06:22:41.4321061Z module = load_inline( 2023-03-31T06:22:41.4321335Z name='inline_extension_cow6quriqr7bswivrijfisjdrcn3dcra3pq6fb32ucqf4e4yfsyt', 2023-03-31T06:22:41.4321422Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4321540Z functions=['call_49'], 2023-03-31T06:22:41.4321923Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4322068Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4322784Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4322808Z 2023-03-31T06:22:41.4322871Z def _wrap_func(f): 2023-03-31T06:22:41.4322944Z def g(args): 2023-03-31T06:22:41.4323189Z return f(args) 2023-03-31T06:22:41.4323298Z return g 2023-03-31T06:22:41.4323447Z call = _wrap_func(module.call_49) 2023-03-31T06:22:41.4323453Z 2023-03-31T06:22:41.4323457Z 2023-03-31T06:22:41.4323562Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4323684Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4323866Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4324099Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4324308Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4324519Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4324727Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4324866Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4324872Z 2023-03-31T06:22:41.4324877Z 2023-03-31T06:22:41.4324958Z if __name__ == "__main__": 2023-03-31T06:22:41.4325038Z import argparse 2023-03-31T06:22:41.4325157Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4325202Z 2023-03-31T06:22:41.4325317Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4325629Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4325970Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4326249Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4326346Z args = parser.parse_args() 2023-03-31T06:22:41.4326353Z 2023-03-31T06:22:41.4326447Z if args.benchmark_kernels: 2023-03-31T06:22:41.4326637Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4326692Z else: 2023-03-31T06:22:41.4326804Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4326901Z benchmark_compiled_module() 2023-03-31T06:22:41.4326907Z 2023-03-31T06:22:41.4326979Z if p: 2023-03-31T06:22:41.4327130Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4327228Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4327368Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4327374Z 2023-03-31T06:22:41.4327652Z [2023-03-31 05:46:04,371] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 42 2023-03-31T06:22:41.4327956Z [2023-03-31 05:46:04,412] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 43 2023-03-31T06:22:41.4328399Z [2023-03-31 05:46:25,847] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/x6/cx6l5cahcgsofc6hybewu5ylqrgpgiizpoihvemdup5ivee6pka2.py 2023-03-31T06:22:41.4328623Z [2023-03-31 05:46:25,847] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4328628Z 2023-03-31T06:22:41.4328727Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4328802Z import torch 2023-03-31T06:22:41.4328873Z import math 2023-03-31T06:22:41.4328946Z import random 2023-03-31T06:22:41.4329013Z import os 2023-03-31T06:22:41.4329078Z import tempfile 2023-03-31T06:22:41.4329197Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4329203Z 2023-03-31T06:22:41.4329324Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4329456Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4329591Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4329597Z 2023-03-31T06:22:41.4329682Z aten = torch.ops.aten 2023-03-31T06:22:41.4329827Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4329924Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4329930Z 2023-03-31T06:22:41.4330008Z async_compile.wait(globals()) 2023-03-31T06:22:41.4330085Z del async_compile 2023-03-31T06:22:41.4330208Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4330284Z wrapper = ( 2023-03-31T06:22:41.4330368Z ''' 2023-03-31T06:22:41.4330446Z #include 2023-03-31T06:22:41.4330524Z #include 2023-03-31T06:22:41.4330529Z 2023-03-31T06:22:41.4330644Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4330650Z 2023-03-31T06:22:41.4330749Z template 2023-03-31T06:22:41.4330874Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4330960Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4331085Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4331185Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4331314Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4331378Z return kernel_cpp; 2023-03-31T06:22:41.4331441Z } 2023-03-31T06:22:41.4331574Z std::vector call_51(std::vector args) { 2023-03-31T06:22:41.4331668Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4331742Z arg0_1 = args[0]; 2023-03-31T06:22:41.4331842Z arg1_1 = args[1]; 2023-03-31T06:22:41.4331913Z arg2_1 = args[2]; 2023-03-31T06:22:41.4331962Z 2023-03-31T06:22:41.4332063Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4332159Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4332247Z .findSchemaOrThrow( 2023-03-31T06:22:41.4332347Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4332414Z "") 2023-03-31T06:22:41.4332485Z .typed< 2023-03-31T06:22:41.4332547Z at::Tensor( 2023-03-31T06:22:41.4332645Z const at::Tensor& input_t, 2023-03-31T06:22:41.4332747Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4332866Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4332956Z c10::string_view attr, 2023-03-31T06:22:41.4333077Z torch::List> scalars, 2023-03-31T06:22:41.4333198Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4333266Z 2023-03-31T06:22:41.4333527Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4333603Z arg1_1.reset(); 2023-03-31T06:22:41.4333675Z arg2_1.reset(); 2023-03-31T06:22:41.4333779Z return std::vector({buf0}); 2023-03-31T06:22:41.4333841Z } 2023-03-31T06:22:41.4333921Z ''' 2023-03-31T06:22:41.4333982Z ) 2023-03-31T06:22:41.4333987Z 2023-03-31T06:22:41.4334052Z module = load_inline( 2023-03-31T06:22:41.4334352Z name='inline_extension_cgzzjygjdtyaeq5ee6eazo4mwcm6w5k4l5xxnkhipg3d5ktdbcnm', 2023-03-31T06:22:41.4334476Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4334591Z functions=['call_51'], 2023-03-31T06:22:41.4334982Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4335137Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4335840Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4335848Z 2023-03-31T06:22:41.4335924Z def _wrap_func(f): 2023-03-31T06:22:41.4335994Z def g(args): 2023-03-31T06:22:41.4336057Z return f(args) 2023-03-31T06:22:41.4336124Z return g 2023-03-31T06:22:41.4336221Z call = _wrap_func(module.call_51) 2023-03-31T06:22:41.4336229Z 2023-03-31T06:22:41.4336233Z 2023-03-31T06:22:41.4336328Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4336452Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4336582Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4336801Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4337001Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4337241Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4337373Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4337378Z 2023-03-31T06:22:41.4337382Z 2023-03-31T06:22:41.4337460Z if __name__ == "__main__": 2023-03-31T06:22:41.4337536Z import argparse 2023-03-31T06:22:41.4337669Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4337677Z 2023-03-31T06:22:41.4337791Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4338104Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4338441Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4338735Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4338829Z args = parser.parse_args() 2023-03-31T06:22:41.4338837Z 2023-03-31T06:22:41.4338926Z if args.benchmark_kernels: 2023-03-31T06:22:41.4339117Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4339184Z else: 2023-03-31T06:22:41.4339296Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4339392Z benchmark_compiled_module() 2023-03-31T06:22:41.4339398Z 2023-03-31T06:22:41.4339468Z if p: 2023-03-31T06:22:41.4339606Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4339703Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4339838Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4339844Z 2023-03-31T06:22:41.4340123Z [2023-03-31 05:46:25,848] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 43 2023-03-31T06:22:41.4340393Z [2023-03-31 05:46:25,903] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 44 2023-03-31T06:22:41.4340796Z [2023-03-31 05:46:50,719] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/p3/cp3phvnqobwh2p4qc7i62xgxlbv4zhqx23lm5xsfgvmrvbefbyby.py 2023-03-31T06:22:41.4341015Z [2023-03-31 05:46:50,719] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4341022Z 2023-03-31T06:22:41.4341148Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4341249Z import torch 2023-03-31T06:22:41.4341307Z import math 2023-03-31T06:22:41.4341380Z import random 2023-03-31T06:22:41.4341448Z import os 2023-03-31T06:22:41.4341523Z import tempfile 2023-03-31T06:22:41.4341640Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4341646Z 2023-03-31T06:22:41.4341766Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4341894Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4342013Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4342019Z 2023-03-31T06:22:41.4342102Z aten = torch.ops.aten 2023-03-31T06:22:41.4342240Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4342335Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4342341Z 2023-03-31T06:22:41.4342522Z async_compile.wait(globals()) 2023-03-31T06:22:41.4342602Z del async_compile 2023-03-31T06:22:41.4342728Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4342790Z wrapper = ( 2023-03-31T06:22:41.4342877Z ''' 2023-03-31T06:22:41.4342954Z #include 2023-03-31T06:22:41.4343031Z #include 2023-03-31T06:22:41.4343036Z 2023-03-31T06:22:41.4343132Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4343137Z 2023-03-31T06:22:41.4343229Z template 2023-03-31T06:22:41.4343354Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4343446Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4343558Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4343695Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4343825Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4343904Z return kernel_cpp; 2023-03-31T06:22:41.4343967Z } 2023-03-31T06:22:41.4344096Z std::vector call_52(std::vector args) { 2023-03-31T06:22:41.4344206Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4344269Z arg0_1 = args[0]; 2023-03-31T06:22:41.4344345Z arg1_1 = args[1]; 2023-03-31T06:22:41.4344415Z arg2_1 = args[2]; 2023-03-31T06:22:41.4344486Z arg3_1 = args[3]; 2023-03-31T06:22:41.4344550Z 2023-03-31T06:22:41.4344649Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4344730Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4344817Z .findSchemaOrThrow( 2023-03-31T06:22:41.4344948Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4345017Z "") 2023-03-31T06:22:41.4345087Z .typed< 2023-03-31T06:22:41.4345166Z at::Tensor( 2023-03-31T06:22:41.4345270Z const at::Tensor& input_t, 2023-03-31T06:22:41.4345358Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4345478Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4345566Z c10::string_view attr, 2023-03-31T06:22:41.4345686Z torch::List> scalars, 2023-03-31T06:22:41.4345810Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4345878Z 2023-03-31T06:22:41.4346103Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "none"); 2023-03-31T06:22:41.4346176Z arg1_1.reset(); 2023-03-31T06:22:41.4346234Z arg2_1.reset(); 2023-03-31T06:22:41.4346305Z arg3_1.reset(); 2023-03-31T06:22:41.4346410Z return std::vector({buf0}); 2023-03-31T06:22:41.4346477Z } 2023-03-31T06:22:41.4346557Z ''' 2023-03-31T06:22:41.4346619Z ) 2023-03-31T06:22:41.4346625Z 2023-03-31T06:22:41.4346702Z module = load_inline( 2023-03-31T06:22:41.4346964Z name='inline_extension_cwidvkkn5ycr44b4b5libljko5pjbaxwy2dstyiheyapho2qbgzi', 2023-03-31T06:22:41.4347051Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4347167Z functions=['call_52'], 2023-03-31T06:22:41.4347585Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4347770Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4348477Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4348487Z 2023-03-31T06:22:41.4348561Z def _wrap_func(f): 2023-03-31T06:22:41.4348633Z def g(args): 2023-03-31T06:22:41.4348697Z return f(args) 2023-03-31T06:22:41.4348764Z return g 2023-03-31T06:22:41.4348859Z call = _wrap_func(module.call_52) 2023-03-31T06:22:41.4348865Z 2023-03-31T06:22:41.4348870Z 2023-03-31T06:22:41.4348964Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4349085Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4349217Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4349433Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4349642Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4349838Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4350059Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4350198Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4350204Z 2023-03-31T06:22:41.4350246Z 2023-03-31T06:22:41.4350330Z if __name__ == "__main__": 2023-03-31T06:22:41.4350407Z import argparse 2023-03-31T06:22:41.4350538Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4350545Z 2023-03-31T06:22:41.4350655Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4350971Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4351311Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4351578Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4351672Z args = parser.parse_args() 2023-03-31T06:22:41.4351717Z 2023-03-31T06:22:41.4351811Z if args.benchmark_kernels: 2023-03-31T06:22:41.4352003Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4352075Z else: 2023-03-31T06:22:41.4352190Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4352288Z benchmark_compiled_module() 2023-03-31T06:22:41.4352294Z 2023-03-31T06:22:41.4352365Z if p: 2023-03-31T06:22:41.4352501Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4352598Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4352738Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4352744Z 2023-03-31T06:22:41.4353025Z [2023-03-31 05:46:50,720] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 44 2023-03-31T06:22:41.4353293Z [2023-03-31 05:46:50,761] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 45 2023-03-31T06:22:41.4353703Z [2023-03-31 05:47:15,887] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/5m/c5mj24wltwcp2ryr6wogyu7oyp7xphwcussrslkwyxiqhd4egzqr.py 2023-03-31T06:22:41.4353930Z [2023-03-31 05:47:15,887] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4353936Z 2023-03-31T06:22:41.4354035Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4354108Z import torch 2023-03-31T06:22:41.4354166Z import math 2023-03-31T06:22:41.4354240Z import random 2023-03-31T06:22:41.4354310Z import os 2023-03-31T06:22:41.4354417Z import tempfile 2023-03-31T06:22:41.4354560Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4354566Z 2023-03-31T06:22:41.4354685Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4354810Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4354931Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4354936Z 2023-03-31T06:22:41.4355019Z aten = torch.ops.aten 2023-03-31T06:22:41.4355157Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4355252Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4355260Z 2023-03-31T06:22:41.4355351Z async_compile.wait(globals()) 2023-03-31T06:22:41.4355426Z del async_compile 2023-03-31T06:22:41.4355550Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4355610Z wrapper = ( 2023-03-31T06:22:41.4355693Z ''' 2023-03-31T06:22:41.4355773Z #include 2023-03-31T06:22:41.4355849Z #include 2023-03-31T06:22:41.4355857Z 2023-03-31T06:22:41.4355952Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4355957Z 2023-03-31T06:22:41.4356050Z template 2023-03-31T06:22:41.4356174Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4356260Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4356372Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4356470Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4356600Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4356681Z return kernel_cpp; 2023-03-31T06:22:41.4356743Z } 2023-03-31T06:22:41.4356934Z std::vector call_53(std::vector args) { 2023-03-31T06:22:41.4357029Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4357090Z arg0_1 = args[0]; 2023-03-31T06:22:41.4357164Z arg1_1 = args[1]; 2023-03-31T06:22:41.4357234Z arg2_1 = args[2]; 2023-03-31T06:22:41.4357297Z 2023-03-31T06:22:41.4357401Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4357497Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4357571Z .findSchemaOrThrow( 2023-03-31T06:22:41.4357672Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4357741Z "") 2023-03-31T06:22:41.4357812Z .typed< 2023-03-31T06:22:41.4357889Z at::Tensor( 2023-03-31T06:22:41.4357988Z const at::Tensor& input_t, 2023-03-31T06:22:41.4358123Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4358243Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4358322Z c10::string_view attr, 2023-03-31T06:22:41.4358445Z torch::List> scalars, 2023-03-31T06:22:41.4358568Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4358635Z 2023-03-31T06:22:41.4358879Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "none"); 2023-03-31T06:22:41.4358956Z arg1_1.reset(); 2023-03-31T06:22:41.4359030Z arg2_1.reset(); 2023-03-31T06:22:41.4359123Z return std::vector({buf0}); 2023-03-31T06:22:41.4359188Z } 2023-03-31T06:22:41.4359271Z ''' 2023-03-31T06:22:41.4359334Z ) 2023-03-31T06:22:41.4359340Z 2023-03-31T06:22:41.4359418Z module = load_inline( 2023-03-31T06:22:41.4359686Z name='inline_extension_cm5x7bscub67egu6kil23wnc2ntd2qsrllom36y3vydcvodxeoao', 2023-03-31T06:22:41.4359776Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4359879Z functions=['call_53'], 2023-03-31T06:22:41.4360260Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4360417Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4361153Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4361193Z 2023-03-31T06:22:41.4361274Z def _wrap_func(f): 2023-03-31T06:22:41.4361350Z def g(args): 2023-03-31T06:22:41.4361426Z return f(args) 2023-03-31T06:22:41.4361495Z return g 2023-03-31T06:22:41.4361594Z call = _wrap_func(module.call_53) 2023-03-31T06:22:41.4361600Z 2023-03-31T06:22:41.4361605Z 2023-03-31T06:22:41.4361687Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4361808Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4361936Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4362155Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4362367Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4362588Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4362719Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4362725Z 2023-03-31T06:22:41.4362730Z 2023-03-31T06:22:41.4362810Z if __name__ == "__main__": 2023-03-31T06:22:41.4362874Z import argparse 2023-03-31T06:22:41.4363150Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4363158Z 2023-03-31T06:22:41.4363271Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4363636Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4363988Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4364266Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4364364Z args = parser.parse_args() 2023-03-31T06:22:41.4364371Z 2023-03-31T06:22:41.4364466Z if args.benchmark_kernels: 2023-03-31T06:22:41.4364657Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4364711Z else: 2023-03-31T06:22:41.4364825Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4364921Z benchmark_compiled_module() 2023-03-31T06:22:41.4364927Z 2023-03-31T06:22:41.4365044Z if p: 2023-03-31T06:22:41.4365195Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4365295Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4365436Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4365442Z 2023-03-31T06:22:41.4365726Z [2023-03-31 05:47:15,887] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 45 2023-03-31T06:22:41.4365984Z [2023-03-31 05:47:15,978] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 46 2023-03-31T06:22:41.4366393Z [2023-03-31 05:47:38,043] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/bp/cbpg2tlkmat2g2eeka4k5qavogt475gelv2yqzwrltsuzlw377vb.py 2023-03-31T06:22:41.4366613Z [2023-03-31 05:47:38,043] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4366619Z 2023-03-31T06:22:41.4366716Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4366791Z import torch 2023-03-31T06:22:41.4366863Z import math 2023-03-31T06:22:41.4366935Z import random 2023-03-31T06:22:41.4367002Z import os 2023-03-31T06:22:41.4367067Z import tempfile 2023-03-31T06:22:41.4367186Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4367192Z 2023-03-31T06:22:41.4367311Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4367438Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4367573Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4367579Z 2023-03-31T06:22:41.4367697Z aten = torch.ops.aten 2023-03-31T06:22:41.4367869Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4367951Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4367970Z 2023-03-31T06:22:41.4368048Z async_compile.wait(globals()) 2023-03-31T06:22:41.4368124Z del async_compile 2023-03-31T06:22:41.4368246Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4368322Z wrapper = ( 2023-03-31T06:22:41.4368409Z ''' 2023-03-31T06:22:41.4368487Z #include 2023-03-31T06:22:41.4368550Z #include 2023-03-31T06:22:41.4368568Z 2023-03-31T06:22:41.4368651Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4368657Z 2023-03-31T06:22:41.4368750Z template 2023-03-31T06:22:41.4368876Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4368961Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4369086Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4369188Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4369316Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4369381Z return kernel_cpp; 2023-03-31T06:22:41.4369443Z } 2023-03-31T06:22:41.4369571Z std::vector call_54(std::vector args) { 2023-03-31T06:22:41.4369676Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4369750Z arg0_1 = args[0]; 2023-03-31T06:22:41.4369824Z arg1_1 = args[1]; 2023-03-31T06:22:41.4369895Z arg2_1 = args[2]; 2023-03-31T06:22:41.4369952Z arg3_1 = args[3]; 2023-03-31T06:22:41.4370015Z 2023-03-31T06:22:41.4370146Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4370240Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4370329Z .findSchemaOrThrow( 2023-03-31T06:22:41.4370427Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4370493Z "") 2023-03-31T06:22:41.4370550Z .typed< 2023-03-31T06:22:41.4370624Z at::Tensor( 2023-03-31T06:22:41.4370724Z const at::Tensor& input_t, 2023-03-31T06:22:41.4370826Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4370944Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4371032Z c10::string_view attr, 2023-03-31T06:22:41.4371156Z torch::List> scalars, 2023-03-31T06:22:41.4371265Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4371357Z 2023-03-31T06:22:41.4371585Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "none"); 2023-03-31T06:22:41.4371658Z arg1_1.reset(); 2023-03-31T06:22:41.4371729Z arg2_1.reset(); 2023-03-31T06:22:41.4371799Z arg3_1.reset(); 2023-03-31T06:22:41.4371904Z return std::vector({buf0}); 2023-03-31T06:22:41.4371954Z } 2023-03-31T06:22:41.4372037Z ''' 2023-03-31T06:22:41.4372099Z ) 2023-03-31T06:22:41.4372104Z 2023-03-31T06:22:41.4372186Z module = load_inline( 2023-03-31T06:22:41.4372459Z name='inline_extension_cwidvkkn5ycr44b4b5libljko5pjbaxwy2dstyiheyapho2qbgzi', 2023-03-31T06:22:41.4372548Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4372664Z functions=['call_54'], 2023-03-31T06:22:41.4373030Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4373189Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4373899Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4373906Z 2023-03-31T06:22:41.4373982Z def _wrap_func(f): 2023-03-31T06:22:41.4374055Z def g(args): 2023-03-31T06:22:41.4374159Z return f(args) 2023-03-31T06:22:41.4374265Z return g 2023-03-31T06:22:41.4374359Z call = _wrap_func(module.call_54) 2023-03-31T06:22:41.4374365Z 2023-03-31T06:22:41.4374370Z 2023-03-31T06:22:41.4374464Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4374571Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4374700Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4374920Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4375131Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4375341Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4375546Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4375682Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4375688Z 2023-03-31T06:22:41.4375693Z 2023-03-31T06:22:41.4375775Z if __name__ == "__main__": 2023-03-31T06:22:41.4375840Z import argparse 2023-03-31T06:22:41.4375973Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4375979Z 2023-03-31T06:22:41.4376091Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4376406Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4376749Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4377055Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4377150Z args = parser.parse_args() 2023-03-31T06:22:41.4377155Z 2023-03-31T06:22:41.4377245Z if args.benchmark_kernels: 2023-03-31T06:22:41.4377435Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4377492Z else: 2023-03-31T06:22:41.4377605Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4377700Z benchmark_compiled_module() 2023-03-31T06:22:41.4377706Z 2023-03-31T06:22:41.4377773Z if p: 2023-03-31T06:22:41.4377925Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4378021Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4378188Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4378194Z 2023-03-31T06:22:41.4378477Z [2023-03-31 05:47:38,043] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 46 2023-03-31T06:22:41.4378733Z [2023-03-31 05:47:38,112] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 47 2023-03-31T06:22:41.4379146Z [2023-03-31 05:48:05,415] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/qc/cqcfq72yroeazbeuun2tko6cdangsarsbqmfxtpxlt6pv7bz7sf2.py 2023-03-31T06:22:41.4379372Z [2023-03-31 05:48:05,415] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4379380Z 2023-03-31T06:22:41.4379478Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4379550Z import torch 2023-03-31T06:22:41.4379621Z import math 2023-03-31T06:22:41.4379695Z import random 2023-03-31T06:22:41.4379762Z import os 2023-03-31T06:22:41.4379824Z import tempfile 2023-03-31T06:22:41.4379944Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4379950Z 2023-03-31T06:22:41.4380068Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4380196Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4380330Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4380336Z 2023-03-31T06:22:41.4380418Z aten = torch.ops.aten 2023-03-31T06:22:41.4380556Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4380637Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4380684Z 2023-03-31T06:22:41.4380787Z async_compile.wait(globals()) 2023-03-31T06:22:41.4380863Z del async_compile 2023-03-31T06:22:41.4380986Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4381058Z wrapper = ( 2023-03-31T06:22:41.4381144Z ''' 2023-03-31T06:22:41.4381219Z #include 2023-03-31T06:22:41.4381281Z #include 2023-03-31T06:22:41.4381299Z 2023-03-31T06:22:41.4381382Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4381387Z 2023-03-31T06:22:41.4381479Z template 2023-03-31T06:22:41.4381604Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4381690Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4381814Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4381912Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4382035Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4382100Z return kernel_cpp; 2023-03-31T06:22:41.4382162Z } 2023-03-31T06:22:41.4382291Z std::vector call_55(std::vector args) { 2023-03-31T06:22:41.4382474Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4382555Z arg0_1 = args[0]; 2023-03-31T06:22:41.4382628Z arg1_1 = args[1]; 2023-03-31T06:22:41.4382702Z arg2_1 = args[2]; 2023-03-31T06:22:41.4382751Z 2023-03-31T06:22:41.4382851Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4382950Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4383039Z .findSchemaOrThrow( 2023-03-31T06:22:41.4383140Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4383246Z "") 2023-03-31T06:22:41.4383303Z .typed< 2023-03-31T06:22:41.4383381Z at::Tensor( 2023-03-31T06:22:41.4383480Z const at::Tensor& input_t, 2023-03-31T06:22:41.4383580Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4383697Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4383789Z c10::string_view attr, 2023-03-31T06:22:41.4383912Z torch::List> scalars, 2023-03-31T06:22:41.4384033Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4384086Z 2023-03-31T06:22:41.4384332Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "none"); 2023-03-31T06:22:41.4384405Z arg1_1.reset(); 2023-03-31T06:22:41.4384506Z arg2_1.reset(); 2023-03-31T06:22:41.4384612Z return std::vector({buf0}); 2023-03-31T06:22:41.4384678Z } 2023-03-31T06:22:41.4384759Z ''' 2023-03-31T06:22:41.4384810Z ) 2023-03-31T06:22:41.4384816Z 2023-03-31T06:22:41.4384894Z module = load_inline( 2023-03-31T06:22:41.4385158Z name='inline_extension_cm5x7bscub67egu6kil23wnc2ntd2qsrllom36y3vydcvodxeoao', 2023-03-31T06:22:41.4385245Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4385361Z functions=['call_55'], 2023-03-31T06:22:41.4385746Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4385902Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4386602Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4386612Z 2023-03-31T06:22:41.4386690Z def _wrap_func(f): 2023-03-31T06:22:41.4386747Z def g(args): 2023-03-31T06:22:41.4386821Z return f(args) 2023-03-31T06:22:41.4386888Z return g 2023-03-31T06:22:41.4386982Z call = _wrap_func(module.call_55) 2023-03-31T06:22:41.4386988Z 2023-03-31T06:22:41.4386993Z 2023-03-31T06:22:41.4387088Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4387232Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4387386Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4387602Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4387801Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4388010Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4388141Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4388146Z 2023-03-31T06:22:41.4388151Z 2023-03-31T06:22:41.4388232Z if __name__ == "__main__": 2023-03-31T06:22:41.4388311Z import argparse 2023-03-31T06:22:41.4388442Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4388447Z 2023-03-31T06:22:41.4388557Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4388869Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4389198Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4389476Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4389568Z args = parser.parse_args() 2023-03-31T06:22:41.4389575Z 2023-03-31T06:22:41.4389665Z if args.benchmark_kernels: 2023-03-31T06:22:41.4389855Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4389923Z else: 2023-03-31T06:22:41.4390065Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4390161Z benchmark_compiled_module() 2023-03-31T06:22:41.4390167Z 2023-03-31T06:22:41.4390223Z if p: 2023-03-31T06:22:41.4390372Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4390469Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4390609Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4390617Z 2023-03-31T06:22:41.4390901Z [2023-03-31 05:48:05,415] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 47 2023-03-31T06:22:41.4391168Z [2023-03-31 05:48:05,472] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 48 2023-03-31T06:22:41.4391604Z [2023-03-31 05:48:27,796] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/fi/cfizkgxhscatzfecf73i7tuu353gsuvkm3wfetccd2bi45j3ekib.py 2023-03-31T06:22:41.4391826Z [2023-03-31 05:48:27,796] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4391833Z 2023-03-31T06:22:41.4391930Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4391990Z import torch 2023-03-31T06:22:41.4392061Z import math 2023-03-31T06:22:41.4392135Z import random 2023-03-31T06:22:41.4392203Z import os 2023-03-31T06:22:41.4392277Z import tempfile 2023-03-31T06:22:41.4392396Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4392404Z 2023-03-31T06:22:41.4392523Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4392636Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4392769Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4392775Z 2023-03-31T06:22:41.4392855Z aten = torch.ops.aten 2023-03-31T06:22:41.4392996Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4393092Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4393097Z 2023-03-31T06:22:41.4393190Z async_compile.wait(globals()) 2023-03-31T06:22:41.4393264Z del async_compile 2023-03-31T06:22:41.4393386Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4393446Z wrapper = ( 2023-03-31T06:22:41.4393530Z ''' 2023-03-31T06:22:41.4393606Z #include 2023-03-31T06:22:41.4393682Z #include 2023-03-31T06:22:41.4393687Z 2023-03-31T06:22:41.4393809Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4393840Z 2023-03-31T06:22:41.4393936Z template 2023-03-31T06:22:41.4394063Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4394137Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4394264Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4394363Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4394492Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4394572Z return kernel_cpp; 2023-03-31T06:22:41.4394634Z } 2023-03-31T06:22:41.4394765Z std::vector call_56(std::vector args) { 2023-03-31T06:22:41.4394857Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4394931Z arg0_1 = args[0]; 2023-03-31T06:22:41.4395006Z arg1_1 = args[1]; 2023-03-31T06:22:41.4395078Z arg2_1 = args[2]; 2023-03-31T06:22:41.4395149Z arg3_1 = args[3]; 2023-03-31T06:22:41.4395212Z 2023-03-31T06:22:41.4395311Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4395394Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4395482Z .findSchemaOrThrow( 2023-03-31T06:22:41.4395582Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4395651Z "") 2023-03-31T06:22:41.4395722Z .typed< 2023-03-31T06:22:41.4395798Z at::Tensor( 2023-03-31T06:22:41.4395896Z const at::Tensor& input_t, 2023-03-31T06:22:41.4395986Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4396104Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4396223Z c10::string_view attr, 2023-03-31T06:22:41.4396347Z torch::List> scalars, 2023-03-31T06:22:41.4396469Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4396536Z 2023-03-31T06:22:41.4396762Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4396825Z arg1_1.reset(); 2023-03-31T06:22:41.4396900Z arg2_1.reset(); 2023-03-31T06:22:41.4396972Z arg3_1.reset(); 2023-03-31T06:22:41.4397079Z return std::vector({buf0}); 2023-03-31T06:22:41.4397144Z } 2023-03-31T06:22:41.4397227Z ''' 2023-03-31T06:22:41.4397290Z ) 2023-03-31T06:22:41.4397295Z 2023-03-31T06:22:41.4397361Z module = load_inline( 2023-03-31T06:22:41.4397680Z name='inline_extension_ctlllk2n4h2r3nrblvgymuc5rhrnmphwpwykxvktrtxfj2cfcnbv', 2023-03-31T06:22:41.4397768Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4397884Z functions=['call_56'], 2023-03-31T06:22:41.4398270Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4398428Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4399140Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4399151Z 2023-03-31T06:22:41.4399228Z def _wrap_func(f): 2023-03-31T06:22:41.4399301Z def g(args): 2023-03-31T06:22:41.4399364Z return f(args) 2023-03-31T06:22:41.4399434Z return g 2023-03-31T06:22:41.4399530Z call = _wrap_func(module.call_56) 2023-03-31T06:22:41.4399536Z 2023-03-31T06:22:41.4399541Z 2023-03-31T06:22:41.4399640Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4399760Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4399890Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4400106Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4400344Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4400573Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4400793Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4400929Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4400934Z 2023-03-31T06:22:41.4400939Z 2023-03-31T06:22:41.4401022Z if __name__ == "__main__": 2023-03-31T06:22:41.4401099Z import argparse 2023-03-31T06:22:41.4401232Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4401238Z 2023-03-31T06:22:41.4401351Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4401663Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4401990Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4402269Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4402366Z args = parser.parse_args() 2023-03-31T06:22:41.4402371Z 2023-03-31T06:22:41.4402462Z if args.benchmark_kernels: 2023-03-31T06:22:41.4402654Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4402721Z else: 2023-03-31T06:22:41.4402837Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4402933Z benchmark_compiled_module() 2023-03-31T06:22:41.4402939Z 2023-03-31T06:22:41.4403191Z if p: 2023-03-31T06:22:41.4403346Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4403442Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4403582Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4403588Z 2023-03-31T06:22:41.4403874Z [2023-03-31 05:48:27,796] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 48 2023-03-31T06:22:41.4404144Z [2023-03-31 05:48:27,839] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 49 2023-03-31T06:22:41.4404543Z [2023-03-31 05:48:52,348] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/z5/cz5wdgu7vh5yaxc76gtz6zvq47ajg63cvmg4kd5anpyuc7maf4ry.py 2023-03-31T06:22:41.4404808Z [2023-03-31 05:48:52,349] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4404816Z 2023-03-31T06:22:41.4404914Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4404976Z import torch 2023-03-31T06:22:41.4405048Z import math 2023-03-31T06:22:41.4405121Z import random 2023-03-31T06:22:41.4405188Z import os 2023-03-31T06:22:41.4405263Z import tempfile 2023-03-31T06:22:41.4405383Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4405390Z 2023-03-31T06:22:41.4405509Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4405623Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4405759Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4405764Z 2023-03-31T06:22:41.4405845Z aten = torch.ops.aten 2023-03-31T06:22:41.4405986Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4406080Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4406086Z 2023-03-31T06:22:41.4406179Z async_compile.wait(globals()) 2023-03-31T06:22:41.4406254Z del async_compile 2023-03-31T06:22:41.4406376Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4406437Z wrapper = ( 2023-03-31T06:22:41.4406521Z ''' 2023-03-31T06:22:41.4406597Z #include 2023-03-31T06:22:41.4406673Z #include 2023-03-31T06:22:41.4406678Z 2023-03-31T06:22:41.4406772Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4406777Z 2023-03-31T06:22:41.4406868Z template 2023-03-31T06:22:41.4407032Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4407180Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4407307Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4407407Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4407534Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4407614Z return kernel_cpp; 2023-03-31T06:22:41.4407679Z } 2023-03-31T06:22:41.4407812Z std::vector call_57(std::vector args) { 2023-03-31T06:22:41.4407891Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4407967Z arg0_1 = args[0]; 2023-03-31T06:22:41.4408043Z arg1_1 = args[1]; 2023-03-31T06:22:41.4408118Z arg2_1 = args[2]; 2023-03-31T06:22:41.4408182Z 2023-03-31T06:22:41.4408284Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4408378Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4408452Z .findSchemaOrThrow( 2023-03-31T06:22:41.4408553Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4408625Z "") 2023-03-31T06:22:41.4408696Z .typed< 2023-03-31T06:22:41.4408774Z at::Tensor( 2023-03-31T06:22:41.4408872Z const at::Tensor& input_t, 2023-03-31T06:22:41.4408973Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4409078Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4409166Z c10::string_view attr, 2023-03-31T06:22:41.4409289Z torch::List> scalars, 2023-03-31T06:22:41.4409415Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4409511Z 2023-03-31T06:22:41.4409759Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4409834Z arg1_1.reset(); 2023-03-31T06:22:41.4409891Z arg2_1.reset(); 2023-03-31T06:22:41.4409996Z return std::vector({buf0}); 2023-03-31T06:22:41.4410059Z } 2023-03-31T06:22:41.4410141Z ''' 2023-03-31T06:22:41.4410205Z ) 2023-03-31T06:22:41.4410212Z 2023-03-31T06:22:41.4410291Z module = load_inline( 2023-03-31T06:22:41.4410557Z name='inline_extension_crz2jjgr4amu57dazhb33qd7qxxdtpn2teqtf6b3zcjdkld77yyo', 2023-03-31T06:22:41.4410643Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4410746Z functions=['call_57'], 2023-03-31T06:22:41.4411126Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4411315Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4412022Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4412029Z 2023-03-31T06:22:41.4412107Z def _wrap_func(f): 2023-03-31T06:22:41.4412182Z def g(args): 2023-03-31T06:22:41.4412257Z return f(args) 2023-03-31T06:22:41.4412325Z return g 2023-03-31T06:22:41.4412406Z call = _wrap_func(module.call_57) 2023-03-31T06:22:41.4412412Z 2023-03-31T06:22:41.4412431Z 2023-03-31T06:22:41.4412512Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4412632Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4412762Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4412980Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4413195Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4413414Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4413541Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4413546Z 2023-03-31T06:22:41.4413551Z 2023-03-31T06:22:41.4413655Z if __name__ == "__main__": 2023-03-31T06:22:41.4413750Z import argparse 2023-03-31T06:22:41.4413881Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4413886Z 2023-03-31T06:22:41.4413996Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4414305Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4414644Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4414921Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4415014Z args = parser.parse_args() 2023-03-31T06:22:41.4415019Z 2023-03-31T06:22:41.4415109Z if args.benchmark_kernels: 2023-03-31T06:22:41.4415284Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4415354Z else: 2023-03-31T06:22:41.4415468Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4415564Z benchmark_compiled_module() 2023-03-31T06:22:41.4415569Z 2023-03-31T06:22:41.4415637Z if p: 2023-03-31T06:22:41.4415787Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4415882Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4416024Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4416030Z 2023-03-31T06:22:41.4416324Z [2023-03-31 05:48:52,349] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 49 2023-03-31T06:22:41.4416591Z [2023-03-31 05:48:52,470] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 50 2023-03-31T06:22:41.4416987Z [2023-03-31 05:49:15,939] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/44/c44ac627bt4c2gnkrizryovklm7qcmmnivbnz7vsev3hjamhn5ri.py 2023-03-31T06:22:41.4417210Z [2023-03-31 05:49:15,940] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4417217Z 2023-03-31T06:22:41.4417315Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4417387Z import torch 2023-03-31T06:22:41.4417458Z import math 2023-03-31T06:22:41.4417531Z import random 2023-03-31T06:22:41.4417586Z import os 2023-03-31T06:22:41.4417661Z import tempfile 2023-03-31T06:22:41.4417811Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4417816Z 2023-03-31T06:22:41.4417936Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4418065Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4418199Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4418205Z 2023-03-31T06:22:41.4418285Z aten = torch.ops.aten 2023-03-31T06:22:41.4418425Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4418507Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4418512Z 2023-03-31T06:22:41.4418607Z async_compile.wait(globals()) 2023-03-31T06:22:41.4418685Z del async_compile 2023-03-31T06:22:41.4418806Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4418878Z wrapper = ( 2023-03-31T06:22:41.4418960Z ''' 2023-03-31T06:22:41.4419036Z #include 2023-03-31T06:22:41.4419099Z #include 2023-03-31T06:22:41.4419104Z 2023-03-31T06:22:41.4419196Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4419204Z 2023-03-31T06:22:41.4419296Z template 2023-03-31T06:22:41.4419419Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4419509Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4419634Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4419732Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4419844Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4419922Z return kernel_cpp; 2023-03-31T06:22:41.4419986Z } 2023-03-31T06:22:41.4420147Z std::vector call_58(std::vector args) { 2023-03-31T06:22:41.4420287Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4420361Z arg0_1 = args[0]; 2023-03-31T06:22:41.4420432Z arg1_1 = args[1]; 2023-03-31T06:22:41.4420491Z arg2_1 = args[2]; 2023-03-31T06:22:41.4420565Z arg3_1 = args[3]; 2023-03-31T06:22:41.4420628Z 2023-03-31T06:22:41.4420727Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4420824Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4420911Z .findSchemaOrThrow( 2023-03-31T06:22:41.4421011Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4421066Z "") 2023-03-31T06:22:41.4421134Z .typed< 2023-03-31T06:22:41.4421209Z at::Tensor( 2023-03-31T06:22:41.4421310Z const at::Tensor& input_t, 2023-03-31T06:22:41.4421410Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4421528Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4421620Z c10::string_view attr, 2023-03-31T06:22:41.4421730Z torch::List> scalars, 2023-03-31T06:22:41.4421853Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4421920Z 2023-03-31T06:22:41.4422146Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4422221Z arg1_1.reset(); 2023-03-31T06:22:41.4422294Z arg2_1.reset(); 2023-03-31T06:22:41.4422364Z arg3_1.reset(); 2023-03-31T06:22:41.4422587Z return std::vector({buf0}); 2023-03-31T06:22:41.4422654Z } 2023-03-31T06:22:41.4422740Z ''' 2023-03-31T06:22:41.4422805Z ) 2023-03-31T06:22:41.4422813Z 2023-03-31T06:22:41.4422893Z module = load_inline( 2023-03-31T06:22:41.4423166Z name='inline_extension_ctlllk2n4h2r3nrblvgymuc5rhrnmphwpwykxvktrtxfj2cfcnbv', 2023-03-31T06:22:41.4423255Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4423358Z functions=['call_58'], 2023-03-31T06:22:41.4423740Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4423899Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4424617Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4424658Z 2023-03-31T06:22:41.4424737Z def _wrap_func(f): 2023-03-31T06:22:41.4424812Z def g(args): 2023-03-31T06:22:41.4424887Z return f(args) 2023-03-31T06:22:41.4424955Z return g 2023-03-31T06:22:41.4425050Z call = _wrap_func(module.call_58) 2023-03-31T06:22:41.4425056Z 2023-03-31T06:22:41.4425061Z 2023-03-31T06:22:41.4425147Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4425269Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4425397Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4425613Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4425820Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4426031Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4426241Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4426375Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4426381Z 2023-03-31T06:22:41.4426386Z 2023-03-31T06:22:41.4426465Z if __name__ == "__main__": 2023-03-31T06:22:41.4426528Z import argparse 2023-03-31T06:22:41.4426658Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4426664Z 2023-03-31T06:22:41.4426804Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4427141Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4427477Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4427756Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4427851Z args = parser.parse_args() 2023-03-31T06:22:41.4427859Z 2023-03-31T06:22:41.4427948Z if args.benchmark_kernels: 2023-03-31T06:22:41.4428124Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4428191Z else: 2023-03-31T06:22:41.4428304Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4428399Z benchmark_compiled_module() 2023-03-31T06:22:41.4428404Z 2023-03-31T06:22:41.4428475Z if p: 2023-03-31T06:22:41.4428627Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4428725Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4428864Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4428870Z 2023-03-31T06:22:41.4429134Z [2023-03-31 05:49:15,940] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 50 2023-03-31T06:22:41.4429403Z [2023-03-31 05:49:15,980] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 51 2023-03-31T06:22:41.4429837Z [2023-03-31 05:49:37,457] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/xs/cxsnqcjknmdpd27binqqbt5xjv2itgyhgneafud63yfscqicvd44.py 2023-03-31T06:22:41.4430060Z [2023-03-31 05:49:37,457] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4430065Z 2023-03-31T06:22:41.4430162Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4430237Z import torch 2023-03-31T06:22:41.4430310Z import math 2023-03-31T06:22:41.4430382Z import random 2023-03-31T06:22:41.4430438Z import os 2023-03-31T06:22:41.4430511Z import tempfile 2023-03-31T06:22:41.4430630Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4430635Z 2023-03-31T06:22:41.4430755Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4430881Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4431046Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4431052Z 2023-03-31T06:22:41.4431135Z aten = torch.ops.aten 2023-03-31T06:22:41.4431276Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4431359Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4431364Z 2023-03-31T06:22:41.4431456Z async_compile.wait(globals()) 2023-03-31T06:22:41.4431531Z del async_compile 2023-03-31T06:22:41.4431654Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4431729Z wrapper = ( 2023-03-31T06:22:41.4431815Z ''' 2023-03-31T06:22:41.4431894Z #include 2023-03-31T06:22:41.4431957Z #include 2023-03-31T06:22:41.4431961Z 2023-03-31T06:22:41.4432053Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4432059Z 2023-03-31T06:22:41.4432150Z template 2023-03-31T06:22:41.4432272Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4432361Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4432485Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4432584Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4432697Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4432776Z return kernel_cpp; 2023-03-31T06:22:41.4432839Z } 2023-03-31T06:22:41.4432966Z std::vector call_59(std::vector args) { 2023-03-31T06:22:41.4433058Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4433133Z arg0_1 = args[0]; 2023-03-31T06:22:41.4433230Z arg1_1 = args[1]; 2023-03-31T06:22:41.4433317Z arg2_1 = args[2]; 2023-03-31T06:22:41.4433380Z 2023-03-31T06:22:41.4433479Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4433571Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4433660Z .findSchemaOrThrow( 2023-03-31T06:22:41.4433759Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4433829Z "") 2023-03-31T06:22:41.4433887Z .typed< 2023-03-31T06:22:41.4433967Z at::Tensor( 2023-03-31T06:22:41.4434065Z const at::Tensor& input_t, 2023-03-31T06:22:41.4434170Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4434289Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4434380Z c10::string_view attr, 2023-03-31T06:22:41.4434504Z torch::List> scalars, 2023-03-31T06:22:41.4434612Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4434681Z 2023-03-31T06:22:41.4434924Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4434999Z arg1_1.reset(); 2023-03-31T06:22:41.4435073Z arg2_1.reset(); 2023-03-31T06:22:41.4435180Z return std::vector({buf0}); 2023-03-31T06:22:41.4435244Z } 2023-03-31T06:22:41.4435326Z ''' 2023-03-31T06:22:41.4435375Z ) 2023-03-31T06:22:41.4435383Z 2023-03-31T06:22:41.4435462Z module = load_inline( 2023-03-31T06:22:41.4435731Z name='inline_extension_crz2jjgr4amu57dazhb33qd7qxxdtpn2teqtf6b3zcjdkld77yyo', 2023-03-31T06:22:41.4435852Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4435971Z functions=['call_59'], 2023-03-31T06:22:41.4436355Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4436512Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4437226Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4437235Z 2023-03-31T06:22:41.4437298Z def _wrap_func(f): 2023-03-31T06:22:41.4437431Z def g(args): 2023-03-31T06:22:41.4437508Z return f(args) 2023-03-31T06:22:41.4437576Z return g 2023-03-31T06:22:41.4437676Z call = _wrap_func(module.call_59) 2023-03-31T06:22:41.4437682Z 2023-03-31T06:22:41.4437687Z 2023-03-31T06:22:41.4437784Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4437904Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4438034Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4438239Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4438453Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4438662Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4438793Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4438798Z 2023-03-31T06:22:41.4438803Z 2023-03-31T06:22:41.4438883Z if __name__ == "__main__": 2023-03-31T06:22:41.4438964Z import argparse 2023-03-31T06:22:41.4439096Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4439101Z 2023-03-31T06:22:41.4439215Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4439511Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4439849Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4440156Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4440275Z args = parser.parse_args() 2023-03-31T06:22:41.4440281Z 2023-03-31T06:22:41.4440370Z if args.benchmark_kernels: 2023-03-31T06:22:41.4440564Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4440633Z else: 2023-03-31T06:22:41.4440750Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4440833Z benchmark_compiled_module() 2023-03-31T06:22:41.4440856Z 2023-03-31T06:22:41.4440912Z if p: 2023-03-31T06:22:41.4441065Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4441162Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4441300Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4441306Z 2023-03-31T06:22:41.4441590Z [2023-03-31 05:49:37,457] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 51 2023-03-31T06:22:41.4441858Z [2023-03-31 05:49:37,523] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 52 2023-03-31T06:22:41.4442265Z [2023-03-31 05:49:59,181] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/g5/cg5u3e7lva3dgxng5u2e6lucaqnhyuqillq6w3eg46yqazuppxmz.py 2023-03-31T06:22:41.4442484Z [2023-03-31 05:49:59,181] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4442493Z 2023-03-31T06:22:41.4442577Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4442677Z import torch 2023-03-31T06:22:41.4442750Z import math 2023-03-31T06:22:41.4442823Z import random 2023-03-31T06:22:41.4442890Z import os 2023-03-31T06:22:41.4442965Z import tempfile 2023-03-31T06:22:41.4443252Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4443261Z 2023-03-31T06:22:41.4443369Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4443499Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4443635Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4443640Z 2023-03-31T06:22:41.4443723Z aten = torch.ops.aten 2023-03-31T06:22:41.4443864Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4443961Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4443966Z 2023-03-31T06:22:41.4444128Z async_compile.wait(globals()) 2023-03-31T06:22:41.4444205Z del async_compile 2023-03-31T06:22:41.4444315Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4444389Z wrapper = ( 2023-03-31T06:22:41.4444482Z ''' 2023-03-31T06:22:41.4444560Z #include 2023-03-31T06:22:41.4444638Z #include 2023-03-31T06:22:41.4444643Z 2023-03-31T06:22:41.4444737Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4444742Z 2023-03-31T06:22:41.4444836Z template 2023-03-31T06:22:41.4444948Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4445038Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4445164Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4445266Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4445392Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4445473Z return kernel_cpp; 2023-03-31T06:22:41.4445535Z } 2023-03-31T06:22:41.4445651Z std::vector call_61(std::vector args) { 2023-03-31T06:22:41.4445756Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4445830Z arg0_1 = args[0]; 2023-03-31T06:22:41.4445904Z arg1_1 = args[1]; 2023-03-31T06:22:41.4445977Z arg2_1 = args[2]; 2023-03-31T06:22:41.4446047Z arg3_1 = args[3]; 2023-03-31T06:22:41.4446110Z 2023-03-31T06:22:41.4446196Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4446288Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4446374Z .findSchemaOrThrow( 2023-03-31T06:22:41.4446510Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4446613Z "") 2023-03-31T06:22:41.4446681Z .typed< 2023-03-31T06:22:41.4446757Z at::Tensor( 2023-03-31T06:22:41.4446841Z const at::Tensor& input_t, 2023-03-31T06:22:41.4446941Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4447060Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4447148Z c10::string_view attr, 2023-03-31T06:22:41.4447271Z torch::List> scalars, 2023-03-31T06:22:41.4447394Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4447459Z 2023-03-31T06:22:41.4447610Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.4447671Z arg1_1.reset(); 2023-03-31T06:22:41.4447742Z arg2_1.reset(); 2023-03-31T06:22:41.4447811Z arg3_1.reset(); 2023-03-31T06:22:41.4447917Z return std::vector({buf0}); 2023-03-31T06:22:41.4447984Z } 2023-03-31T06:22:41.4448072Z ''' 2023-03-31T06:22:41.4448120Z ) 2023-03-31T06:22:41.4448140Z 2023-03-31T06:22:41.4448206Z module = load_inline( 2023-03-31T06:22:41.4448475Z name='inline_extension_cchmpo3apapn7nym7i52nb22x4wy2ymilgftavlve4qpxwqdrpqg', 2023-03-31T06:22:41.4448563Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4448683Z functions=['call_61'], 2023-03-31T06:22:41.4449067Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4449257Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4449972Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4449980Z 2023-03-31T06:22:41.4450057Z def _wrap_func(f): 2023-03-31T06:22:41.4450115Z def g(args): 2023-03-31T06:22:41.4450191Z return f(args) 2023-03-31T06:22:41.4450258Z return g 2023-03-31T06:22:41.4450351Z call = _wrap_func(module.call_61) 2023-03-31T06:22:41.4450357Z 2023-03-31T06:22:41.4450361Z 2023-03-31T06:22:41.4450456Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4450603Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4450731Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4450946Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4451141Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4451350Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4451570Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4451706Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4451711Z 2023-03-31T06:22:41.4451716Z 2023-03-31T06:22:41.4451798Z if __name__ == "__main__": 2023-03-31T06:22:41.4451874Z import argparse 2023-03-31T06:22:41.4452006Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4452015Z 2023-03-31T06:22:41.4452128Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4452424Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4452765Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4453041Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4453134Z args = parser.parse_args() 2023-03-31T06:22:41.4453140Z 2023-03-31T06:22:41.4453264Z if args.benchmark_kernels: 2023-03-31T06:22:41.4453483Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4453551Z else: 2023-03-31T06:22:41.4453663Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4453746Z benchmark_compiled_module() 2023-03-31T06:22:41.4453764Z 2023-03-31T06:22:41.4453819Z if p: 2023-03-31T06:22:41.4453971Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4454066Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4454205Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4454212Z 2023-03-31T06:22:41.4454491Z [2023-03-31 05:49:59,181] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 52 2023-03-31T06:22:41.4454760Z [2023-03-31 05:49:59,232] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 53 2023-03-31T06:22:41.4455173Z [2023-03-31 05:50:23,140] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ad/caddy3vlbh5y2azgmvdjpfsf6bflgnk4kzh4xxacgufk2k3zpmuv.py 2023-03-31T06:22:41.4455395Z [2023-03-31 05:50:23,141] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4455401Z 2023-03-31T06:22:41.4455500Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4455559Z import torch 2023-03-31T06:22:41.4455636Z import math 2023-03-31T06:22:41.4455709Z import random 2023-03-31T06:22:41.4455776Z import os 2023-03-31T06:22:41.4455849Z import tempfile 2023-03-31T06:22:41.4455999Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4456006Z 2023-03-31T06:22:41.4456114Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4456239Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4456370Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4456376Z 2023-03-31T06:22:41.4456460Z aten = torch.ops.aten 2023-03-31T06:22:41.4456600Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4456695Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4456700Z 2023-03-31T06:22:41.4456791Z async_compile.wait(globals()) 2023-03-31T06:22:41.4456867Z del async_compile 2023-03-31T06:22:41.4456978Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4457050Z wrapper = ( 2023-03-31T06:22:41.4457160Z ''' 2023-03-31T06:22:41.4457237Z #include 2023-03-31T06:22:41.4457313Z #include 2023-03-31T06:22:41.4457318Z 2023-03-31T06:22:41.4457412Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4457418Z 2023-03-31T06:22:41.4457512Z template 2023-03-31T06:22:41.4457623Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4457708Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4457833Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4457930Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4458057Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4458138Z return kernel_cpp; 2023-03-31T06:22:41.4458202Z } 2023-03-31T06:22:41.4458318Z std::vector call_63(std::vector args) { 2023-03-31T06:22:41.4458408Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4458482Z arg0_1 = args[0]; 2023-03-31T06:22:41.4458555Z arg1_1 = args[1]; 2023-03-31T06:22:41.4458629Z arg2_1 = args[2]; 2023-03-31T06:22:41.4458691Z 2023-03-31T06:22:41.4458789Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4458871Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4458958Z .findSchemaOrThrow( 2023-03-31T06:22:41.4459055Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4459123Z "") 2023-03-31T06:22:41.4459191Z .typed< 2023-03-31T06:22:41.4459265Z at::Tensor( 2023-03-31T06:22:41.4459361Z const at::Tensor& input_t, 2023-03-31T06:22:41.4459511Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4459656Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4459745Z c10::string_view attr, 2023-03-31T06:22:41.4459867Z torch::List> scalars, 2023-03-31T06:22:41.4459990Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4460059Z 2023-03-31T06:22:41.4460227Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.4460302Z arg1_1.reset(); 2023-03-31T06:22:41.4460363Z arg2_1.reset(); 2023-03-31T06:22:41.4460472Z return std::vector({buf0}); 2023-03-31T06:22:41.4460539Z } 2023-03-31T06:22:41.4460626Z ''' 2023-03-31T06:22:41.4460691Z ) 2023-03-31T06:22:41.4460697Z 2023-03-31T06:22:41.4460778Z module = load_inline( 2023-03-31T06:22:41.4461048Z name='inline_extension_cxmtqkjphkfev7ejgp2tmn6rjh6v7wt5i32fvaq375ug5fuqtged', 2023-03-31T06:22:41.4461123Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4461244Z functions=['call_63'], 2023-03-31T06:22:41.4461628Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4461784Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4462608Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4462619Z 2023-03-31T06:22:41.4462700Z def _wrap_func(f): 2023-03-31T06:22:41.4462775Z def g(args): 2023-03-31T06:22:41.4462852Z return f(args) 2023-03-31T06:22:41.4462907Z return g 2023-03-31T06:22:41.4463007Z call = _wrap_func(module.call_63) 2023-03-31T06:22:41.4463014Z 2023-03-31T06:22:41.4463019Z 2023-03-31T06:22:41.4463115Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4463236Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4463364Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4463585Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4463830Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4464049Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4464178Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4464184Z 2023-03-31T06:22:41.4464188Z 2023-03-31T06:22:41.4464254Z if __name__ == "__main__": 2023-03-31T06:22:41.4464331Z import argparse 2023-03-31T06:22:41.4464462Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4464468Z 2023-03-31T06:22:41.4464580Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4464890Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4465228Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4465504Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4465603Z args = parser.parse_args() 2023-03-31T06:22:41.4465608Z 2023-03-31T06:22:41.4465700Z if args.benchmark_kernels: 2023-03-31T06:22:41.4465877Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4465947Z else: 2023-03-31T06:22:41.4466062Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4466158Z benchmark_compiled_module() 2023-03-31T06:22:41.4466164Z 2023-03-31T06:22:41.4466232Z if p: 2023-03-31T06:22:41.4466412Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4466559Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4466684Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4466689Z 2023-03-31T06:22:41.4466972Z [2023-03-31 05:50:23,141] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 53 2023-03-31T06:22:41.4467241Z [2023-03-31 05:50:23,205] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 54 2023-03-31T06:22:41.4467650Z [2023-03-31 05:50:45,468] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/qx/cqxgvw6ktujsnnmlvf6a4tzwuoujqwxvuuqvwezr22urlorfcncg.py 2023-03-31T06:22:41.4467870Z [2023-03-31 05:50:45,468] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4467876Z 2023-03-31T06:22:41.4467973Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4468047Z import torch 2023-03-31T06:22:41.4468121Z import math 2023-03-31T06:22:41.4468180Z import random 2023-03-31T06:22:41.4468248Z import os 2023-03-31T06:22:41.4468322Z import tempfile 2023-03-31T06:22:41.4468441Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4468446Z 2023-03-31T06:22:41.4468564Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4468687Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4468825Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4468831Z 2023-03-31T06:22:41.4468913Z aten = torch.ops.aten 2023-03-31T06:22:41.4469071Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4469171Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4469177Z 2023-03-31T06:22:41.4469269Z async_compile.wait(globals()) 2023-03-31T06:22:41.4469344Z del async_compile 2023-03-31T06:22:41.4469465Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4469537Z wrapper = ( 2023-03-31T06:22:41.4469622Z ''' 2023-03-31T06:22:41.4469687Z #include 2023-03-31T06:22:41.4469762Z #include 2023-03-31T06:22:41.4469768Z 2023-03-31T06:22:41.4469861Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4469866Z 2023-03-31T06:22:41.4469958Z template 2023-03-31T06:22:41.4470082Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4470202Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4470327Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4470412Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4470540Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4470619Z return kernel_cpp; 2023-03-31T06:22:41.4470682Z } 2023-03-31T06:22:41.4470812Z std::vector call_65(std::vector args) { 2023-03-31T06:22:41.4470916Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4470991Z arg0_1 = args[0]; 2023-03-31T06:22:41.4471052Z arg1_1 = args[1]; 2023-03-31T06:22:41.4471126Z arg2_1 = args[2]; 2023-03-31T06:22:41.4471197Z arg3_1 = args[3]; 2023-03-31T06:22:41.4471260Z 2023-03-31T06:22:41.4471359Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4471454Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4471540Z .findSchemaOrThrow( 2023-03-31T06:22:41.4471625Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4471694Z "") 2023-03-31T06:22:41.4471762Z .typed< 2023-03-31T06:22:41.4471838Z at::Tensor( 2023-03-31T06:22:41.4471939Z const at::Tensor& input_t, 2023-03-31T06:22:41.4472040Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4472159Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4472234Z c10::string_view attr, 2023-03-31T06:22:41.4472356Z torch::List> scalars, 2023-03-31T06:22:41.4472502Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4472595Z 2023-03-31T06:22:41.4472747Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.4472822Z arg1_1.reset(); 2023-03-31T06:22:41.4472895Z arg2_1.reset(); 2023-03-31T06:22:41.4472967Z arg3_1.reset(); 2023-03-31T06:22:41.4473059Z return std::vector({buf0}); 2023-03-31T06:22:41.4473126Z } 2023-03-31T06:22:41.4473209Z ''' 2023-03-31T06:22:41.4473271Z ) 2023-03-31T06:22:41.4473277Z 2023-03-31T06:22:41.4473356Z module = load_inline( 2023-03-31T06:22:41.4473633Z name='inline_extension_cchmpo3apapn7nym7i52nb22x4wy2ymilgftavlve4qpxwqdrpqg', 2023-03-31T06:22:41.4473722Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4473824Z functions=['call_65'], 2023-03-31T06:22:41.4474201Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4474360Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4475069Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4475078Z 2023-03-31T06:22:41.4475156Z def _wrap_func(f): 2023-03-31T06:22:41.4475232Z def g(args): 2023-03-31T06:22:41.4475309Z return f(args) 2023-03-31T06:22:41.4475406Z return g 2023-03-31T06:22:41.4475490Z call = _wrap_func(module.call_65) 2023-03-31T06:22:41.4475509Z 2023-03-31T06:22:41.4475513Z 2023-03-31T06:22:41.4475595Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4475713Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4475842Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4476061Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4476274Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4476484Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4476692Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4476859Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4476865Z 2023-03-31T06:22:41.4476869Z 2023-03-31T06:22:41.4476937Z if __name__ == "__main__": 2023-03-31T06:22:41.4477015Z import argparse 2023-03-31T06:22:41.4477146Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4477152Z 2023-03-31T06:22:41.4477263Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4477578Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4477922Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4478199Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4478295Z args = parser.parse_args() 2023-03-31T06:22:41.4478300Z 2023-03-31T06:22:41.4478392Z if args.benchmark_kernels: 2023-03-31T06:22:41.4478571Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4478638Z else: 2023-03-31T06:22:41.4478754Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4478851Z benchmark_compiled_module() 2023-03-31T06:22:41.4478856Z 2023-03-31T06:22:41.4478927Z if p: 2023-03-31T06:22:41.4479075Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4479172Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4479327Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4479358Z 2023-03-31T06:22:41.4479643Z [2023-03-31 05:50:45,468] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 54 2023-03-31T06:22:41.4479912Z [2023-03-31 05:50:45,528] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 55 2023-03-31T06:22:41.4480329Z [2023-03-31 05:51:07,499] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/wf/cwfxzrgszyenylp2jf2ypdfiutpaeztvuasjzq4q57pbpohu75bw.py 2023-03-31T06:22:41.4480556Z [2023-03-31 05:51:07,499] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4480562Z 2023-03-31T06:22:41.4480661Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4480735Z import torch 2023-03-31T06:22:41.4480807Z import math 2023-03-31T06:22:41.4480867Z import random 2023-03-31T06:22:41.4480935Z import os 2023-03-31T06:22:41.4481011Z import tempfile 2023-03-31T06:22:41.4481132Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4481140Z 2023-03-31T06:22:41.4481258Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4481382Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4481516Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4481522Z 2023-03-31T06:22:41.4481604Z aten = torch.ops.aten 2023-03-31T06:22:41.4481732Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4481827Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4481833Z 2023-03-31T06:22:41.4481951Z async_compile.wait(globals()) 2023-03-31T06:22:41.4482028Z del async_compile 2023-03-31T06:22:41.4482149Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4482223Z wrapper = ( 2023-03-31T06:22:41.4482305Z ''' 2023-03-31T06:22:41.4482368Z #include 2023-03-31T06:22:41.4482443Z #include 2023-03-31T06:22:41.4482448Z 2023-03-31T06:22:41.4482542Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4482549Z 2023-03-31T06:22:41.4482644Z template 2023-03-31T06:22:41.4482768Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4482857Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4482981Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4483200Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4483384Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4483466Z return kernel_cpp; 2023-03-31T06:22:41.4483531Z } 2023-03-31T06:22:41.4483664Z std::vector call_67(std::vector args) { 2023-03-31T06:22:41.4483759Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4483835Z arg0_1 = args[0]; 2023-03-31T06:22:41.4483895Z arg1_1 = args[1]; 2023-03-31T06:22:41.4483969Z arg2_1 = args[2]; 2023-03-31T06:22:41.4484032Z 2023-03-31T06:22:41.4484131Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4484228Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4484320Z .findSchemaOrThrow( 2023-03-31T06:22:41.4484422Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4484475Z "") 2023-03-31T06:22:41.4484544Z .typed< 2023-03-31T06:22:41.4484622Z at::Tensor( 2023-03-31T06:22:41.4484722Z const at::Tensor& input_t, 2023-03-31T06:22:41.4484825Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4484949Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4485040Z c10::string_view attr, 2023-03-31T06:22:41.4485165Z torch::List> scalars, 2023-03-31T06:22:41.4485274Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4485339Z 2023-03-31T06:22:41.4485502Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.4485576Z arg1_1.reset(); 2023-03-31T06:22:41.4485682Z arg2_1.reset(); 2023-03-31T06:22:41.4485820Z return std::vector({buf0}); 2023-03-31T06:22:41.4485884Z } 2023-03-31T06:22:41.4485955Z ''' 2023-03-31T06:22:41.4486017Z ) 2023-03-31T06:22:41.4486024Z 2023-03-31T06:22:41.4486104Z module = load_inline( 2023-03-31T06:22:41.4486373Z name='inline_extension_cxmtqkjphkfev7ejgp2tmn6rjh6v7wt5i32fvaq375ug5fuqtged', 2023-03-31T06:22:41.4486460Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4486578Z functions=['call_67'], 2023-03-31T06:22:41.4486966Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4487108Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4487817Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4487839Z 2023-03-31T06:22:41.4487903Z def _wrap_func(f): 2023-03-31T06:22:41.4487974Z def g(args): 2023-03-31T06:22:41.4488048Z return f(args) 2023-03-31T06:22:41.4488117Z return g 2023-03-31T06:22:41.4488212Z call = _wrap_func(module.call_67) 2023-03-31T06:22:41.4488221Z 2023-03-31T06:22:41.4488225Z 2023-03-31T06:22:41.4488322Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4488443Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4488598Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4488816Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4489027Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4489236Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4489365Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4489371Z 2023-03-31T06:22:41.4489376Z 2023-03-31T06:22:41.4489454Z if __name__ == "__main__": 2023-03-31T06:22:41.4489531Z import argparse 2023-03-31T06:22:41.4489662Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4489668Z 2023-03-31T06:22:41.4489777Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4490112Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4490450Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4490725Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4490820Z args = parser.parse_args() 2023-03-31T06:22:41.4490825Z 2023-03-31T06:22:41.4490918Z if args.benchmark_kernels: 2023-03-31T06:22:41.4491108Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4491176Z else: 2023-03-31T06:22:41.4491289Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4491372Z benchmark_compiled_module() 2023-03-31T06:22:41.4491377Z 2023-03-31T06:22:41.4491446Z if p: 2023-03-31T06:22:41.4491600Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4491697Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4491837Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4491843Z 2023-03-31T06:22:41.4492123Z [2023-03-31 05:51:07,500] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 55 2023-03-31T06:22:41.4492390Z [2023-03-31 05:51:07,589] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 56 2023-03-31T06:22:41.4492830Z [2023-03-31 05:51:31,463] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/fs/cfs5wjtwmpl23dufyvt4octrdgnkdzkc4ywpjv25xwxr5yvmpjt3.py 2023-03-31T06:22:41.4493062Z [2023-03-31 05:51:31,463] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4493080Z 2023-03-31T06:22:41.4493165Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4493237Z import torch 2023-03-31T06:22:41.4493312Z import math 2023-03-31T06:22:41.4493389Z import random 2023-03-31T06:22:41.4493459Z import os 2023-03-31T06:22:41.4493533Z import tempfile 2023-03-31T06:22:41.4493641Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4493660Z 2023-03-31T06:22:41.4493766Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4493891Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4494025Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4494030Z 2023-03-31T06:22:41.4494111Z aten = torch.ops.aten 2023-03-31T06:22:41.4494253Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4494350Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4494355Z 2023-03-31T06:22:41.4494447Z async_compile.wait(globals()) 2023-03-31T06:22:41.4494510Z del async_compile 2023-03-31T06:22:41.4494632Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4494707Z wrapper = ( 2023-03-31T06:22:41.4494793Z ''' 2023-03-31T06:22:41.4494869Z #include 2023-03-31T06:22:41.4494949Z #include 2023-03-31T06:22:41.4494954Z 2023-03-31T06:22:41.4495046Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4495078Z 2023-03-31T06:22:41.4495173Z template 2023-03-31T06:22:41.4495284Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4495371Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4495496Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4495595Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4495723Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4495803Z return kernel_cpp; 2023-03-31T06:22:41.4495867Z } 2023-03-31T06:22:41.4495985Z std::vector call_68(std::vector args) { 2023-03-31T06:22:41.4496090Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4496165Z arg0_1 = args[0]; 2023-03-31T06:22:41.4496237Z arg1_1 = args[1]; 2023-03-31T06:22:41.4496341Z arg2_1 = args[2]; 2023-03-31T06:22:41.4496414Z arg3_1 = args[3]; 2023-03-31T06:22:41.4496464Z 2023-03-31T06:22:41.4496568Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4496661Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4496751Z .findSchemaOrThrow( 2023-03-31T06:22:41.4496848Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4496916Z "") 2023-03-31T06:22:41.4496983Z .typed< 2023-03-31T06:22:41.4497045Z at::Tensor( 2023-03-31T06:22:41.4497145Z const at::Tensor& input_t, 2023-03-31T06:22:41.4497248Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4497364Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4497452Z c10::string_view attr, 2023-03-31T06:22:41.4497574Z torch::List> scalars, 2023-03-31T06:22:41.4497697Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4497765Z 2023-03-31T06:22:41.4497976Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "swish", {-1}, ""); 2023-03-31T06:22:41.4498051Z arg1_1.reset(); 2023-03-31T06:22:41.4498122Z arg2_1.reset(); 2023-03-31T06:22:41.4498191Z arg3_1.reset(); 2023-03-31T06:22:41.4498298Z return std::vector({buf0}); 2023-03-31T06:22:41.4498360Z } 2023-03-31T06:22:41.4498441Z ''' 2023-03-31T06:22:41.4498489Z ) 2023-03-31T06:22:41.4498495Z 2023-03-31T06:22:41.4498573Z module = load_inline( 2023-03-31T06:22:41.4498871Z name='inline_extension_co2j6sdzmzk3o6xkazchr6wfthup3h4hy5pmjbkj5ysafin6kr5b', 2023-03-31T06:22:41.4498984Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4499100Z functions=['call_68'], 2023-03-31T06:22:41.4499480Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4499634Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4500345Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4500352Z 2023-03-31T06:22:41.4500428Z def _wrap_func(f): 2023-03-31T06:22:41.4500485Z def g(args): 2023-03-31T06:22:41.4500559Z return f(args) 2023-03-31T06:22:41.4500630Z return g 2023-03-31T06:22:41.4500728Z call = _wrap_func(module.call_68) 2023-03-31T06:22:41.4500734Z 2023-03-31T06:22:41.4500739Z 2023-03-31T06:22:41.4500835Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4500955Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4501082Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4501283Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4501493Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4501727Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4501943Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4502077Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4502083Z 2023-03-31T06:22:41.4502088Z 2023-03-31T06:22:41.4502169Z if __name__ == "__main__": 2023-03-31T06:22:41.4502249Z import argparse 2023-03-31T06:22:41.4502459Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4502465Z 2023-03-31T06:22:41.4502581Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4502878Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4503254Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4503537Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4503633Z args = parser.parse_args() 2023-03-31T06:22:41.4503638Z 2023-03-31T06:22:41.4503730Z if args.benchmark_kernels: 2023-03-31T06:22:41.4503923Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4503992Z else: 2023-03-31T06:22:41.4504107Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4504192Z benchmark_compiled_module() 2023-03-31T06:22:41.4504198Z 2023-03-31T06:22:41.4504267Z if p: 2023-03-31T06:22:41.4504417Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4504515Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4504652Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4504661Z 2023-03-31T06:22:41.4504941Z [2023-03-31 05:51:31,464] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 56 2023-03-31T06:22:41.4505207Z [2023-03-31 05:51:31,521] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 57 2023-03-31T06:22:41.4505614Z [2023-03-31 05:51:54,092] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/qm/cqmwij5ye57znaybavttfcicv7yjvi6fy3qf4fztpgrt37i6zxly.py 2023-03-31T06:22:41.4505866Z [2023-03-31 05:51:54,092] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4505899Z 2023-03-31T06:22:41.4505985Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4506060Z import torch 2023-03-31T06:22:41.4506131Z import math 2023-03-31T06:22:41.4506204Z import random 2023-03-31T06:22:41.4506272Z import os 2023-03-31T06:22:41.4506346Z import tempfile 2023-03-31T06:22:41.4506451Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4506472Z 2023-03-31T06:22:41.4506580Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4506709Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4506842Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4506848Z 2023-03-31T06:22:41.4506929Z aten = torch.ops.aten 2023-03-31T06:22:41.4507067Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4507161Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4507167Z 2023-03-31T06:22:41.4507260Z async_compile.wait(globals()) 2023-03-31T06:22:41.4507324Z del async_compile 2023-03-31T06:22:41.4507447Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4507519Z wrapper = ( 2023-03-31T06:22:41.4507600Z ''' 2023-03-31T06:22:41.4507676Z #include 2023-03-31T06:22:41.4507752Z #include 2023-03-31T06:22:41.4507757Z 2023-03-31T06:22:41.4507849Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4507857Z 2023-03-31T06:22:41.4507949Z template 2023-03-31T06:22:41.4508059Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4508180Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4508307Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4508405Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4508531Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4508612Z return kernel_cpp; 2023-03-31T06:22:41.4508677Z } 2023-03-31T06:22:41.4508796Z std::vector call_69(std::vector args) { 2023-03-31T06:22:41.4508892Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4508965Z arg0_1 = args[0]; 2023-03-31T06:22:41.4509038Z arg1_1 = args[1]; 2023-03-31T06:22:41.4509108Z arg2_1 = args[2]; 2023-03-31T06:22:41.4509171Z 2023-03-31T06:22:41.4509255Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4509348Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4509502Z .findSchemaOrThrow( 2023-03-31T06:22:41.4509601Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4509668Z "") 2023-03-31T06:22:41.4509738Z .typed< 2023-03-31T06:22:41.4509818Z at::Tensor( 2023-03-31T06:22:41.4509904Z const at::Tensor& input_t, 2023-03-31T06:22:41.4510005Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4510126Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4510215Z c10::string_view attr, 2023-03-31T06:22:41.4510340Z torch::List> scalars, 2023-03-31T06:22:41.4510466Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4510534Z 2023-03-31T06:22:41.4510777Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "swish", {-1}, ""); 2023-03-31T06:22:41.4510837Z arg1_1.reset(); 2023-03-31T06:22:41.4510910Z arg2_1.reset(); 2023-03-31T06:22:41.4511022Z return std::vector({buf0}); 2023-03-31T06:22:41.4511087Z } 2023-03-31T06:22:41.4511169Z ''' 2023-03-31T06:22:41.4511232Z ) 2023-03-31T06:22:41.4511238Z 2023-03-31T06:22:41.4511319Z module = load_inline( 2023-03-31T06:22:41.4511576Z name='inline_extension_cqvj3c7evbaptsjxt2y7hhph6qcdvq6y4ukqluourdvjmnmle7j7', 2023-03-31T06:22:41.4511664Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4511784Z functions=['call_69'], 2023-03-31T06:22:41.4512191Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4512385Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4513093Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4513103Z 2023-03-31T06:22:41.4513184Z def _wrap_func(f): 2023-03-31T06:22:41.4513259Z def g(args): 2023-03-31T06:22:41.4513336Z return f(args) 2023-03-31T06:22:41.4513392Z return g 2023-03-31T06:22:41.4513488Z call = _wrap_func(module.call_69) 2023-03-31T06:22:41.4513494Z 2023-03-31T06:22:41.4513498Z 2023-03-31T06:22:41.4513594Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4513714Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4513845Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4514061Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4514273Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4514493Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4514610Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4514615Z 2023-03-31T06:22:41.4514634Z 2023-03-31T06:22:41.4514700Z if __name__ == "__main__": 2023-03-31T06:22:41.4514806Z import argparse 2023-03-31T06:22:41.4514941Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4514947Z 2023-03-31T06:22:41.4515059Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4515372Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4515712Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4515993Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4516087Z args = parser.parse_args() 2023-03-31T06:22:41.4516093Z 2023-03-31T06:22:41.4516170Z if args.benchmark_kernels: 2023-03-31T06:22:41.4516389Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4516457Z else: 2023-03-31T06:22:41.4516574Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4516672Z benchmark_compiled_module() 2023-03-31T06:22:41.4516677Z 2023-03-31T06:22:41.4516747Z if p: 2023-03-31T06:22:41.4516898Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4516981Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4517122Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4517129Z 2023-03-31T06:22:41.4517408Z [2023-03-31 05:51:54,093] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 57 2023-03-31T06:22:41.4523620Z [2023-03-31 05:51:54,159] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 58 2023-03-31T06:22:41.4524112Z [2023-03-31 05:52:15,135] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/uu/cuuvwnu2fzo7y2wruhke4yfhhe735efvei6r6eslubujcpfkk3zo.py 2023-03-31T06:22:41.4524353Z [2023-03-31 05:52:15,135] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4524362Z 2023-03-31T06:22:41.4524465Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4524539Z import torch 2023-03-31T06:22:41.4524611Z import math 2023-03-31T06:22:41.4524672Z import random 2023-03-31T06:22:41.4524742Z import os 2023-03-31T06:22:41.4524819Z import tempfile 2023-03-31T06:22:41.4525081Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4525136Z 2023-03-31T06:22:41.4525260Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4525389Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4525526Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4525532Z 2023-03-31T06:22:41.4525600Z aten = torch.ops.aten 2023-03-31T06:22:41.4525745Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4525844Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4525849Z 2023-03-31T06:22:41.4525945Z async_compile.wait(globals()) 2023-03-31T06:22:41.4526024Z del async_compile 2023-03-31T06:22:41.4526147Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4526222Z wrapper = ( 2023-03-31T06:22:41.4526296Z ''' 2023-03-31T06:22:41.4526375Z #include 2023-03-31T06:22:41.4526455Z #include 2023-03-31T06:22:41.4526460Z 2023-03-31T06:22:41.4526565Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4526573Z 2023-03-31T06:22:41.4526669Z template 2023-03-31T06:22:41.4526798Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4526886Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4527012Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4527097Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4527225Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4527309Z return kernel_cpp; 2023-03-31T06:22:41.4527375Z } 2023-03-31T06:22:41.4527547Z std::vector call_70(std::vector args) { 2023-03-31T06:22:41.4527656Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4527729Z arg0_1 = args[0]; 2023-03-31T06:22:41.4527788Z arg1_1 = args[1]; 2023-03-31T06:22:41.4527861Z arg2_1 = args[2]; 2023-03-31T06:22:41.4527933Z arg3_1 = args[3]; 2023-03-31T06:22:41.4527998Z 2023-03-31T06:22:41.4528102Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4528199Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4528288Z .findSchemaOrThrow( 2023-03-31T06:22:41.4528375Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4528445Z "") 2023-03-31T06:22:41.4528514Z .typed< 2023-03-31T06:22:41.4528595Z at::Tensor( 2023-03-31T06:22:41.4528694Z const at::Tensor& input_t, 2023-03-31T06:22:41.4528833Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4528954Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4529033Z c10::string_view attr, 2023-03-31T06:22:41.4529159Z torch::List> scalars, 2023-03-31T06:22:41.4529284Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4529350Z 2023-03-31T06:22:41.4529583Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "swish", {-1}, ""); 2023-03-31T06:22:41.4529657Z arg1_1.reset(); 2023-03-31T06:22:41.4529732Z arg2_1.reset(); 2023-03-31T06:22:41.4529791Z arg3_1.reset(); 2023-03-31T06:22:41.4529897Z return std::vector({buf0}); 2023-03-31T06:22:41.4529963Z } 2023-03-31T06:22:41.4530043Z ''' 2023-03-31T06:22:41.4530105Z ) 2023-03-31T06:22:41.4530111Z 2023-03-31T06:22:41.4530189Z module = load_inline( 2023-03-31T06:22:41.4530459Z name='inline_extension_co2j6sdzmzk3o6xkazchr6wfthup3h4hy5pmjbkj5ysafin6kr5b', 2023-03-31T06:22:41.4530535Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4530651Z functions=['call_70'], 2023-03-31T06:22:41.4531044Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4531202Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4531955Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4531990Z 2023-03-31T06:22:41.4532068Z def _wrap_func(f): 2023-03-31T06:22:41.4532141Z def g(args): 2023-03-31T06:22:41.4532215Z return f(args) 2023-03-31T06:22:41.4532283Z return g 2023-03-31T06:22:41.4532368Z call = _wrap_func(module.call_70) 2023-03-31T06:22:41.4532374Z 2023-03-31T06:22:41.4532379Z 2023-03-31T06:22:41.4532474Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4532595Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4532724Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4532940Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4533150Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4533363Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4533570Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4533692Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4533697Z 2023-03-31T06:22:41.4533714Z 2023-03-31T06:22:41.4533780Z if __name__ == "__main__": 2023-03-31T06:22:41.4533858Z import argparse 2023-03-31T06:22:41.4533988Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4533993Z 2023-03-31T06:22:41.4534132Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4534444Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4534783Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4535065Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4535159Z args = parser.parse_args() 2023-03-31T06:22:41.4535165Z 2023-03-31T06:22:41.4535241Z if args.benchmark_kernels: 2023-03-31T06:22:41.4535434Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4535505Z else: 2023-03-31T06:22:41.4535618Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4535750Z benchmark_compiled_module() 2023-03-31T06:22:41.4535755Z 2023-03-31T06:22:41.4535824Z if p: 2023-03-31T06:22:41.4535978Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4536063Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4536200Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4536206Z 2023-03-31T06:22:41.4536488Z [2023-03-31 05:52:15,136] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 58 2023-03-31T06:22:41.4536758Z [2023-03-31 05:52:15,186] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 59 2023-03-31T06:22:41.4537172Z [2023-03-31 05:52:36,152] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ek/cekscpy2s5bjnvzcmqxdmsxfdcikismwtzlvudrpakwmspuga2zx.py 2023-03-31T06:22:41.4537392Z [2023-03-31 05:52:36,152] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4537401Z 2023-03-31T06:22:41.4537498Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4537570Z import torch 2023-03-31T06:22:41.4537642Z import math 2023-03-31T06:22:41.4537701Z import random 2023-03-31T06:22:41.4537769Z import os 2023-03-31T06:22:41.4537848Z import tempfile 2023-03-31T06:22:41.4537966Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4537972Z 2023-03-31T06:22:41.4538091Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4538246Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4538404Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4538410Z 2023-03-31T06:22:41.4538478Z aten = torch.ops.aten 2023-03-31T06:22:41.4538615Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4538710Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4538715Z 2023-03-31T06:22:41.4538807Z async_compile.wait(globals()) 2023-03-31T06:22:41.4538884Z del async_compile 2023-03-31T06:22:41.4539005Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4539078Z wrapper = ( 2023-03-31T06:22:41.4539147Z ''' 2023-03-31T06:22:41.4539226Z #include 2023-03-31T06:22:41.4539303Z #include 2023-03-31T06:22:41.4539309Z 2023-03-31T06:22:41.4539401Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4539406Z 2023-03-31T06:22:41.4539499Z template 2023-03-31T06:22:41.4539624Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4539712Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4539840Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4539925Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4540050Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4540130Z return kernel_cpp; 2023-03-31T06:22:41.4540193Z } 2023-03-31T06:22:41.4540324Z std::vector call_71(std::vector args) { 2023-03-31T06:22:41.4540416Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4540489Z arg0_1 = args[0]; 2023-03-31T06:22:41.4540549Z arg1_1 = args[1]; 2023-03-31T06:22:41.4540649Z arg2_1 = args[2]; 2023-03-31T06:22:41.4540712Z 2023-03-31T06:22:41.4540811Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4540905Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4540992Z .findSchemaOrThrow( 2023-03-31T06:22:41.4541090Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4541146Z "") 2023-03-31T06:22:41.4541216Z .typed< 2023-03-31T06:22:41.4541294Z at::Tensor( 2023-03-31T06:22:41.4541394Z const at::Tensor& input_t, 2023-03-31T06:22:41.4541498Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4541616Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4541704Z c10::string_view attr, 2023-03-31T06:22:41.4541812Z torch::List> scalars, 2023-03-31T06:22:41.4541974Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4542040Z 2023-03-31T06:22:41.4542280Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "swish", {-1}, ""); 2023-03-31T06:22:41.4542353Z arg1_1.reset(); 2023-03-31T06:22:41.4542533Z arg2_1.reset(); 2023-03-31T06:22:41.4542644Z return std::vector({buf0}); 2023-03-31T06:22:41.4542694Z } 2023-03-31T06:22:41.4542778Z ''' 2023-03-31T06:22:41.4542842Z ) 2023-03-31T06:22:41.4542848Z 2023-03-31T06:22:41.4542930Z module = load_inline( 2023-03-31T06:22:41.4543207Z name='inline_extension_cqvj3c7evbaptsjxt2y7hhph6qcdvq6y4ukqluourdvjmnmle7j7', 2023-03-31T06:22:41.4543299Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4543416Z functions=['call_71'], 2023-03-31T06:22:41.4543801Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4543948Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4544663Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4544670Z 2023-03-31T06:22:41.4544748Z def _wrap_func(f): 2023-03-31T06:22:41.4544858Z def g(args): 2023-03-31T06:22:41.4544963Z return f(args) 2023-03-31T06:22:41.4545032Z return g 2023-03-31T06:22:41.4545127Z call = _wrap_func(module.call_71) 2023-03-31T06:22:41.4545133Z 2023-03-31T06:22:41.4545138Z 2023-03-31T06:22:41.4545232Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4545337Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4545464Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4545683Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4545896Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4546105Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4546233Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4546238Z 2023-03-31T06:22:41.4546245Z 2023-03-31T06:22:41.4546324Z if __name__ == "__main__": 2023-03-31T06:22:41.4546402Z import argparse 2023-03-31T06:22:41.4546533Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4546539Z 2023-03-31T06:22:41.4546636Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4546946Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4547285Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4547594Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4547691Z args = parser.parse_args() 2023-03-31T06:22:41.4547696Z 2023-03-31T06:22:41.4547787Z if args.benchmark_kernels: 2023-03-31T06:22:41.4547977Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4548043Z else: 2023-03-31T06:22:41.4548146Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4548243Z benchmark_compiled_module() 2023-03-31T06:22:41.4548248Z 2023-03-31T06:22:41.4548316Z if p: 2023-03-31T06:22:41.4548466Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4548561Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4548699Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4548747Z 2023-03-31T06:22:41.4549028Z [2023-03-31 05:52:36,153] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 59 2023-03-31T06:22:41.4549296Z [2023-03-31 05:52:36,213] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 60 2023-03-31T06:22:41.4549689Z [2023-03-31 05:52:56,983] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3b/c3brywphmsk3lmknhn5co5ocd7hfa37sneltdutyxb62qdb6c7oy.py 2023-03-31T06:22:41.4549912Z [2023-03-31 05:52:56,984] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4549919Z 2023-03-31T06:22:41.4550015Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4550087Z import torch 2023-03-31T06:22:41.4550157Z import math 2023-03-31T06:22:41.4550229Z import random 2023-03-31T06:22:41.4550296Z import os 2023-03-31T06:22:41.4550369Z import tempfile 2023-03-31T06:22:41.4550473Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4550482Z 2023-03-31T06:22:41.4550601Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4550728Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4550868Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4550874Z 2023-03-31T06:22:41.4550955Z aten = torch.ops.aten 2023-03-31T06:22:41.4551097Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4551193Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4551198Z 2023-03-31T06:22:41.4551292Z async_compile.wait(globals()) 2023-03-31T06:22:41.4551382Z del async_compile 2023-03-31T06:22:41.4551536Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4551609Z wrapper = ( 2023-03-31T06:22:41.4551692Z ''' 2023-03-31T06:22:41.4551766Z #include 2023-03-31T06:22:41.4551842Z #include 2023-03-31T06:22:41.4551848Z 2023-03-31T06:22:41.4551938Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4551943Z 2023-03-31T06:22:41.4552026Z template 2023-03-31T06:22:41.4552148Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4552234Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4552362Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4552459Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4552584Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4552664Z return kernel_cpp; 2023-03-31T06:22:41.4552714Z } 2023-03-31T06:22:41.4552843Z std::vector call_72(std::vector args) { 2023-03-31T06:22:41.4552950Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4553023Z arg0_1 = args[0]; 2023-03-31T06:22:41.4553097Z arg1_1 = args[1]; 2023-03-31T06:22:41.4553170Z arg2_1 = args[2]; 2023-03-31T06:22:41.4553241Z arg3_1 = args[3]; 2023-03-31T06:22:41.4553291Z 2023-03-31T06:22:41.4553391Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4553487Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4553577Z .findSchemaOrThrow( 2023-03-31T06:22:41.4553675Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4553742Z "") 2023-03-31T06:22:41.4553841Z .typed< 2023-03-31T06:22:41.4553904Z at::Tensor( 2023-03-31T06:22:41.4554002Z const at::Tensor& input_t, 2023-03-31T06:22:41.4554103Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4554220Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4554308Z c10::string_view attr, 2023-03-31T06:22:41.4554433Z torch::List> scalars, 2023-03-31T06:22:41.4554557Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4554609Z 2023-03-31T06:22:41.4554846Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.4554921Z arg1_1.reset(); 2023-03-31T06:22:41.4555021Z arg2_1.reset(); 2023-03-31T06:22:41.4555092Z arg3_1.reset(); 2023-03-31T06:22:41.4555197Z return std::vector({buf0}); 2023-03-31T06:22:41.4555258Z } 2023-03-31T06:22:41.4555329Z ''' 2023-03-31T06:22:41.4555389Z ) 2023-03-31T06:22:41.4555397Z 2023-03-31T06:22:41.4555475Z module = load_inline( 2023-03-31T06:22:41.4555741Z name='inline_extension_cet5mw6ckpn5sdinqb23taoug7nnqbjxmhb6i27zj645y6mt3saj', 2023-03-31T06:22:41.4555826Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4555942Z functions=['call_72'], 2023-03-31T06:22:41.4556328Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4556483Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4557181Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4557205Z 2023-03-31T06:22:41.4557269Z def _wrap_func(f): 2023-03-31T06:22:41.4557341Z def g(args): 2023-03-31T06:22:41.4557416Z return f(args) 2023-03-31T06:22:41.4557483Z return g 2023-03-31T06:22:41.4557578Z call = _wrap_func(module.call_72) 2023-03-31T06:22:41.4557584Z 2023-03-31T06:22:41.4557589Z 2023-03-31T06:22:41.4557683Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4557830Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4557970Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4558186Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4558392Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4558602Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4558819Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4558956Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4558962Z 2023-03-31T06:22:41.4558967Z 2023-03-31T06:22:41.4559047Z if __name__ == "__main__": 2023-03-31T06:22:41.4559123Z import argparse 2023-03-31T06:22:41.4559255Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4559261Z 2023-03-31T06:22:41.4559361Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4559675Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4560012Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4560290Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4560385Z args = parser.parse_args() 2023-03-31T06:22:41.4560391Z 2023-03-31T06:22:41.4560516Z if args.benchmark_kernels: 2023-03-31T06:22:41.4560707Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4560774Z else: 2023-03-31T06:22:41.4560873Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4560969Z benchmark_compiled_module() 2023-03-31T06:22:41.4560974Z 2023-03-31T06:22:41.4561043Z if p: 2023-03-31T06:22:41.4561194Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4561294Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4561430Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4561436Z 2023-03-31T06:22:41.4561716Z [2023-03-31 05:52:56,984] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 60 2023-03-31T06:22:41.4562012Z [2023-03-31 05:52:57,042] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 61 2023-03-31T06:22:41.4562419Z [2023-03-31 05:53:18,008] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/hq/chqx2hbjk3ml4sq73pta2izzsylwctpyq6kqax2yj7rx2fvmkjho.py 2023-03-31T06:22:41.4562625Z [2023-03-31 05:53:18,009] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4562631Z 2023-03-31T06:22:41.4562727Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4562800Z import torch 2023-03-31T06:22:41.4562875Z import math 2023-03-31T06:22:41.4562950Z import random 2023-03-31T06:22:41.4563179Z import os 2023-03-31T06:22:41.4563256Z import tempfile 2023-03-31T06:22:41.4563362Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4563368Z 2023-03-31T06:22:41.4563489Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4563620Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4563758Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4563764Z 2023-03-31T06:22:41.4563847Z aten = torch.ops.aten 2023-03-31T06:22:41.4563991Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4564089Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4564094Z 2023-03-31T06:22:41.4564187Z async_compile.wait(globals()) 2023-03-31T06:22:41.4564249Z del async_compile 2023-03-31T06:22:41.4564373Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4564448Z wrapper = ( 2023-03-31T06:22:41.4564532Z ''' 2023-03-31T06:22:41.4564665Z #include 2023-03-31T06:22:41.4564784Z #include 2023-03-31T06:22:41.4564790Z 2023-03-31T06:22:41.4564889Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4564894Z 2023-03-31T06:22:41.4564973Z template 2023-03-31T06:22:41.4565098Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4565186Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4565315Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4565414Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4565544Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4565625Z return kernel_cpp; 2023-03-31T06:22:41.4565675Z } 2023-03-31T06:22:41.4565806Z std::vector call_73(std::vector args) { 2023-03-31T06:22:41.4565901Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4565977Z arg0_1 = args[0]; 2023-03-31T06:22:41.4566053Z arg1_1 = args[1]; 2023-03-31T06:22:41.4566127Z arg2_1 = args[2]; 2023-03-31T06:22:41.4566194Z 2023-03-31T06:22:41.4566279Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4566374Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4566463Z .findSchemaOrThrow( 2023-03-31T06:22:41.4566566Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4566634Z "") 2023-03-31T06:22:41.4566705Z .typed< 2023-03-31T06:22:41.4566784Z at::Tensor( 2023-03-31T06:22:41.4566867Z const at::Tensor& input_t, 2023-03-31T06:22:41.4566970Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4567159Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4567251Z c10::string_view attr, 2023-03-31T06:22:41.4567373Z torch::List> scalars, 2023-03-31T06:22:41.4567496Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4567566Z 2023-03-31T06:22:41.4567825Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.4567889Z arg1_1.reset(); 2023-03-31T06:22:41.4567962Z arg2_1.reset(); 2023-03-31T06:22:41.4568069Z return std::vector({buf0}); 2023-03-31T06:22:41.4568134Z } 2023-03-31T06:22:41.4568216Z ''' 2023-03-31T06:22:41.4568279Z ) 2023-03-31T06:22:41.4568285Z 2023-03-31T06:22:41.4568388Z module = load_inline( 2023-03-31T06:22:41.4568660Z name='inline_extension_ceyanris47ohp6ch643vxttxgfgln7ral5h3xnryxnbufbtgql2n', 2023-03-31T06:22:41.4568746Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4568867Z functions=['call_73'], 2023-03-31T06:22:41.4569249Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4569406Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4570113Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4570121Z 2023-03-31T06:22:41.4570196Z def _wrap_func(f): 2023-03-31T06:22:41.4570266Z def g(args): 2023-03-31T06:22:41.4570331Z return f(args) 2023-03-31T06:22:41.4570401Z return g 2023-03-31T06:22:41.4570496Z call = _wrap_func(module.call_73) 2023-03-31T06:22:41.4570502Z 2023-03-31T06:22:41.4570509Z 2023-03-31T06:22:41.4570603Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4570721Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4570847Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4571065Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4571307Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4571537Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4571670Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4571676Z 2023-03-31T06:22:41.4571680Z 2023-03-31T06:22:41.4571757Z if __name__ == "__main__": 2023-03-31T06:22:41.4571834Z import argparse 2023-03-31T06:22:41.4571967Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4571973Z 2023-03-31T06:22:41.4572083Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4572396Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4572731Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4573010Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4573092Z args = parser.parse_args() 2023-03-31T06:22:41.4573097Z 2023-03-31T06:22:41.4573186Z if args.benchmark_kernels: 2023-03-31T06:22:41.4573375Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4573441Z else: 2023-03-31T06:22:41.4573554Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4573655Z benchmark_compiled_module() 2023-03-31T06:22:41.4573661Z 2023-03-31T06:22:41.4573729Z if p: 2023-03-31T06:22:41.4573891Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4573989Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4574126Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4574132Z 2023-03-31T06:22:41.4574413Z [2023-03-31 05:53:18,009] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 61 2023-03-31T06:22:41.4574683Z [2023-03-31 05:53:18,079] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 62 2023-03-31T06:22:41.4575089Z [2023-03-31 05:53:41,028] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/uw/cuwrq6hayvxwve3njcjx33pg43ushcavnehz6ngwxpfhllh3p4uo.py 2023-03-31T06:22:41.4575309Z [2023-03-31 05:53:41,028] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4575341Z 2023-03-31T06:22:41.4575441Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4575513Z import torch 2023-03-31T06:22:41.4575571Z import math 2023-03-31T06:22:41.4575645Z import random 2023-03-31T06:22:41.4575714Z import os 2023-03-31T06:22:41.4575789Z import tempfile 2023-03-31T06:22:41.4575906Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4575912Z 2023-03-31T06:22:41.4576031Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4576155Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4576279Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4576299Z 2023-03-31T06:22:41.4576365Z aten = torch.ops.aten 2023-03-31T06:22:41.4576505Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4576600Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4576606Z 2023-03-31T06:22:41.4576696Z async_compile.wait(globals()) 2023-03-31T06:22:41.4576773Z del async_compile 2023-03-31T06:22:41.4576895Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4576971Z wrapper = ( 2023-03-31T06:22:41.4577041Z ''' 2023-03-31T06:22:41.4577118Z #include 2023-03-31T06:22:41.4577193Z #include 2023-03-31T06:22:41.4577198Z 2023-03-31T06:22:41.4577290Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4577295Z 2023-03-31T06:22:41.4577386Z template 2023-03-31T06:22:41.4577508Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4577595Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4577735Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4577877Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4578001Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4578079Z return kernel_cpp; 2023-03-31T06:22:41.4578144Z } 2023-03-31T06:22:41.4578272Z std::vector call_74(std::vector args) { 2023-03-31T06:22:41.4578377Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4578437Z arg0_1 = args[0]; 2023-03-31T06:22:41.4578509Z arg1_1 = args[1]; 2023-03-31T06:22:41.4578582Z arg2_1 = args[2]; 2023-03-31T06:22:41.4578652Z arg3_1 = args[3]; 2023-03-31T06:22:41.4578714Z 2023-03-31T06:22:41.4578813Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4578907Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4578982Z .findSchemaOrThrow( 2023-03-31T06:22:41.4579080Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4579148Z "") 2023-03-31T06:22:41.4579217Z .typed< 2023-03-31T06:22:41.4579294Z at::Tensor( 2023-03-31T06:22:41.4579395Z const at::Tensor& input_t, 2023-03-31T06:22:41.4579497Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4579600Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4579688Z c10::string_view attr, 2023-03-31T06:22:41.4579816Z torch::List> scalars, 2023-03-31T06:22:41.4579936Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4580032Z 2023-03-31T06:22:41.4580269Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.4580341Z arg1_1.reset(); 2023-03-31T06:22:41.4580400Z arg2_1.reset(); 2023-03-31T06:22:41.4580468Z arg3_1.reset(); 2023-03-31T06:22:41.4580573Z return std::vector({buf0}); 2023-03-31T06:22:41.4580636Z } 2023-03-31T06:22:41.4580719Z ''' 2023-03-31T06:22:41.4580782Z ) 2023-03-31T06:22:41.4580787Z 2023-03-31T06:22:41.4580868Z module = load_inline( 2023-03-31T06:22:41.4581120Z name='inline_extension_cet5mw6ckpn5sdinqb23taoug7nnqbjxmhb6i27zj645y6mt3saj', 2023-03-31T06:22:41.4581205Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4581319Z functions=['call_74'], 2023-03-31T06:22:41.4581698Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4581887Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4582700Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4582710Z 2023-03-31T06:22:41.4582789Z def _wrap_func(f): 2023-03-31T06:22:41.4582862Z def g(args): 2023-03-31T06:22:41.4582938Z return f(args) 2023-03-31T06:22:41.4582993Z return g 2023-03-31T06:22:41.4583090Z call = _wrap_func(module.call_74) 2023-03-31T06:22:41.4583096Z 2023-03-31T06:22:41.4583101Z 2023-03-31T06:22:41.4583197Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4583319Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4583447Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4583668Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4583878Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4584090Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4584284Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4584453Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4584485Z 2023-03-31T06:22:41.4584490Z 2023-03-31T06:22:41.4584571Z if __name__ == "__main__": 2023-03-31T06:22:41.4584649Z import argparse 2023-03-31T06:22:41.4584780Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4584786Z 2023-03-31T06:22:41.4584898Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4585209Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4585544Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4585823Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4585902Z args = parser.parse_args() 2023-03-31T06:22:41.4585908Z 2023-03-31T06:22:41.4585998Z if args.benchmark_kernels: 2023-03-31T06:22:41.4586191Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4586258Z else: 2023-03-31T06:22:41.4586371Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4586465Z benchmark_compiled_module() 2023-03-31T06:22:41.4586471Z 2023-03-31T06:22:41.4586538Z if p: 2023-03-31T06:22:41.4586686Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4586773Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4586943Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4586949Z 2023-03-31T06:22:41.4587232Z [2023-03-31 05:53:41,029] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 62 2023-03-31T06:22:41.4587499Z [2023-03-31 05:53:41,075] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 63 2023-03-31T06:22:41.4587912Z [2023-03-31 05:54:02,159] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/te/ctemjcmrdlq6pu7qqywuoqrnmtafwfeec7phmspyn4ilijyl742k.py 2023-03-31T06:22:41.4588135Z [2023-03-31 05:54:02,159] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4588141Z 2023-03-31T06:22:41.4588239Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4588311Z import torch 2023-03-31T06:22:41.4588369Z import math 2023-03-31T06:22:41.4588484Z import random 2023-03-31T06:22:41.4588555Z import os 2023-03-31T06:22:41.4588630Z import tempfile 2023-03-31T06:22:41.4588748Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4588757Z 2023-03-31T06:22:41.4588877Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4589004Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4589124Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4589143Z 2023-03-31T06:22:41.4589211Z aten = torch.ops.aten 2023-03-31T06:22:41.4589352Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4589450Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4589456Z 2023-03-31T06:22:41.4589546Z async_compile.wait(globals()) 2023-03-31T06:22:41.4589623Z del async_compile 2023-03-31T06:22:41.4589746Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4589818Z wrapper = ( 2023-03-31T06:22:41.4589889Z ''' 2023-03-31T06:22:41.4589966Z #include 2023-03-31T06:22:41.4590045Z #include 2023-03-31T06:22:41.4590050Z 2023-03-31T06:22:41.4590143Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4590149Z 2023-03-31T06:22:41.4590244Z template 2023-03-31T06:22:41.4590367Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4590453Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4590565Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4590663Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4590819Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4590927Z return kernel_cpp; 2023-03-31T06:22:41.4590992Z } 2023-03-31T06:22:41.4591120Z std::vector call_75(std::vector args) { 2023-03-31T06:22:41.4591213Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4591274Z arg0_1 = args[0]; 2023-03-31T06:22:41.4591345Z arg1_1 = args[1]; 2023-03-31T06:22:41.4591415Z arg2_1 = args[2]; 2023-03-31T06:22:41.4591484Z 2023-03-31T06:22:41.4591583Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4591677Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4591766Z .findSchemaOrThrow( 2023-03-31T06:22:41.4591852Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4591918Z "") 2023-03-31T06:22:41.4591988Z .typed< 2023-03-31T06:22:41.4592064Z at::Tensor( 2023-03-31T06:22:41.4592160Z const at::Tensor& input_t, 2023-03-31T06:22:41.4592261Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4592381Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4592458Z c10::string_view attr, 2023-03-31T06:22:41.4592579Z torch::List> scalars, 2023-03-31T06:22:41.4592701Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4592768Z 2023-03-31T06:22:41.4593020Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.4593095Z arg1_1.reset(); 2023-03-31T06:22:41.4593165Z arg2_1.reset(); 2023-03-31T06:22:41.4593297Z return std::vector({buf0}); 2023-03-31T06:22:41.4593348Z } 2023-03-31T06:22:41.4593428Z ''' 2023-03-31T06:22:41.4593490Z ) 2023-03-31T06:22:41.4593495Z 2023-03-31T06:22:41.4593572Z module = load_inline( 2023-03-31T06:22:41.4593842Z name='inline_extension_ceyanris47ohp6ch643vxttxgfgln7ral5h3xnryxnbufbtgql2n', 2023-03-31T06:22:41.4593927Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4594043Z functions=['call_75'], 2023-03-31T06:22:41.4594415Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4594569Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4595273Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4595308Z 2023-03-31T06:22:41.4595385Z def _wrap_func(f): 2023-03-31T06:22:41.4595457Z def g(args): 2023-03-31T06:22:41.4595531Z return f(args) 2023-03-31T06:22:41.4595598Z return g 2023-03-31T06:22:41.4595694Z call = _wrap_func(module.call_75) 2023-03-31T06:22:41.4595700Z 2023-03-31T06:22:41.4595707Z 2023-03-31T06:22:41.4595803Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4595908Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4596034Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4596249Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4596458Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4596667Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4596796Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4596801Z 2023-03-31T06:22:41.4596806Z 2023-03-31T06:22:41.4596884Z if __name__ == "__main__": 2023-03-31T06:22:41.4596960Z import argparse 2023-03-31T06:22:41.4597078Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4597083Z 2023-03-31T06:22:41.4597194Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4597531Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4597895Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4598170Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4598265Z args = parser.parse_args() 2023-03-31T06:22:41.4598271Z 2023-03-31T06:22:41.4598360Z if args.benchmark_kernels: 2023-03-31T06:22:41.4598552Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4598607Z else: 2023-03-31T06:22:41.4598719Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4598813Z benchmark_compiled_module() 2023-03-31T06:22:41.4598818Z 2023-03-31T06:22:41.4598887Z if p: 2023-03-31T06:22:41.4599039Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4599138Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4599273Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4599279Z 2023-03-31T06:22:41.4599558Z [2023-03-31 05:54:02,159] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 63 2023-03-31T06:22:41.4599830Z [2023-03-31 05:54:02,213] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 64 2023-03-31T06:22:41.4600250Z [2023-03-31 05:54:23,345] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3c/c3c6obdsdmqu7l76kdrmg5jlnkqdhcyciwcr626jibbc32rqj37k.py 2023-03-31T06:22:41.4600471Z [2023-03-31 05:54:23,345] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4600476Z 2023-03-31T06:22:41.4600578Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4600651Z import torch 2023-03-31T06:22:41.4600724Z import math 2023-03-31T06:22:41.4600798Z import random 2023-03-31T06:22:41.4600868Z import os 2023-03-31T06:22:41.4600929Z import tempfile 2023-03-31T06:22:41.4601049Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4601055Z 2023-03-31T06:22:41.4601177Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4601304Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4601470Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4601476Z 2023-03-31T06:22:41.4601556Z aten = torch.ops.aten 2023-03-31T06:22:41.4601699Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4601796Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4601801Z 2023-03-31T06:22:41.4601878Z async_compile.wait(globals()) 2023-03-31T06:22:41.4601955Z del async_compile 2023-03-31T06:22:41.4602079Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4602154Z wrapper = ( 2023-03-31T06:22:41.4602237Z ''' 2023-03-31T06:22:41.4602317Z #include 2023-03-31T06:22:41.4602396Z #include 2023-03-31T06:22:41.4602402Z 2023-03-31T06:22:41.4602481Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4602486Z 2023-03-31T06:22:41.4602580Z template 2023-03-31T06:22:41.4602707Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4602795Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4602924Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4603206Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4603337Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4603403Z return kernel_cpp; 2023-03-31T06:22:41.4603467Z } 2023-03-31T06:22:41.4603597Z std::vector call_76(std::vector args) { 2023-03-31T06:22:41.4603705Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4603780Z arg0_1 = args[0]; 2023-03-31T06:22:41.4603855Z arg1_1 = args[1]; 2023-03-31T06:22:41.4603983Z arg2_1 = args[2]; 2023-03-31T06:22:41.4604074Z arg3_1 = args[3]; 2023-03-31T06:22:41.4604137Z 2023-03-31T06:22:41.4604237Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4604332Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4604421Z .findSchemaOrThrow( 2023-03-31T06:22:41.4604521Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4604589Z "") 2023-03-31T06:22:41.4604647Z .typed< 2023-03-31T06:22:41.4604724Z at::Tensor( 2023-03-31T06:22:41.4604824Z const at::Tensor& input_t, 2023-03-31T06:22:41.4604932Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4605052Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4605143Z c10::string_view attr, 2023-03-31T06:22:41.4605265Z torch::List> scalars, 2023-03-31T06:22:41.4605389Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4605446Z 2023-03-31T06:22:41.4605674Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.4605746Z arg1_1.reset(); 2023-03-31T06:22:41.4605819Z arg2_1.reset(); 2023-03-31T06:22:41.4605891Z arg3_1.reset(); 2023-03-31T06:22:41.4605995Z return std::vector({buf0}); 2023-03-31T06:22:41.4606058Z } 2023-03-31T06:22:41.4606126Z ''' 2023-03-31T06:22:41.4606198Z ) 2023-03-31T06:22:41.4606204Z 2023-03-31T06:22:41.4606282Z module = load_inline( 2023-03-31T06:22:41.4606581Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.4606673Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4606789Z functions=['call_76'], 2023-03-31T06:22:41.4607170Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4607314Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4608024Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4608032Z 2023-03-31T06:22:41.4608143Z def _wrap_func(f): 2023-03-31T06:22:41.4608215Z def g(args): 2023-03-31T06:22:41.4608290Z return f(args) 2023-03-31T06:22:41.4608359Z return g 2023-03-31T06:22:41.4608457Z call = _wrap_func(module.call_76) 2023-03-31T06:22:41.4608463Z 2023-03-31T06:22:41.4608468Z 2023-03-31T06:22:41.4608562Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4608679Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4608793Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4609013Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4609223Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4609430Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4609650Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4609783Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4609791Z 2023-03-31T06:22:41.4609796Z 2023-03-31T06:22:41.4609874Z if __name__ == "__main__": 2023-03-31T06:22:41.4609951Z import argparse 2023-03-31T06:22:41.4610068Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4610074Z 2023-03-31T06:22:41.4610183Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4610494Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4610855Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4611170Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4611263Z args = parser.parse_args() 2023-03-31T06:22:41.4611268Z 2023-03-31T06:22:41.4611358Z if args.benchmark_kernels: 2023-03-31T06:22:41.4611551Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4611606Z else: 2023-03-31T06:22:41.4611719Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4611816Z benchmark_compiled_module() 2023-03-31T06:22:41.4611821Z 2023-03-31T06:22:41.4611891Z if p: 2023-03-31T06:22:41.4612040Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4612135Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4612274Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4612282Z 2023-03-31T06:22:41.4612560Z [2023-03-31 05:54:23,346] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 64 2023-03-31T06:22:41.4612827Z [2023-03-31 05:54:23,396] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 65 2023-03-31T06:22:41.4613226Z [2023-03-31 05:54:45,553] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/br/cbrvuabdwirpmeuivefedhge6blgztzoouhxsz6acgt5mbllhkza.py 2023-03-31T06:22:41.4613479Z [2023-03-31 05:54:45,553] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4613485Z 2023-03-31T06:22:41.4613586Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4613658Z import torch 2023-03-31T06:22:41.4613728Z import math 2023-03-31T06:22:41.4613799Z import random 2023-03-31T06:22:41.4613867Z import os 2023-03-31T06:22:41.4613929Z import tempfile 2023-03-31T06:22:41.4614047Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4614055Z 2023-03-31T06:22:41.4614176Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4614302Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4614433Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4614440Z 2023-03-31T06:22:41.4614520Z aten = torch.ops.aten 2023-03-31T06:22:41.4614660Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4614781Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4614787Z 2023-03-31T06:22:41.4614864Z async_compile.wait(globals()) 2023-03-31T06:22:41.4614941Z del async_compile 2023-03-31T06:22:41.4615062Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4615134Z wrapper = ( 2023-03-31T06:22:41.4615216Z ''' 2023-03-31T06:22:41.4615291Z #include 2023-03-31T06:22:41.4615367Z #include 2023-03-31T06:22:41.4615373Z 2023-03-31T06:22:41.4615451Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4615457Z 2023-03-31T06:22:41.4615551Z template 2023-03-31T06:22:41.4615677Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4615762Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4615887Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4615987Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4616110Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4616178Z return kernel_cpp; 2023-03-31T06:22:41.4616239Z } 2023-03-31T06:22:41.4616366Z std::vector call_77(std::vector args) { 2023-03-31T06:22:41.4616461Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4616536Z arg0_1 = args[0]; 2023-03-31T06:22:41.4616609Z arg1_1 = args[1]; 2023-03-31T06:22:41.4616680Z arg2_1 = args[2]; 2023-03-31T06:22:41.4616729Z 2023-03-31T06:22:41.4616826Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4616919Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4617039Z .findSchemaOrThrow( 2023-03-31T06:22:41.4617205Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4617272Z "") 2023-03-31T06:22:41.4617342Z .typed< 2023-03-31T06:22:41.4617405Z at::Tensor( 2023-03-31T06:22:41.4617504Z const at::Tensor& input_t, 2023-03-31T06:22:41.4617604Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4617731Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4617821Z c10::string_view attr, 2023-03-31T06:22:41.4617944Z torch::List> scalars, 2023-03-31T06:22:41.4618070Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4618137Z 2023-03-31T06:22:41.4618360Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.4618435Z arg1_1.reset(); 2023-03-31T06:22:41.4618507Z arg2_1.reset(); 2023-03-31T06:22:41.4618615Z return std::vector({buf0}); 2023-03-31T06:22:41.4618678Z } 2023-03-31T06:22:41.4618758Z ''' 2023-03-31T06:22:41.4618819Z ) 2023-03-31T06:22:41.4618824Z 2023-03-31T06:22:41.4618889Z module = load_inline( 2023-03-31T06:22:41.4619162Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.4619248Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4619363Z functions=['call_77'], 2023-03-31T06:22:41.4619754Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4619934Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4620639Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4620647Z 2023-03-31T06:22:41.4620722Z def _wrap_func(f): 2023-03-31T06:22:41.4620792Z def g(args): 2023-03-31T06:22:41.4620854Z return f(args) 2023-03-31T06:22:41.4620921Z return g 2023-03-31T06:22:41.4621017Z call = _wrap_func(module.call_77) 2023-03-31T06:22:41.4621022Z 2023-03-31T06:22:41.4621027Z 2023-03-31T06:22:41.4621151Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4621268Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4621398Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4621611Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4621809Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4622027Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4622156Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4622163Z 2023-03-31T06:22:41.4622168Z 2023-03-31T06:22:41.4622247Z if __name__ == "__main__": 2023-03-31T06:22:41.4622322Z import argparse 2023-03-31T06:22:41.4622593Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4622601Z 2023-03-31T06:22:41.4622715Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4623035Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4623378Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4623642Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4623735Z args = parser.parse_args() 2023-03-31T06:22:41.4623741Z 2023-03-31T06:22:41.4623831Z if args.benchmark_kernels: 2023-03-31T06:22:41.4624056Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4624153Z else: 2023-03-31T06:22:41.4624269Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4624364Z benchmark_compiled_module() 2023-03-31T06:22:41.4624370Z 2023-03-31T06:22:41.4624439Z if p: 2023-03-31T06:22:41.4624575Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4624676Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4624813Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4624821Z 2023-03-31T06:22:41.4625098Z [2023-03-31 05:54:45,553] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 65 2023-03-31T06:22:41.4625365Z [2023-03-31 05:54:45,607] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 66 2023-03-31T06:22:41.4625775Z [2023-03-31 05:55:07,201] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ek/cekdxzg2i2bxvpzugoojj5jaxdxyimckbscvgispfpcqiay5smle.py 2023-03-31T06:22:41.4625999Z [2023-03-31 05:55:07,201] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4626005Z 2023-03-31T06:22:41.4626101Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4626172Z import torch 2023-03-31T06:22:41.4626231Z import math 2023-03-31T06:22:41.4626307Z import random 2023-03-31T06:22:41.4626378Z import os 2023-03-31T06:22:41.4626451Z import tempfile 2023-03-31T06:22:41.4626567Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4626604Z 2023-03-31T06:22:41.4626723Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4626847Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4626967Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4626972Z 2023-03-31T06:22:41.4627052Z aten = torch.ops.aten 2023-03-31T06:22:41.4627193Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4627289Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4627294Z 2023-03-31T06:22:41.4627385Z async_compile.wait(globals()) 2023-03-31T06:22:41.4627460Z del async_compile 2023-03-31T06:22:41.4627584Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4627643Z wrapper = ( 2023-03-31T06:22:41.4627724Z ''' 2023-03-31T06:22:41.4627800Z #include 2023-03-31T06:22:41.4627903Z #include 2023-03-31T06:22:41.4627908Z 2023-03-31T06:22:41.4628002Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4628007Z 2023-03-31T06:22:41.4628100Z template 2023-03-31T06:22:41.4628223Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4628296Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4628419Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4628514Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4628643Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4628722Z return kernel_cpp; 2023-03-31T06:22:41.4628786Z } 2023-03-31T06:22:41.4628913Z std::vector call_78(std::vector args) { 2023-03-31T06:22:41.4629016Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4629077Z arg0_1 = args[0]; 2023-03-31T06:22:41.4629149Z arg1_1 = args[1]; 2023-03-31T06:22:41.4629220Z arg2_1 = args[2]; 2023-03-31T06:22:41.4629292Z arg3_1 = args[3]; 2023-03-31T06:22:41.4629354Z 2023-03-31T06:22:41.4629452Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4629534Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4629621Z .findSchemaOrThrow( 2023-03-31T06:22:41.4629718Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4629785Z "") 2023-03-31T06:22:41.4629854Z .typed< 2023-03-31T06:22:41.4629928Z at::Tensor( 2023-03-31T06:22:41.4630027Z const at::Tensor& input_t, 2023-03-31T06:22:41.4630149Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4630305Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4630394Z c10::string_view attr, 2023-03-31T06:22:41.4630513Z torch::List> scalars, 2023-03-31T06:22:41.4630632Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4630698Z 2023-03-31T06:22:41.4630922Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.4630995Z arg1_1.reset(); 2023-03-31T06:22:41.4631054Z arg2_1.reset(); 2023-03-31T06:22:41.4631128Z arg3_1.reset(); 2023-03-31T06:22:41.4631235Z return std::vector({buf0}); 2023-03-31T06:22:41.4631299Z } 2023-03-31T06:22:41.4631381Z ''' 2023-03-31T06:22:41.4631444Z ) 2023-03-31T06:22:41.4631450Z 2023-03-31T06:22:41.4631527Z module = load_inline( 2023-03-31T06:22:41.4631777Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.4631866Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4631983Z functions=['call_78'], 2023-03-31T06:22:41.4632368Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4632525Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4633275Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4633285Z 2023-03-31T06:22:41.4633362Z def _wrap_func(f): 2023-03-31T06:22:41.4633435Z def g(args): 2023-03-31T06:22:41.4633497Z return f(args) 2023-03-31T06:22:41.4633564Z return g 2023-03-31T06:22:41.4633659Z call = _wrap_func(module.call_78) 2023-03-31T06:22:41.4633667Z 2023-03-31T06:22:41.4633672Z 2023-03-31T06:22:41.4633769Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4633886Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4634014Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4634228Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4634476Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4634674Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4634881Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4635014Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4635020Z 2023-03-31T06:22:41.4635024Z 2023-03-31T06:22:41.4635101Z if __name__ == "__main__": 2023-03-31T06:22:41.4635177Z import argparse 2023-03-31T06:22:41.4635310Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4635318Z 2023-03-31T06:22:41.4635427Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4635733Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4636071Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4636338Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4636432Z args = parser.parse_args() 2023-03-31T06:22:41.4636438Z 2023-03-31T06:22:41.4636527Z if args.benchmark_kernels: 2023-03-31T06:22:41.4636715Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4636781Z else: 2023-03-31T06:22:41.4636893Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4637018Z benchmark_compiled_module() 2023-03-31T06:22:41.4637047Z 2023-03-31T06:22:41.4637115Z if p: 2023-03-31T06:22:41.4637252Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4637348Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4637486Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4637495Z 2023-03-31T06:22:41.4637772Z [2023-03-31 05:55:07,202] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 66 2023-03-31T06:22:41.4638040Z [2023-03-31 05:55:07,241] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 67 2023-03-31T06:22:41.4638436Z [2023-03-31 05:55:29,005] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/qs/cqs655l6u4nt6jilq35ky3uvewukprxh7wsz44ilskab6uki73f3.py 2023-03-31T06:22:41.4638658Z [2023-03-31 05:55:29,005] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4638666Z 2023-03-31T06:22:41.4638763Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4638836Z import torch 2023-03-31T06:22:41.4638894Z import math 2023-03-31T06:22:41.4638967Z import random 2023-03-31T06:22:41.4639033Z import os 2023-03-31T06:22:41.4639107Z import tempfile 2023-03-31T06:22:41.4639226Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4639231Z 2023-03-31T06:22:41.4639354Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4639480Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4639628Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4639635Z 2023-03-31T06:22:41.4639718Z aten = torch.ops.aten 2023-03-31T06:22:41.4639858Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4639955Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4639961Z 2023-03-31T06:22:41.4640054Z async_compile.wait(globals()) 2023-03-31T06:22:41.4640137Z del async_compile 2023-03-31T06:22:41.4640262Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4640322Z wrapper = ( 2023-03-31T06:22:41.4640406Z ''' 2023-03-31T06:22:41.4640484Z #include 2023-03-31T06:22:41.4640561Z #include 2023-03-31T06:22:41.4640566Z 2023-03-31T06:22:41.4640660Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4640666Z 2023-03-31T06:22:41.4640760Z template 2023-03-31T06:22:41.4640917Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4640991Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4641120Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4641219Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4641346Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4641426Z return kernel_cpp; 2023-03-31T06:22:41.4641491Z } 2023-03-31T06:22:41.4641619Z std::vector call_79(std::vector args) { 2023-03-31T06:22:41.4641702Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4641784Z arg0_1 = args[0]; 2023-03-31T06:22:41.4641856Z arg1_1 = args[1]; 2023-03-31T06:22:41.4641928Z arg2_1 = args[2]; 2023-03-31T06:22:41.4641994Z 2023-03-31T06:22:41.4642093Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4642188Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4642263Z .findSchemaOrThrow( 2023-03-31T06:22:41.4642366Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4642434Z "") 2023-03-31T06:22:41.4642503Z .typed< 2023-03-31T06:22:41.4642581Z at::Tensor( 2023-03-31T06:22:41.4642682Z const at::Tensor& input_t, 2023-03-31T06:22:41.4642785Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4642890Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4642980Z c10::string_view attr, 2023-03-31T06:22:41.4643287Z torch::List> scalars, 2023-03-31T06:22:41.4643445Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4643513Z 2023-03-31T06:22:41.4643752Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.4643826Z arg1_1.reset(); 2023-03-31T06:22:41.4643899Z arg2_1.reset(); 2023-03-31T06:22:41.4643993Z return std::vector({buf0}); 2023-03-31T06:22:41.4644059Z } 2023-03-31T06:22:41.4644141Z ''' 2023-03-31T06:22:41.4644204Z ) 2023-03-31T06:22:41.4644209Z 2023-03-31T06:22:41.4644289Z module = load_inline( 2023-03-31T06:22:41.4644562Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.4644649Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4644753Z functions=['call_79'], 2023-03-31T06:22:41.4645140Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4645297Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4646005Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4646014Z 2023-03-31T06:22:41.4646089Z def _wrap_func(f): 2023-03-31T06:22:41.4646161Z def g(args): 2023-03-31T06:22:41.4646270Z return f(args) 2023-03-31T06:22:41.4646339Z return g 2023-03-31T06:22:41.4646433Z call = _wrap_func(module.call_79) 2023-03-31T06:22:41.4646439Z 2023-03-31T06:22:41.4646445Z 2023-03-31T06:22:41.4646528Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4646644Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4646774Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4646993Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4647203Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4647411Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4647537Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4647583Z 2023-03-31T06:22:41.4647588Z 2023-03-31T06:22:41.4647666Z if __name__ == "__main__": 2023-03-31T06:22:41.4647730Z import argparse 2023-03-31T06:22:41.4647864Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4647869Z 2023-03-31T06:22:41.4647980Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4648289Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4648628Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4648907Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4649000Z args = parser.parse_args() 2023-03-31T06:22:41.4649005Z 2023-03-31T06:22:41.4649094Z if args.benchmark_kernels: 2023-03-31T06:22:41.4649287Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4649343Z else: 2023-03-31T06:22:41.4649456Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4649556Z benchmark_compiled_module() 2023-03-31T06:22:41.4649561Z 2023-03-31T06:22:41.4649630Z if p: 2023-03-31T06:22:41.4649779Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4649876Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4650013Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4650045Z 2023-03-31T06:22:41.4650354Z [2023-03-31 05:55:29,006] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 67 2023-03-31T06:22:41.4650609Z [2023-03-31 05:55:29,061] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 68 2023-03-31T06:22:41.4651011Z [2023-03-31 05:55:50,601] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/d4/cd4tucm5exx4yqkinxq6uxeds4trmsmvvn74qcu56qg3jy6bq455.py 2023-03-31T06:22:41.4651233Z [2023-03-31 05:55:50,601] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4651241Z 2023-03-31T06:22:41.4651341Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4651413Z import torch 2023-03-31T06:22:41.4651483Z import math 2023-03-31T06:22:41.4651554Z import random 2023-03-31T06:22:41.4651609Z import os 2023-03-31T06:22:41.4651682Z import tempfile 2023-03-31T06:22:41.4651798Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4651804Z 2023-03-31T06:22:41.4651923Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4652050Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4652183Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4652189Z 2023-03-31T06:22:41.4652270Z aten = torch.ops.aten 2023-03-31T06:22:41.4652409Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4652494Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4652500Z 2023-03-31T06:22:41.4652589Z async_compile.wait(globals()) 2023-03-31T06:22:41.4652663Z del async_compile 2023-03-31T06:22:41.4652812Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4652888Z wrapper = ( 2023-03-31T06:22:41.4652971Z ''' 2023-03-31T06:22:41.4653047Z #include 2023-03-31T06:22:41.4653110Z #include 2023-03-31T06:22:41.4653115Z 2023-03-31T06:22:41.4653205Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4653211Z 2023-03-31T06:22:41.4653304Z template 2023-03-31T06:22:41.4653429Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4653515Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4653638Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4653735Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4653859Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4653956Z return kernel_cpp; 2023-03-31T06:22:41.4654018Z } 2023-03-31T06:22:41.4654147Z std::vector call_80(std::vector args) { 2023-03-31T06:22:41.4654253Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4654326Z arg0_1 = args[0]; 2023-03-31T06:22:41.4654398Z arg1_1 = args[1]; 2023-03-31T06:22:41.4654468Z arg2_1 = args[2]; 2023-03-31T06:22:41.4654527Z arg3_1 = args[3]; 2023-03-31T06:22:41.4654589Z 2023-03-31T06:22:41.4654686Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4654779Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4654867Z .findSchemaOrThrow( 2023-03-31T06:22:41.4654969Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4655023Z "") 2023-03-31T06:22:41.4655090Z .typed< 2023-03-31T06:22:41.4655165Z at::Tensor( 2023-03-31T06:22:41.4655260Z const at::Tensor& input_t, 2023-03-31T06:22:41.4655362Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4655480Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4655568Z c10::string_view attr, 2023-03-31T06:22:41.4655691Z torch::List> scalars, 2023-03-31T06:22:41.4655800Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4655865Z 2023-03-31T06:22:41.4656091Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4656163Z arg1_1.reset(); 2023-03-31T06:22:41.4656235Z arg2_1.reset(); 2023-03-31T06:22:41.4656329Z arg3_1.reset(); 2023-03-31T06:22:41.4656471Z return std::vector({buf0}); 2023-03-31T06:22:41.4656521Z } 2023-03-31T06:22:41.4656602Z ''' 2023-03-31T06:22:41.4656664Z ) 2023-03-31T06:22:41.4656669Z 2023-03-31T06:22:41.4656747Z module = load_inline( 2023-03-31T06:22:41.4657018Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.4657106Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4657220Z functions=['call_80'], 2023-03-31T06:22:41.4657596Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4657752Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4658461Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4658470Z 2023-03-31T06:22:41.4658547Z def _wrap_func(f): 2023-03-31T06:22:41.4658620Z def g(args): 2023-03-31T06:22:41.4658696Z return f(args) 2023-03-31T06:22:41.4658765Z return g 2023-03-31T06:22:41.4658860Z call = _wrap_func(module.call_80) 2023-03-31T06:22:41.4658868Z 2023-03-31T06:22:41.4658873Z 2023-03-31T06:22:41.4658969Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4659074Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4659227Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4659444Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4659652Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4659863Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4660080Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4660214Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4660219Z 2023-03-31T06:22:41.4660224Z 2023-03-31T06:22:41.4660302Z if __name__ == "__main__": 2023-03-31T06:22:41.4660365Z import argparse 2023-03-31T06:22:41.4660528Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4660535Z 2023-03-31T06:22:41.4660643Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4660956Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4661292Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4661568Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4661662Z args = parser.parse_args() 2023-03-31T06:22:41.4661667Z 2023-03-31T06:22:41.4661756Z if args.benchmark_kernels: 2023-03-31T06:22:41.4661944Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4661999Z else: 2023-03-31T06:22:41.4662111Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4662206Z benchmark_compiled_module() 2023-03-31T06:22:41.4662214Z 2023-03-31T06:22:41.4662281Z if p: 2023-03-31T06:22:41.4662516Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4662617Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4662756Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4662761Z 2023-03-31T06:22:41.4663042Z [2023-03-31 05:55:50,602] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 68 2023-03-31T06:22:41.4663330Z [2023-03-31 05:55:50,644] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 69 2023-03-31T06:22:41.4663761Z [2023-03-31 05:56:12,227] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/vh/cvh7v5p7qozeotxbkkaukiav6kzur7lmufufqrjjh7e4ob5xtbar.py 2023-03-31T06:22:41.4663982Z [2023-03-31 05:56:12,227] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4663990Z 2023-03-31T06:22:41.4664088Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4664160Z import torch 2023-03-31T06:22:41.4664232Z import math 2023-03-31T06:22:41.4664305Z import random 2023-03-31T06:22:41.4664362Z import os 2023-03-31T06:22:41.4664435Z import tempfile 2023-03-31T06:22:41.4664552Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4664557Z 2023-03-31T06:22:41.4664676Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4664801Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4664936Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4664944Z 2023-03-31T06:22:41.4665024Z aten = torch.ops.aten 2023-03-31T06:22:41.4665162Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4665245Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4665250Z 2023-03-31T06:22:41.4665340Z async_compile.wait(globals()) 2023-03-31T06:22:41.4665415Z del async_compile 2023-03-31T06:22:41.4665537Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4665610Z wrapper = ( 2023-03-31T06:22:41.4665691Z ''' 2023-03-31T06:22:41.4665767Z #include 2023-03-31T06:22:41.4665857Z #include 2023-03-31T06:22:41.4665863Z 2023-03-31T06:22:41.4665956Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4665961Z 2023-03-31T06:22:41.4666053Z template 2023-03-31T06:22:41.4666175Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4666260Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4666389Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4666489Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4666614Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4666680Z return kernel_cpp; 2023-03-31T06:22:41.4666742Z } 2023-03-31T06:22:41.4666868Z std::vector call_81(std::vector args) { 2023-03-31T06:22:41.4666961Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4667061Z arg0_1 = args[0]; 2023-03-31T06:22:41.4667132Z arg1_1 = args[1]; 2023-03-31T06:22:41.4667202Z arg2_1 = args[2]; 2023-03-31T06:22:41.4667253Z 2023-03-31T06:22:41.4667351Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4667443Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4667529Z .findSchemaOrThrow( 2023-03-31T06:22:41.4667626Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4667692Z "") 2023-03-31T06:22:41.4667747Z .typed< 2023-03-31T06:22:41.4667823Z at::Tensor( 2023-03-31T06:22:41.4667922Z const at::Tensor& input_t, 2023-03-31T06:22:41.4668023Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4668139Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4668228Z c10::string_view attr, 2023-03-31T06:22:41.4668348Z torch::List> scalars, 2023-03-31T06:22:41.4668472Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4668525Z 2023-03-31T06:22:41.4668765Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4668837Z arg1_1.reset(); 2023-03-31T06:22:41.4668908Z arg2_1.reset(); 2023-03-31T06:22:41.4669013Z return std::vector({buf0}); 2023-03-31T06:22:41.4669075Z } 2023-03-31T06:22:41.4669156Z ''' 2023-03-31T06:22:41.4669204Z ) 2023-03-31T06:22:41.4669210Z 2023-03-31T06:22:41.4669286Z module = load_inline( 2023-03-31T06:22:41.4669616Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.4669730Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4669848Z functions=['call_81'], 2023-03-31T06:22:41.4670229Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4670387Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4671100Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4671107Z 2023-03-31T06:22:41.4671186Z def _wrap_func(f): 2023-03-31T06:22:41.4671244Z def g(args): 2023-03-31T06:22:41.4671322Z return f(args) 2023-03-31T06:22:41.4671393Z return g 2023-03-31T06:22:41.4671490Z call = _wrap_func(module.call_81) 2023-03-31T06:22:41.4671496Z 2023-03-31T06:22:41.4671500Z 2023-03-31T06:22:41.4671595Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4671714Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4671842Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4672046Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4672259Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4672502Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4672631Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4672637Z 2023-03-31T06:22:41.4672642Z 2023-03-31T06:22:41.4672719Z if __name__ == "__main__": 2023-03-31T06:22:41.4672795Z import argparse 2023-03-31T06:22:41.4672926Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4672934Z 2023-03-31T06:22:41.4673044Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4673359Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4673686Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4673991Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4674084Z args = parser.parse_args() 2023-03-31T06:22:41.4674090Z 2023-03-31T06:22:41.4674183Z if args.benchmark_kernels: 2023-03-31T06:22:41.4674372Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4674440Z else: 2023-03-31T06:22:41.4674555Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4674653Z benchmark_compiled_module() 2023-03-31T06:22:41.4674661Z 2023-03-31T06:22:41.4674717Z if p: 2023-03-31T06:22:41.4674865Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4674961Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4675099Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4675105Z 2023-03-31T06:22:41.4675386Z [2023-03-31 05:56:12,227] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 69 2023-03-31T06:22:41.4675655Z [2023-03-31 05:56:12,281] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 70 2023-03-31T06:22:41.4676052Z [2023-03-31 05:56:33,598] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/lm/clm72zrhdyq4f4juo3ydudji2cu7q5w2bz4effp7ymb73g6rtp3q.py 2023-03-31T06:22:41.4676272Z [2023-03-31 05:56:33,598] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4676303Z 2023-03-31T06:22:41.4676429Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4676489Z import torch 2023-03-31T06:22:41.4676559Z import math 2023-03-31T06:22:41.4676631Z import random 2023-03-31T06:22:41.4676698Z import os 2023-03-31T06:22:41.4676773Z import tempfile 2023-03-31T06:22:41.4676890Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4676896Z 2023-03-31T06:22:41.4677017Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4677131Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4677266Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4677272Z 2023-03-31T06:22:41.4677354Z aten = torch.ops.aten 2023-03-31T06:22:41.4677498Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4677594Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4677600Z 2023-03-31T06:22:41.4677695Z async_compile.wait(globals()) 2023-03-31T06:22:41.4677772Z del async_compile 2023-03-31T06:22:41.4677884Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4677960Z wrapper = ( 2023-03-31T06:22:41.4678043Z ''' 2023-03-31T06:22:41.4678119Z #include 2023-03-31T06:22:41.4678197Z #include 2023-03-31T06:22:41.4678202Z 2023-03-31T06:22:41.4678296Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4678301Z 2023-03-31T06:22:41.4678395Z template 2023-03-31T06:22:41.4678523Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4678598Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4678751Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4678851Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4678975Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4679057Z return kernel_cpp; 2023-03-31T06:22:41.4679121Z } 2023-03-31T06:22:41.4679250Z std::vector call_82(std::vector args) { 2023-03-31T06:22:41.4679347Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4679427Z arg0_1 = args[0]; 2023-03-31T06:22:41.4679503Z arg1_1 = args[1]; 2023-03-31T06:22:41.4679576Z arg2_1 = args[2]; 2023-03-31T06:22:41.4679647Z arg3_1 = args[3]; 2023-03-31T06:22:41.4679715Z 2023-03-31T06:22:41.4679800Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4679895Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4680025Z .findSchemaOrThrow( 2023-03-31T06:22:41.4680124Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4680196Z "") 2023-03-31T06:22:41.4680265Z .typed< 2023-03-31T06:22:41.4680344Z at::Tensor( 2023-03-31T06:22:41.4680430Z const at::Tensor& input_t, 2023-03-31T06:22:41.4680533Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4680651Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4680742Z c10::string_view attr, 2023-03-31T06:22:41.4680867Z torch::List> scalars, 2023-03-31T06:22:41.4680992Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4681059Z 2023-03-31T06:22:41.4681289Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4681350Z arg1_1.reset(); 2023-03-31T06:22:41.4681422Z arg2_1.reset(); 2023-03-31T06:22:41.4681493Z arg3_1.reset(); 2023-03-31T06:22:41.4681603Z return std::vector({buf0}); 2023-03-31T06:22:41.4681668Z } 2023-03-31T06:22:41.4681750Z ''' 2023-03-31T06:22:41.4681813Z ) 2023-03-31T06:22:41.4681821Z 2023-03-31T06:22:41.4681888Z module = load_inline( 2023-03-31T06:22:41.4682160Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.4682249Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4682364Z functions=['call_82'], 2023-03-31T06:22:41.4682775Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4682969Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4683922Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4683933Z 2023-03-31T06:22:41.4684010Z def _wrap_func(f): 2023-03-31T06:22:41.4684084Z def g(args): 2023-03-31T06:22:41.4684147Z return f(args) 2023-03-31T06:22:41.4684215Z return g 2023-03-31T06:22:41.4684312Z call = _wrap_func(module.call_82) 2023-03-31T06:22:41.4684318Z 2023-03-31T06:22:41.4684322Z 2023-03-31T06:22:41.4684417Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4684538Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4684669Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4684886Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4685080Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4685290Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4685499Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4685635Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4685693Z 2023-03-31T06:22:41.4685698Z 2023-03-31T06:22:41.4685782Z if __name__ == "__main__": 2023-03-31T06:22:41.4685860Z import argparse 2023-03-31T06:22:41.4685992Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4685998Z 2023-03-31T06:22:41.4686110Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4686426Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4686756Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4687033Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4687168Z args = parser.parse_args() 2023-03-31T06:22:41.4687173Z 2023-03-31T06:22:41.4687264Z if args.benchmark_kernels: 2023-03-31T06:22:41.4687458Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4687527Z else: 2023-03-31T06:22:41.4687640Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4687736Z benchmark_compiled_module() 2023-03-31T06:22:41.4687741Z 2023-03-31T06:22:41.4687797Z if p: 2023-03-31T06:22:41.4687947Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4688046Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4688184Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4688189Z 2023-03-31T06:22:41.4688468Z [2023-03-31 05:56:33,599] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 70 2023-03-31T06:22:41.4688732Z [2023-03-31 05:56:33,646] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 71 2023-03-31T06:22:41.4689141Z [2023-03-31 05:56:54,921] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/yl/cyl7goeopu3vqli5wlhbn6xeodsu4qrxtw4kwze2rug27arj2k62.py 2023-03-31T06:22:41.4689361Z [2023-03-31 05:56:54,921] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4689368Z 2023-03-31T06:22:41.4689464Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4689523Z import torch 2023-03-31T06:22:41.4689593Z import math 2023-03-31T06:22:41.4689665Z import random 2023-03-31T06:22:41.4689765Z import os 2023-03-31T06:22:41.4689870Z import tempfile 2023-03-31T06:22:41.4689987Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4689992Z 2023-03-31T06:22:41.4690112Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4690226Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4690363Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4690371Z 2023-03-31T06:22:41.4690452Z aten = torch.ops.aten 2023-03-31T06:22:41.4690590Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4690688Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4690694Z 2023-03-31T06:22:41.4690785Z async_compile.wait(globals()) 2023-03-31T06:22:41.4690859Z del async_compile 2023-03-31T06:22:41.4690968Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4691039Z wrapper = ( 2023-03-31T06:22:41.4691123Z ''' 2023-03-31T06:22:41.4691200Z #include 2023-03-31T06:22:41.4691279Z #include 2023-03-31T06:22:41.4691286Z 2023-03-31T06:22:41.4691378Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4691383Z 2023-03-31T06:22:41.4691474Z template 2023-03-31T06:22:41.4691598Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4691672Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4691795Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4691895Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4692021Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4692125Z return kernel_cpp; 2023-03-31T06:22:41.4692188Z } 2023-03-31T06:22:41.4692317Z std::vector call_83(std::vector args) { 2023-03-31T06:22:41.4692397Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4692469Z arg0_1 = args[0]; 2023-03-31T06:22:41.4692541Z arg1_1 = args[1]; 2023-03-31T06:22:41.4692612Z arg2_1 = args[2]; 2023-03-31T06:22:41.4692674Z 2023-03-31T06:22:41.4692774Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4692868Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4692943Z .findSchemaOrThrow( 2023-03-31T06:22:41.4693039Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4693106Z "") 2023-03-31T06:22:41.4693174Z .typed< 2023-03-31T06:22:41.4693249Z at::Tensor( 2023-03-31T06:22:41.4693384Z const at::Tensor& input_t, 2023-03-31T06:22:41.4693485Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4693593Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4693681Z c10::string_view attr, 2023-03-31T06:22:41.4693802Z torch::List> scalars, 2023-03-31T06:22:41.4693922Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4693987Z 2023-03-31T06:22:41.4694231Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.4694306Z arg1_1.reset(); 2023-03-31T06:22:41.4694365Z arg2_1.reset(); 2023-03-31T06:22:41.4694469Z return std::vector({buf0}); 2023-03-31T06:22:41.4694532Z } 2023-03-31T06:22:41.4694612Z ''' 2023-03-31T06:22:41.4694673Z ) 2023-03-31T06:22:41.4694678Z 2023-03-31T06:22:41.4694756Z module = load_inline( 2023-03-31T06:22:41.4695018Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.4695094Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4695208Z functions=['call_83'], 2023-03-31T06:22:41.4695589Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4695743Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4696473Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4696505Z 2023-03-31T06:22:41.4696581Z def _wrap_func(f): 2023-03-31T06:22:41.4696654Z def g(args): 2023-03-31T06:22:41.4696728Z return f(args) 2023-03-31T06:22:41.4696797Z return g 2023-03-31T06:22:41.4696878Z call = _wrap_func(module.call_83) 2023-03-31T06:22:41.4696884Z 2023-03-31T06:22:41.4696888Z 2023-03-31T06:22:41.4696984Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4697101Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4697228Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4697441Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4697652Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4697862Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4697989Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4697994Z 2023-03-31T06:22:41.4697999Z 2023-03-31T06:22:41.4698076Z if __name__ == "__main__": 2023-03-31T06:22:41.4698139Z import argparse 2023-03-31T06:22:41.4698268Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4698275Z 2023-03-31T06:22:41.4698385Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4698723Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4699068Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4699347Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4699444Z args = parser.parse_args() 2023-03-31T06:22:41.4699450Z 2023-03-31T06:22:41.4699540Z if args.benchmark_kernels: 2023-03-31T06:22:41.4699717Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4699784Z else: 2023-03-31T06:22:41.4699896Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4699991Z benchmark_compiled_module() 2023-03-31T06:22:41.4700026Z 2023-03-31T06:22:41.4700095Z if p: 2023-03-31T06:22:41.4700243Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4700341Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4700466Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4700483Z 2023-03-31T06:22:41.4700754Z [2023-03-31 05:56:54,922] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 71 2023-03-31T06:22:41.4701022Z [2023-03-31 05:56:54,976] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 72 2023-03-31T06:22:41.4701436Z [2023-03-31 05:57:16,393] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/bu/cbuuw7vlmhrkdzqvpofblmx6xf7cysfcq7ckloxh2g7tcrejwz6n.py 2023-03-31T06:22:41.4701654Z [2023-03-31 05:57:16,393] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4701660Z 2023-03-31T06:22:41.4701760Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4701833Z import torch 2023-03-31T06:22:41.4701904Z import math 2023-03-31T06:22:41.4701978Z import random 2023-03-31T06:22:41.4702036Z import os 2023-03-31T06:22:41.4702112Z import tempfile 2023-03-31T06:22:41.4702229Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4702235Z 2023-03-31T06:22:41.4702354Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4702578Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4702750Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4702793Z 2023-03-31T06:22:41.4702875Z aten = torch.ops.aten 2023-03-31T06:22:41.4703001Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4703095Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4703100Z 2023-03-31T06:22:41.4703195Z async_compile.wait(globals()) 2023-03-31T06:22:41.4703273Z del async_compile 2023-03-31T06:22:41.4703395Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4703471Z wrapper = ( 2023-03-31T06:22:41.4703556Z ''' 2023-03-31T06:22:41.4703618Z #include 2023-03-31T06:22:41.4703700Z #include 2023-03-31T06:22:41.4703705Z 2023-03-31T06:22:41.4703800Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4703805Z 2023-03-31T06:22:41.4703901Z template 2023-03-31T06:22:41.4704025Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4704113Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4704243Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4704344Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4704457Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4704536Z return kernel_cpp; 2023-03-31T06:22:41.4704601Z } 2023-03-31T06:22:41.4704729Z std::vector call_84(std::vector args) { 2023-03-31T06:22:41.4704832Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4704907Z arg0_1 = args[0]; 2023-03-31T06:22:41.4704979Z arg1_1 = args[1]; 2023-03-31T06:22:41.4705038Z arg2_1 = args[2]; 2023-03-31T06:22:41.4705140Z arg3_1 = args[3]; 2023-03-31T06:22:41.4705203Z 2023-03-31T06:22:41.4705300Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4705393Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4705481Z .findSchemaOrThrow( 2023-03-31T06:22:41.4705579Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4705633Z "") 2023-03-31T06:22:41.4705701Z .typed< 2023-03-31T06:22:41.4705778Z at::Tensor( 2023-03-31T06:22:41.4705875Z const at::Tensor& input_t, 2023-03-31T06:22:41.4705974Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4706094Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4706182Z c10::string_view attr, 2023-03-31T06:22:41.4706287Z torch::List> scalars, 2023-03-31T06:22:41.4706438Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4706504Z 2023-03-31T06:22:41.4706726Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.4706798Z arg1_1.reset(); 2023-03-31T06:22:41.4706870Z arg2_1.reset(); 2023-03-31T06:22:41.4706939Z arg3_1.reset(); 2023-03-31T06:22:41.4707031Z return std::vector({buf0}); 2023-03-31T06:22:41.4707093Z } 2023-03-31T06:22:41.4707173Z ''' 2023-03-31T06:22:41.4707235Z ) 2023-03-31T06:22:41.4707240Z 2023-03-31T06:22:41.4707319Z module = load_inline( 2023-03-31T06:22:41.4707585Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.4707671Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4707774Z functions=['call_84'], 2023-03-31T06:22:41.4708153Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4708314Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4709015Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4709022Z 2023-03-31T06:22:41.4709098Z def _wrap_func(f): 2023-03-31T06:22:41.4709200Z def g(args): 2023-03-31T06:22:41.4709303Z return f(args) 2023-03-31T06:22:41.4709370Z return g 2023-03-31T06:22:41.4709464Z call = _wrap_func(module.call_84) 2023-03-31T06:22:41.4709470Z 2023-03-31T06:22:41.4709475Z 2023-03-31T06:22:41.4709557Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4709675Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4709807Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4710020Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4710231Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4710440Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4710655Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4710787Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4710795Z 2023-03-31T06:22:41.4710802Z 2023-03-31T06:22:41.4710881Z if __name__ == "__main__": 2023-03-31T06:22:41.4710944Z import argparse 2023-03-31T06:22:41.4711072Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4711078Z 2023-03-31T06:22:41.4711187Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4711496Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4711865Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4712139Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4712232Z args = parser.parse_args() 2023-03-31T06:22:41.4712237Z 2023-03-31T06:22:41.4712326Z if args.benchmark_kernels: 2023-03-31T06:22:41.4712506Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4712574Z else: 2023-03-31T06:22:41.4712687Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4712780Z benchmark_compiled_module() 2023-03-31T06:22:41.4712785Z 2023-03-31T06:22:41.4712856Z if p: 2023-03-31T06:22:41.4713005Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4713135Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4713260Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4713278Z 2023-03-31T06:22:41.4713548Z [2023-03-31 05:57:16,394] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 72 2023-03-31T06:22:41.4713813Z [2023-03-31 05:57:16,434] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 73 2023-03-31T06:22:41.4714222Z [2023-03-31 05:57:37,690] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/py/cpyt3xd7l6z5ateuz3sdavrrriuygzku6mvum2jbedxow6xb7dgu.py 2023-03-31T06:22:41.4714445Z [2023-03-31 05:57:37,690] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4714451Z 2023-03-31T06:22:41.4714548Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4714620Z import torch 2023-03-31T06:22:41.4714691Z import math 2023-03-31T06:22:41.4714762Z import random 2023-03-31T06:22:41.4714817Z import os 2023-03-31T06:22:41.4714894Z import tempfile 2023-03-31T06:22:41.4715011Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4715017Z 2023-03-31T06:22:41.4715138Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4715262Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4715396Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4715407Z 2023-03-31T06:22:41.4715486Z aten = torch.ops.aten 2023-03-31T06:22:41.4715612Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4715741Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4715772Z 2023-03-31T06:22:41.4715866Z async_compile.wait(globals()) 2023-03-31T06:22:41.4715939Z del async_compile 2023-03-31T06:22:41.4716062Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4716136Z wrapper = ( 2023-03-31T06:22:41.4716220Z ''' 2023-03-31T06:22:41.4716283Z #include 2023-03-31T06:22:41.4716362Z #include 2023-03-31T06:22:41.4716370Z 2023-03-31T06:22:41.4716463Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4716468Z 2023-03-31T06:22:41.4716562Z template 2023-03-31T06:22:41.4716689Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4716778Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4716904Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4717003Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4717115Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4717197Z return kernel_cpp; 2023-03-31T06:22:41.4717264Z } 2023-03-31T06:22:41.4717392Z std::vector call_85(std::vector args) { 2023-03-31T06:22:41.4717488Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4717563Z arg0_1 = args[0]; 2023-03-31T06:22:41.4717638Z arg1_1 = args[1]; 2023-03-31T06:22:41.4717696Z arg2_1 = args[2]; 2023-03-31T06:22:41.4717760Z 2023-03-31T06:22:41.4717863Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4717957Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4718046Z .findSchemaOrThrow( 2023-03-31T06:22:41.4718175Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4718245Z "") 2023-03-31T06:22:41.4718300Z .typed< 2023-03-31T06:22:41.4718375Z at::Tensor( 2023-03-31T06:22:41.4718473Z const at::Tensor& input_t, 2023-03-31T06:22:41.4718574Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4718695Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4718788Z c10::string_view attr, 2023-03-31T06:22:41.4718911Z torch::List> scalars, 2023-03-31T06:22:41.4719019Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4719087Z 2023-03-31T06:22:41.4719324Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.4719470Z arg1_1.reset(); 2023-03-31T06:22:41.4719544Z arg2_1.reset(); 2023-03-31T06:22:41.4719651Z return std::vector({buf0}); 2023-03-31T06:22:41.4719715Z } 2023-03-31T06:22:41.4719786Z ''' 2023-03-31T06:22:41.4719850Z ) 2023-03-31T06:22:41.4719855Z 2023-03-31T06:22:41.4719937Z module = load_inline( 2023-03-31T06:22:41.4720204Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.4720291Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4720409Z functions=['call_85'], 2023-03-31T06:22:41.4720794Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4720952Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4721653Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4721675Z 2023-03-31T06:22:41.4721739Z def _wrap_func(f): 2023-03-31T06:22:41.4721812Z def g(args): 2023-03-31T06:22:41.4721889Z return f(args) 2023-03-31T06:22:41.4721958Z return g 2023-03-31T06:22:41.4722053Z call = _wrap_func(module.call_85) 2023-03-31T06:22:41.4722059Z 2023-03-31T06:22:41.4722065Z 2023-03-31T06:22:41.4722189Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4722330Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4722459Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4722661Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4722873Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4723253Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4723383Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4723391Z 2023-03-31T06:22:41.4723396Z 2023-03-31T06:22:41.4723478Z if __name__ == "__main__": 2023-03-31T06:22:41.4723556Z import argparse 2023-03-31T06:22:41.4723691Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4723697Z 2023-03-31T06:22:41.4723808Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4724114Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4724456Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4724731Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4724829Z args = parser.parse_args() 2023-03-31T06:22:41.4724834Z 2023-03-31T06:22:41.4724924Z if args.benchmark_kernels: 2023-03-31T06:22:41.4725173Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4725243Z else: 2023-03-31T06:22:41.4725359Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4725441Z benchmark_compiled_module() 2023-03-31T06:22:41.4725460Z 2023-03-31T06:22:41.4725516Z if p: 2023-03-31T06:22:41.4725665Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4725765Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4725906Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4725912Z 2023-03-31T06:22:41.4726195Z [2023-03-31 05:57:37,691] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 73 2023-03-31T06:22:41.4726460Z [2023-03-31 05:57:37,759] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 74 2023-03-31T06:22:41.4726915Z [2023-03-31 05:57:59,186] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ai/caijhs6j6whw5gg64sku527u2hyodzy4q7wi4g2wkmqbdhmtslwa.py 2023-03-31T06:22:41.4727133Z [2023-03-31 05:57:59,186] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4727139Z 2023-03-31T06:22:41.4727223Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4727294Z import torch 2023-03-31T06:22:41.4727363Z import math 2023-03-31T06:22:41.4727434Z import random 2023-03-31T06:22:41.4727505Z import os 2023-03-31T06:22:41.4727581Z import tempfile 2023-03-31T06:22:41.4727697Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4727703Z 2023-03-31T06:22:41.4727809Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4727934Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4728067Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4728076Z 2023-03-31T06:22:41.4728157Z aten = torch.ops.aten 2023-03-31T06:22:41.4728296Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4728398Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4728403Z 2023-03-31T06:22:41.4728495Z async_compile.wait(globals()) 2023-03-31T06:22:41.4728569Z del async_compile 2023-03-31T06:22:41.4728678Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4728749Z wrapper = ( 2023-03-31T06:22:41.4728829Z ''' 2023-03-31T06:22:41.4728905Z #include 2023-03-31T06:22:41.4728981Z #include 2023-03-31T06:22:41.4729021Z 2023-03-31T06:22:41.4729147Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4729152Z 2023-03-31T06:22:41.4729243Z template 2023-03-31T06:22:41.4729354Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4729443Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4729567Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4729667Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4729792Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4729871Z return kernel_cpp; 2023-03-31T06:22:41.4729934Z } 2023-03-31T06:22:41.4730050Z std::vector call_86(std::vector args) { 2023-03-31T06:22:41.4730152Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4730224Z arg0_1 = args[0]; 2023-03-31T06:22:41.4730294Z arg1_1 = args[1]; 2023-03-31T06:22:41.4730365Z arg2_1 = args[2]; 2023-03-31T06:22:41.4730435Z arg3_1 = args[3]; 2023-03-31T06:22:41.4730499Z 2023-03-31T06:22:41.4730586Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4730679Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4730766Z .findSchemaOrThrow( 2023-03-31T06:22:41.4730864Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4730930Z "") 2023-03-31T06:22:41.4730997Z .typed< 2023-03-31T06:22:41.4731070Z at::Tensor( 2023-03-31T06:22:41.4731158Z const at::Tensor& input_t, 2023-03-31T06:22:41.4731257Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4731403Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4731492Z c10::string_view attr, 2023-03-31T06:22:41.4731613Z torch::List> scalars, 2023-03-31T06:22:41.4731732Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4731797Z 2023-03-31T06:22:41.4732010Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.4732084Z arg1_1.reset(); 2023-03-31T06:22:41.4732155Z arg2_1.reset(); 2023-03-31T06:22:41.4732224Z arg3_1.reset(); 2023-03-31T06:22:41.4732329Z return std::vector({buf0}); 2023-03-31T06:22:41.4732392Z } 2023-03-31T06:22:41.4732472Z ''' 2023-03-31T06:22:41.4732521Z ) 2023-03-31T06:22:41.4732526Z 2023-03-31T06:22:41.4732635Z module = load_inline( 2023-03-31T06:22:41.4732902Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.4732988Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4733107Z functions=['call_86'], 2023-03-31T06:22:41.4733488Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4733643Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4734345Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4734353Z 2023-03-31T06:22:41.4734428Z def _wrap_func(f): 2023-03-31T06:22:41.4734487Z def g(args): 2023-03-31T06:22:41.4734563Z return f(args) 2023-03-31T06:22:41.4734631Z return g 2023-03-31T06:22:41.4734726Z call = _wrap_func(module.call_86) 2023-03-31T06:22:41.4734732Z 2023-03-31T06:22:41.4734739Z 2023-03-31T06:22:41.4734833Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4734950Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4735076Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4735288Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4735514Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4735766Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4735974Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4736112Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4736117Z 2023-03-31T06:22:41.4736125Z 2023-03-31T06:22:41.4736206Z if __name__ == "__main__": 2023-03-31T06:22:41.4736282Z import argparse 2023-03-31T06:22:41.4736413Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4736419Z 2023-03-31T06:22:41.4736529Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4736824Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4737164Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4737443Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4737536Z args = parser.parse_args() 2023-03-31T06:22:41.4737541Z 2023-03-31T06:22:41.4737629Z if args.benchmark_kernels: 2023-03-31T06:22:41.4737819Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4737889Z else: 2023-03-31T06:22:41.4738002Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4738085Z benchmark_compiled_module() 2023-03-31T06:22:41.4738134Z 2023-03-31T06:22:41.4738191Z if p: 2023-03-31T06:22:41.4738342Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4738439Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4738575Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4738580Z 2023-03-31T06:22:41.4738863Z [2023-03-31 05:57:59,186] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 74 2023-03-31T06:22:41.4739134Z [2023-03-31 05:57:59,233] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 75 2023-03-31T06:22:41.4739543Z [2023-03-31 05:58:20,749] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/fm/cfmglh5vicmugkvxvvhqbkdzja3dpkxlws7txgicnp2tjq5mn2sl.py 2023-03-31T06:22:41.4739791Z [2023-03-31 05:58:20,749] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4739797Z 2023-03-31T06:22:41.4739884Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4739957Z import torch 2023-03-31T06:22:41.4740028Z import math 2023-03-31T06:22:41.4740100Z import random 2023-03-31T06:22:41.4740168Z import os 2023-03-31T06:22:41.4740244Z import tempfile 2023-03-31T06:22:41.4740361Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4740366Z 2023-03-31T06:22:41.4740474Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4740601Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4740733Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4740738Z 2023-03-31T06:22:41.4740818Z aten = torch.ops.aten 2023-03-31T06:22:41.4740957Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4741051Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4741059Z 2023-03-31T06:22:41.4741149Z async_compile.wait(globals()) 2023-03-31T06:22:41.4741224Z del async_compile 2023-03-31T06:22:41.4741335Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4741407Z wrapper = ( 2023-03-31T06:22:41.4741488Z ''' 2023-03-31T06:22:41.4741564Z #include 2023-03-31T06:22:41.4741640Z #include 2023-03-31T06:22:41.4741645Z 2023-03-31T06:22:41.4741737Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4741741Z 2023-03-31T06:22:41.4741833Z template 2023-03-31T06:22:41.4741973Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4742086Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4742211Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4742308Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4742525Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4742609Z return kernel_cpp; 2023-03-31T06:22:41.4742676Z } 2023-03-31T06:22:41.4742791Z std::vector call_87(std::vector args) { 2023-03-31T06:22:41.4742884Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4742961Z arg0_1 = args[0]; 2023-03-31T06:22:41.4743038Z arg1_1 = args[1]; 2023-03-31T06:22:41.4743111Z arg2_1 = args[2]; 2023-03-31T06:22:41.4743174Z 2023-03-31T06:22:41.4743273Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4743353Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4743442Z .findSchemaOrThrow( 2023-03-31T06:22:41.4743545Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4743615Z "") 2023-03-31T06:22:41.4743685Z .typed< 2023-03-31T06:22:41.4743761Z at::Tensor( 2023-03-31T06:22:41.4743860Z const at::Tensor& input_t, 2023-03-31T06:22:41.4743950Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4744066Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4744156Z c10::string_view attr, 2023-03-31T06:22:41.4744277Z torch::List> scalars, 2023-03-31T06:22:41.4744430Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4744496Z 2023-03-31T06:22:41.4744745Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.4744817Z arg1_1.reset(); 2023-03-31T06:22:41.4744876Z arg2_1.reset(); 2023-03-31T06:22:41.4744979Z return std::vector({buf0}); 2023-03-31T06:22:41.4745043Z } 2023-03-31T06:22:41.4745125Z ''' 2023-03-31T06:22:41.4745188Z ) 2023-03-31T06:22:41.4745194Z 2023-03-31T06:22:41.4745270Z module = load_inline( 2023-03-31T06:22:41.4745535Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.4745608Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4745726Z functions=['call_87'], 2023-03-31T06:22:41.4746108Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4746295Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4746997Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4747007Z 2023-03-31T06:22:41.4747084Z def _wrap_func(f): 2023-03-31T06:22:41.4747155Z def g(args): 2023-03-31T06:22:41.4747229Z return f(args) 2023-03-31T06:22:41.4747284Z return g 2023-03-31T06:22:41.4747377Z call = _wrap_func(module.call_87) 2023-03-31T06:22:41.4747383Z 2023-03-31T06:22:41.4747388Z 2023-03-31T06:22:41.4747480Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4747601Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4747728Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4747944Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4748155Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4748363Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4748478Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4748495Z 2023-03-31T06:22:41.4748530Z 2023-03-31T06:22:41.4748637Z if __name__ == "__main__": 2023-03-31T06:22:41.4748713Z import argparse 2023-03-31T06:22:41.4748843Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4748848Z 2023-03-31T06:22:41.4748959Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4749270Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4749614Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4749893Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4749986Z args = parser.parse_args() 2023-03-31T06:22:41.4749992Z 2023-03-31T06:22:41.4750070Z if args.benchmark_kernels: 2023-03-31T06:22:41.4750261Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4750333Z else: 2023-03-31T06:22:41.4750449Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4750543Z benchmark_compiled_module() 2023-03-31T06:22:41.4750548Z 2023-03-31T06:22:41.4750620Z if p: 2023-03-31T06:22:41.4750771Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4750866Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4750994Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4751000Z 2023-03-31T06:22:41.4751313Z [2023-03-31 05:58:20,749] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 75 2023-03-31T06:22:41.4751590Z [2023-03-31 05:58:20,826] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 76 2023-03-31T06:22:41.4751999Z [2023-03-31 05:58:42,437] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/4d/c4dpkli2majowrvmwzt6qlzo5ohxdixuoetggql5elozegsibs2x.py 2023-03-31T06:22:41.4752220Z [2023-03-31 05:58:42,437] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4752226Z 2023-03-31T06:22:41.4752322Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4752394Z import torch 2023-03-31T06:22:41.4752464Z import math 2023-03-31T06:22:41.4752524Z import random 2023-03-31T06:22:41.4752592Z import os 2023-03-31T06:22:41.4752695Z import tempfile 2023-03-31T06:22:41.4752811Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4752817Z 2023-03-31T06:22:41.4752936Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4753064Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4753198Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4753203Z 2023-03-31T06:22:41.4753285Z aten = torch.ops.aten 2023-03-31T06:22:41.4753411Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4753508Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4753516Z 2023-03-31T06:22:41.4753612Z async_compile.wait(globals()) 2023-03-31T06:22:41.4753688Z del async_compile 2023-03-31T06:22:41.4753811Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4753884Z wrapper = ( 2023-03-31T06:22:41.4753971Z ''' 2023-03-31T06:22:41.4754034Z #include 2023-03-31T06:22:41.4754111Z #include 2023-03-31T06:22:41.4754117Z 2023-03-31T06:22:41.4754213Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4754219Z 2023-03-31T06:22:41.4754313Z template 2023-03-31T06:22:41.4754441Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4754532Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4754659Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4754744Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4754871Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4754954Z return kernel_cpp; 2023-03-31T06:22:41.4755048Z } 2023-03-31T06:22:41.4755202Z std::vector call_88(std::vector args) { 2023-03-31T06:22:41.4755305Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4755380Z arg0_1 = args[0]; 2023-03-31T06:22:41.4755440Z arg1_1 = args[1]; 2023-03-31T06:22:41.4755513Z arg2_1 = args[2]; 2023-03-31T06:22:41.4755585Z arg3_1 = args[3]; 2023-03-31T06:22:41.4755649Z 2023-03-31T06:22:41.4755752Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4755846Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4755935Z .findSchemaOrThrow( 2023-03-31T06:22:41.4756024Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4756093Z "") 2023-03-31T06:22:41.4756162Z .typed< 2023-03-31T06:22:41.4756238Z at::Tensor( 2023-03-31T06:22:41.4756337Z const at::Tensor& input_t, 2023-03-31T06:22:41.4756440Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4756560Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4756638Z c10::string_view attr, 2023-03-31T06:22:41.4756761Z torch::List> scalars, 2023-03-31T06:22:41.4756882Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4756949Z 2023-03-31T06:22:41.4757181Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardswish", {-1}, ""); 2023-03-31T06:22:41.4757258Z arg1_1.reset(); 2023-03-31T06:22:41.4757331Z arg2_1.reset(); 2023-03-31T06:22:41.4757390Z arg3_1.reset(); 2023-03-31T06:22:41.4757527Z return std::vector({buf0}); 2023-03-31T06:22:41.4757593Z } 2023-03-31T06:22:41.4757676Z ''' 2023-03-31T06:22:41.4757739Z ) 2023-03-31T06:22:41.4757744Z 2023-03-31T06:22:41.4757824Z module = load_inline( 2023-03-31T06:22:41.4758100Z name='inline_extension_cudwzcjmrkzefzcijxyxj2xexarvj4uoxq3zkd6yjvof2g3iyxm6', 2023-03-31T06:22:41.4758187Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4758294Z functions=['call_88'], 2023-03-31T06:22:41.4758677Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4758834Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4759541Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4759577Z 2023-03-31T06:22:41.4764374Z def _wrap_func(f): 2023-03-31T06:22:41.4764479Z def g(args): 2023-03-31T06:22:41.4764557Z return f(args) 2023-03-31T06:22:41.4764628Z return g 2023-03-31T06:22:41.4764714Z call = _wrap_func(module.call_88) 2023-03-31T06:22:41.4764723Z 2023-03-31T06:22:41.4764748Z 2023-03-31T06:22:41.4764836Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4764960Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4765094Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4765341Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4765553Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4765771Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4765993Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4766118Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4766138Z 2023-03-31T06:22:41.4766144Z 2023-03-31T06:22:41.4766210Z if __name__ == "__main__": 2023-03-31T06:22:41.4766289Z import argparse 2023-03-31T06:22:41.4766542Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4766585Z 2023-03-31T06:22:41.4766700Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4767017Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4767358Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4767642Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4767740Z args = parser.parse_args() 2023-03-31T06:22:41.4767745Z 2023-03-31T06:22:41.4767823Z if args.benchmark_kernels: 2023-03-31T06:22:41.4768015Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4768084Z else: 2023-03-31T06:22:41.4768199Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4768296Z benchmark_compiled_module() 2023-03-31T06:22:41.4768306Z 2023-03-31T06:22:41.4768378Z if p: 2023-03-31T06:22:41.4768531Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4768633Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4768760Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4768766Z 2023-03-31T06:22:41.4769047Z [2023-03-31 05:58:42,437] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 76 2023-03-31T06:22:41.4769354Z [2023-03-31 05:58:42,487] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 77 2023-03-31T06:22:41.4769765Z [2023-03-31 05:59:03,991] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3j/c3jwfmf6vrhxjq7lf73sqclmabcvplq7fnl4zmbi2klslptekh6k.py 2023-03-31T06:22:41.4769987Z [2023-03-31 05:59:03,992] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4769994Z 2023-03-31T06:22:41.4770097Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4770173Z import torch 2023-03-31T06:22:41.4770245Z import math 2023-03-31T06:22:41.4770305Z import random 2023-03-31T06:22:41.4770375Z import os 2023-03-31T06:22:41.4770451Z import tempfile 2023-03-31T06:22:41.4770573Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4770579Z 2023-03-31T06:22:41.4770700Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4770874Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4771010Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4771016Z 2023-03-31T06:22:41.4771100Z aten = torch.ops.aten 2023-03-31T06:22:41.4771227Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4771324Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4771329Z 2023-03-31T06:22:41.4771422Z async_compile.wait(globals()) 2023-03-31T06:22:41.4771499Z del async_compile 2023-03-31T06:22:41.4771625Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4771701Z wrapper = ( 2023-03-31T06:22:41.4771788Z ''' 2023-03-31T06:22:41.4771853Z #include 2023-03-31T06:22:41.4771930Z #include 2023-03-31T06:22:41.4771935Z 2023-03-31T06:22:41.4772035Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4772041Z 2023-03-31T06:22:41.4772134Z template 2023-03-31T06:22:41.4772261Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4772351Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4772477Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4772564Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4772691Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4772771Z return kernel_cpp; 2023-03-31T06:22:41.4772837Z } 2023-03-31T06:22:41.4772964Z std::vector call_89(std::vector args) { 2023-03-31T06:22:41.4773058Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4773165Z arg0_1 = args[0]; 2023-03-31T06:22:41.4773249Z arg1_1 = args[1]; 2023-03-31T06:22:41.4773320Z arg2_1 = args[2]; 2023-03-31T06:22:41.4773383Z 2023-03-31T06:22:41.4773485Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4773580Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4773666Z .findSchemaOrThrow( 2023-03-31T06:22:41.4773765Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4773822Z "") 2023-03-31T06:22:41.4773891Z .typed< 2023-03-31T06:22:41.4773966Z at::Tensor( 2023-03-31T06:22:41.4774066Z const at::Tensor& input_t, 2023-03-31T06:22:41.4774166Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4774282Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4774371Z c10::string_view attr, 2023-03-31T06:22:41.4774479Z torch::List> scalars, 2023-03-31T06:22:41.4774603Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4774672Z 2023-03-31T06:22:41.4774919Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardswish", {-1}, ""); 2023-03-31T06:22:41.4774992Z arg1_1.reset(); 2023-03-31T06:22:41.4775063Z arg2_1.reset(); 2023-03-31T06:22:41.4775168Z return std::vector({buf0}); 2023-03-31T06:22:41.4775231Z } 2023-03-31T06:22:41.4775302Z ''' 2023-03-31T06:22:41.4775367Z ) 2023-03-31T06:22:41.4775372Z 2023-03-31T06:22:41.4775451Z module = load_inline( 2023-03-31T06:22:41.4775780Z name='inline_extension_cnvk2g3e2rdsnmm65pobm4q7xt5awnshzz37lydqy4gckkvhk5qf', 2023-03-31T06:22:41.4775868Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4775983Z functions=['call_89'], 2023-03-31T06:22:41.4776367Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4776509Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4777219Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4777289Z 2023-03-31T06:22:41.4777355Z def _wrap_func(f): 2023-03-31T06:22:41.4777427Z def g(args): 2023-03-31T06:22:41.4777504Z return f(args) 2023-03-31T06:22:41.4777572Z return g 2023-03-31T06:22:41.4777668Z call = _wrap_func(module.call_89) 2023-03-31T06:22:41.4777674Z 2023-03-31T06:22:41.4777679Z 2023-03-31T06:22:41.4777773Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4777891Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4778005Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4778223Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4778435Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4778651Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4778778Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4778784Z 2023-03-31T06:22:41.4778792Z 2023-03-31T06:22:41.4778870Z if __name__ == "__main__": 2023-03-31T06:22:41.4778946Z import argparse 2023-03-31T06:22:41.4779076Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4779084Z 2023-03-31T06:22:41.4779182Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4779491Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4779859Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4780161Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4780257Z args = parser.parse_args() 2023-03-31T06:22:41.4780261Z 2023-03-31T06:22:41.4780351Z if args.benchmark_kernels: 2023-03-31T06:22:41.4780540Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4780610Z else: 2023-03-31T06:22:41.4780709Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4780805Z benchmark_compiled_module() 2023-03-31T06:22:41.4780814Z 2023-03-31T06:22:41.4780884Z if p: 2023-03-31T06:22:41.4781033Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4781130Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4781266Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4781272Z 2023-03-31T06:22:41.4781557Z [2023-03-31 05:59:03,992] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 77 2023-03-31T06:22:41.4781827Z [2023-03-31 05:59:04,055] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 78 2023-03-31T06:22:41.4782217Z [2023-03-31 05:59:25,579] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ek/cekhknnmw7ym226mhff7a4h4lq4m3t43x5k6774b5u7aqodriz6t.py 2023-03-31T06:22:41.4782529Z [2023-03-31 05:59:25,579] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4782550Z 2023-03-31T06:22:41.4782683Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4782758Z import torch 2023-03-31T06:22:41.4782830Z import math 2023-03-31T06:22:41.4782904Z import random 2023-03-31T06:22:41.4782973Z import os 2023-03-31T06:22:41.4783048Z import tempfile 2023-03-31T06:22:41.4783153Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4783159Z 2023-03-31T06:22:41.4783282Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4783409Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4783545Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4783550Z 2023-03-31T06:22:41.4783635Z aten = torch.ops.aten 2023-03-31T06:22:41.4783781Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4783877Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4783912Z 2023-03-31T06:22:41.4784005Z async_compile.wait(globals()) 2023-03-31T06:22:41.4784067Z del async_compile 2023-03-31T06:22:41.4784190Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4784271Z wrapper = ( 2023-03-31T06:22:41.4784357Z ''' 2023-03-31T06:22:41.4784434Z #include 2023-03-31T06:22:41.4784513Z #include 2023-03-31T06:22:41.4784518Z 2023-03-31T06:22:41.4784612Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4784617Z 2023-03-31T06:22:41.4784696Z template 2023-03-31T06:22:41.4784823Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4784911Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4785038Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4785138Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4785262Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4785340Z return kernel_cpp; 2023-03-31T06:22:41.4785393Z } 2023-03-31T06:22:41.4785519Z std::vector call_90(std::vector args) { 2023-03-31T06:22:41.4785622Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4785699Z arg0_1 = args[0]; 2023-03-31T06:22:41.4785771Z arg1_1 = args[1]; 2023-03-31T06:22:41.4785842Z arg2_1 = args[2]; 2023-03-31T06:22:41.4785913Z arg3_1 = args[3]; 2023-03-31T06:22:41.4785963Z 2023-03-31T06:22:41.4786062Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4786153Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4786242Z .findSchemaOrThrow( 2023-03-31T06:22:41.4786370Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4786461Z "") 2023-03-31T06:22:41.4786528Z .typed< 2023-03-31T06:22:41.4786591Z at::Tensor( 2023-03-31T06:22:41.4786688Z const at::Tensor& input_t, 2023-03-31T06:22:41.4786790Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4786906Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4786997Z c10::string_view attr, 2023-03-31T06:22:41.4787117Z torch::List> scalars, 2023-03-31T06:22:41.4787240Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4787305Z 2023-03-31T06:22:41.4787524Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardswish", {-1}, ""); 2023-03-31T06:22:41.4787596Z arg1_1.reset(); 2023-03-31T06:22:41.4787667Z arg2_1.reset(); 2023-03-31T06:22:41.4787736Z arg3_1.reset(); 2023-03-31T06:22:41.4787844Z return std::vector({buf0}); 2023-03-31T06:22:41.4787908Z } 2023-03-31T06:22:41.4787988Z ''' 2023-03-31T06:22:41.4788037Z ) 2023-03-31T06:22:41.4788043Z 2023-03-31T06:22:41.4788120Z module = load_inline( 2023-03-31T06:22:41.4788398Z name='inline_extension_cudwzcjmrkzefzcijxyxj2xexarvj4uoxq3zkd6yjvof2g3iyxm6', 2023-03-31T06:22:41.4788485Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4788600Z functions=['call_90'], 2023-03-31T06:22:41.4789019Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4789174Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4789881Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4789891Z 2023-03-31T06:22:41.4789954Z def _wrap_func(f): 2023-03-31T06:22:41.4790024Z def g(args): 2023-03-31T06:22:41.4790098Z return f(args) 2023-03-31T06:22:41.4790167Z return g 2023-03-31T06:22:41.4790261Z call = _wrap_func(module.call_90) 2023-03-31T06:22:41.4790267Z 2023-03-31T06:22:41.4790300Z 2023-03-31T06:22:41.4790395Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4790514Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4790642Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4790843Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4791049Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4791258Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4791467Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4791603Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4791609Z 2023-03-31T06:22:41.4791614Z 2023-03-31T06:22:41.4791692Z if __name__ == "__main__": 2023-03-31T06:22:41.4791770Z import argparse 2023-03-31T06:22:41.4791900Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4791908Z 2023-03-31T06:22:41.4792007Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4792320Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4792658Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4792933Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4793054Z args = parser.parse_args() 2023-03-31T06:22:41.4793097Z 2023-03-31T06:22:41.4793188Z if args.benchmark_kernels: 2023-03-31T06:22:41.4793379Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4793447Z else: 2023-03-31T06:22:41.4793546Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4793642Z benchmark_compiled_module() 2023-03-31T06:22:41.4793647Z 2023-03-31T06:22:41.4793717Z if p: 2023-03-31T06:22:41.4793867Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4793966Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4794102Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4794108Z 2023-03-31T06:22:41.4794390Z [2023-03-31 05:59:25,580] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 78 2023-03-31T06:22:41.4794658Z [2023-03-31 05:59:25,639] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 79 2023-03-31T06:22:41.4795063Z [2023-03-31 05:59:47,233] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/mp/cmp2ci6vcnosqtw7377d7ox4kyd5arelmzlezdp7nbw4twofxmee.py 2023-03-31T06:22:41.4795272Z [2023-03-31 05:59:47,234] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4795292Z 2023-03-31T06:22:41.4795376Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4795451Z import torch 2023-03-31T06:22:41.4795522Z import math 2023-03-31T06:22:41.4795595Z import random 2023-03-31T06:22:41.4795664Z import os 2023-03-31T06:22:41.4795773Z import tempfile 2023-03-31T06:22:41.4795879Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4795885Z 2023-03-31T06:22:41.4796003Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4796126Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4796258Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4796263Z 2023-03-31T06:22:41.4796346Z aten = torch.ops.aten 2023-03-31T06:22:41.4796487Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4796582Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4796587Z 2023-03-31T06:22:41.4796677Z async_compile.wait(globals()) 2023-03-31T06:22:41.4796739Z del async_compile 2023-03-31T06:22:41.4796860Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4796960Z wrapper = ( 2023-03-31T06:22:41.4797043Z ''' 2023-03-31T06:22:41.4797119Z #include 2023-03-31T06:22:41.4797194Z #include 2023-03-31T06:22:41.4797199Z 2023-03-31T06:22:41.4797294Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4797300Z 2023-03-31T06:22:41.4797379Z template 2023-03-31T06:22:41.4797502Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4797588Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4797711Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4797810Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4797936Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4798014Z return kernel_cpp; 2023-03-31T06:22:41.4798063Z } 2023-03-31T06:22:41.4798191Z std::vector call_91(std::vector args) { 2023-03-31T06:22:41.4798283Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4798356Z arg0_1 = args[0]; 2023-03-31T06:22:41.4798435Z arg1_1 = args[1]; 2023-03-31T06:22:41.4798509Z arg2_1 = args[2]; 2023-03-31T06:22:41.4798574Z 2023-03-31T06:22:41.4798661Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4798753Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4798839Z .findSchemaOrThrow( 2023-03-31T06:22:41.4798938Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4799005Z "") 2023-03-31T06:22:41.4799072Z .typed< 2023-03-31T06:22:41.4799147Z at::Tensor( 2023-03-31T06:22:41.4799259Z const at::Tensor& input_t, 2023-03-31T06:22:41.4799398Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4799515Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4799604Z c10::string_view attr, 2023-03-31T06:22:41.4799725Z torch::List> scalars, 2023-03-31T06:22:41.4799845Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4799914Z 2023-03-31T06:22:41.4800160Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardswish", {-1}, ""); 2023-03-31T06:22:41.4800223Z arg1_1.reset(); 2023-03-31T06:22:41.4800295Z arg2_1.reset(); 2023-03-31T06:22:41.4800400Z return std::vector({buf0}); 2023-03-31T06:22:41.4800463Z } 2023-03-31T06:22:41.4800544Z ''' 2023-03-31T06:22:41.4800605Z ) 2023-03-31T06:22:41.4800611Z 2023-03-31T06:22:41.4800686Z module = load_inline( 2023-03-31T06:22:41.4800941Z name='inline_extension_cnvk2g3e2rdsnmm65pobm4q7xt5awnshzz37lydqy4gckkvhk5qf', 2023-03-31T06:22:41.4801029Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4801144Z functions=['call_91'], 2023-03-31T06:22:41.4801524Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4801678Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4802424Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4802432Z 2023-03-31T06:22:41.4802508Z def _wrap_func(f): 2023-03-31T06:22:41.4802579Z def g(args): 2023-03-31T06:22:41.4802652Z return f(args) 2023-03-31T06:22:41.4802707Z return g 2023-03-31T06:22:41.4802805Z call = _wrap_func(module.call_91) 2023-03-31T06:22:41.4802813Z 2023-03-31T06:22:41.4802818Z 2023-03-31T06:22:41.4802913Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4803220Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4803350Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4803567Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4803849Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4804045Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4804174Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4804179Z 2023-03-31T06:22:41.4804184Z 2023-03-31T06:22:41.4804263Z if __name__ == "__main__": 2023-03-31T06:22:41.4804342Z import argparse 2023-03-31T06:22:41.4804474Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4804482Z 2023-03-31T06:22:41.4804599Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4804915Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4805259Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4805539Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4805619Z args = parser.parse_args() 2023-03-31T06:22:41.4805641Z 2023-03-31T06:22:41.4805719Z if args.benchmark_kernels: 2023-03-31T06:22:41.4805910Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4805980Z else: 2023-03-31T06:22:41.4806094Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4806190Z benchmark_compiled_module() 2023-03-31T06:22:41.4806195Z 2023-03-31T06:22:41.4806304Z if p: 2023-03-31T06:22:41.4806490Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4806573Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4806711Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4806717Z 2023-03-31T06:22:41.4807000Z [2023-03-31 05:59:47,234] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 79 2023-03-31T06:22:41.4807269Z [2023-03-31 05:59:47,302] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 80 2023-03-31T06:22:41.4807685Z [2023-03-31 06:00:08,901] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zx/czxtcpt5midawaqegqkzbqun5linjtjjqerqjt2dnthdyuzowx2d.py 2023-03-31T06:22:41.4807904Z [2023-03-31 06:00:08,901] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4807911Z 2023-03-31T06:22:41.4808009Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4808087Z import torch 2023-03-31T06:22:41.4808162Z import math 2023-03-31T06:22:41.4808221Z import random 2023-03-31T06:22:41.4808291Z import os 2023-03-31T06:22:41.4808367Z import tempfile 2023-03-31T06:22:41.4808486Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4808492Z 2023-03-31T06:22:41.4808613Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4808743Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4808877Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4808883Z 2023-03-31T06:22:41.4808985Z aten = torch.ops.aten 2023-03-31T06:22:41.4809127Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4809226Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4809232Z 2023-03-31T06:22:41.4809326Z async_compile.wait(globals()) 2023-03-31T06:22:41.4809402Z del async_compile 2023-03-31T06:22:41.4809524Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4809602Z wrapper = ( 2023-03-31T06:22:41.4809675Z ''' 2023-03-31T06:22:41.4809752Z #include 2023-03-31T06:22:41.4809830Z #include 2023-03-31T06:22:41.4809835Z 2023-03-31T06:22:41.4809929Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4809934Z 2023-03-31T06:22:41.4810029Z template 2023-03-31T06:22:41.4810154Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4810280Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4810393Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4810492Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4810616Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4810697Z return kernel_cpp; 2023-03-31T06:22:41.4810763Z } 2023-03-31T06:22:41.4810890Z std::vector call_93(std::vector args) { 2023-03-31T06:22:41.4810996Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4811074Z arg0_1 = args[0]; 2023-03-31T06:22:41.4811136Z arg1_1 = args[1]; 2023-03-31T06:22:41.4811208Z arg2_1 = args[2]; 2023-03-31T06:22:41.4811279Z arg3_1 = args[3]; 2023-03-31T06:22:41.4811341Z 2023-03-31T06:22:41.4811440Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4811533Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4811608Z .findSchemaOrThrow( 2023-03-31T06:22:41.4811706Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4811775Z "") 2023-03-31T06:22:41.4811842Z .typed< 2023-03-31T06:22:41.4811916Z at::Tensor( 2023-03-31T06:22:41.4812017Z const at::Tensor& input_t, 2023-03-31T06:22:41.4812117Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4812222Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4812311Z c10::string_view attr, 2023-03-31T06:22:41.4812430Z torch::List> scalars, 2023-03-31T06:22:41.4812577Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4812669Z 2023-03-31T06:22:41.4812831Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4812904Z arg1_1.reset(); 2023-03-31T06:22:41.4812974Z arg2_1.reset(); 2023-03-31T06:22:41.4813030Z arg3_1.reset(); 2023-03-31T06:22:41.4813135Z return std::vector({buf0}); 2023-03-31T06:22:41.4813200Z } 2023-03-31T06:22:41.4813283Z ''' 2023-03-31T06:22:41.4813345Z ) 2023-03-31T06:22:41.4813350Z 2023-03-31T06:22:41.4813429Z module = load_inline( 2023-03-31T06:22:41.4813698Z name='inline_extension_c2hs5opcafpw4goqfmhdn4sfm5mn5hcfq3zgnku6z7hzzbefhon2', 2023-03-31T06:22:41.4813772Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4813888Z functions=['call_93'], 2023-03-31T06:22:41.4814274Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4814431Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4815136Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4815146Z 2023-03-31T06:22:41.4815222Z def _wrap_func(f): 2023-03-31T06:22:41.4815293Z def g(args): 2023-03-31T06:22:41.4815400Z return f(args) 2023-03-31T06:22:41.4815456Z return g 2023-03-31T06:22:41.4815550Z call = _wrap_func(module.call_93) 2023-03-31T06:22:41.4815555Z 2023-03-31T06:22:41.4815560Z 2023-03-31T06:22:41.4815659Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4815776Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4815904Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4816120Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4816324Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4816534Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4816747Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4816898Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4816903Z 2023-03-31T06:22:41.4816910Z 2023-03-31T06:22:41.4816989Z if __name__ == "__main__": 2023-03-31T06:22:41.4817064Z import argparse 2023-03-31T06:22:41.4817193Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4817199Z 2023-03-31T06:22:41.4817308Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4817618Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4817955Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4818230Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4818311Z args = parser.parse_args() 2023-03-31T06:22:41.4818330Z 2023-03-31T06:22:41.4818407Z if args.benchmark_kernels: 2023-03-31T06:22:41.4818596Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4818665Z else: 2023-03-31T06:22:41.4818778Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4818873Z benchmark_compiled_module() 2023-03-31T06:22:41.4818879Z 2023-03-31T06:22:41.4818948Z if p: 2023-03-31T06:22:41.4819100Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4819212Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4819383Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4819389Z 2023-03-31T06:22:41.4819669Z [2023-03-31 06:00:08,901] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 80 2023-03-31T06:22:41.4819935Z [2023-03-31 06:00:08,951] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 81 2023-03-31T06:22:41.4820345Z [2023-03-31 06:00:30,489] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/io/ciosxiweumih3aq6jjpucxtkz32mpoelbkxd3iqg65b4q3h5wgh6.py 2023-03-31T06:22:41.4820565Z [2023-03-31 06:00:30,489] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4820571Z 2023-03-31T06:22:41.4820669Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4820741Z import torch 2023-03-31T06:22:41.4820811Z import math 2023-03-31T06:22:41.4820870Z import random 2023-03-31T06:22:41.4820938Z import os 2023-03-31T06:22:41.4821014Z import tempfile 2023-03-31T06:22:41.4821135Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4821140Z 2023-03-31T06:22:41.4821259Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4821384Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4821517Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4821522Z 2023-03-31T06:22:41.4821593Z aten = torch.ops.aten 2023-03-31T06:22:41.4821733Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4821827Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4821860Z 2023-03-31T06:22:41.4821952Z async_compile.wait(globals()) 2023-03-31T06:22:41.4822027Z del async_compile 2023-03-31T06:22:41.4822148Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4822221Z wrapper = ( 2023-03-31T06:22:41.4822290Z ''' 2023-03-31T06:22:41.4822448Z #include 2023-03-31T06:22:41.4822540Z #include 2023-03-31T06:22:41.4822548Z 2023-03-31T06:22:41.4822645Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4822650Z 2023-03-31T06:22:41.4822745Z template 2023-03-31T06:22:41.4822868Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4822956Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4823081Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4823166Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4823342Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4823424Z return kernel_cpp; 2023-03-31T06:22:41.4823491Z } 2023-03-31T06:22:41.4823621Z std::vector call_95(std::vector args) { 2023-03-31T06:22:41.4823716Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4823790Z arg0_1 = args[0]; 2023-03-31T06:22:41.4823850Z arg1_1 = args[1]; 2023-03-31T06:22:41.4823922Z arg2_1 = args[2]; 2023-03-31T06:22:41.4823986Z 2023-03-31T06:22:41.4824086Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4824180Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4824266Z .findSchemaOrThrow( 2023-03-31T06:22:41.4824353Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4824422Z "") 2023-03-31T06:22:41.4824491Z .typed< 2023-03-31T06:22:41.4824568Z at::Tensor( 2023-03-31T06:22:41.4824668Z const at::Tensor& input_t, 2023-03-31T06:22:41.4824771Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4824892Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4824984Z c10::string_view attr, 2023-03-31T06:22:41.4825092Z torch::List> scalars, 2023-03-31T06:22:41.4825212Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4825278Z 2023-03-31T06:22:41.4825451Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4825553Z arg1_1.reset(); 2023-03-31T06:22:41.4825689Z arg2_1.reset(); 2023-03-31T06:22:41.4825795Z return std::vector({buf0}); 2023-03-31T06:22:41.4825846Z } 2023-03-31T06:22:41.4825934Z ''' 2023-03-31T06:22:41.4825996Z ) 2023-03-31T06:22:41.4826006Z 2023-03-31T06:22:41.4826087Z module = load_inline( 2023-03-31T06:22:41.4826353Z name='inline_extension_cabai5zl24epsf4tb5et443eooog5gqm5vfqz6torsz4rh3eskkg', 2023-03-31T06:22:41.4826443Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4826559Z functions=['call_95'], 2023-03-31T06:22:41.4826935Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4827093Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4827810Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4827818Z 2023-03-31T06:22:41.4827895Z def _wrap_func(f): 2023-03-31T06:22:41.4827966Z def g(args): 2023-03-31T06:22:41.4828041Z return f(args) 2023-03-31T06:22:41.4828109Z return g 2023-03-31T06:22:41.4828206Z call = _wrap_func(module.call_95) 2023-03-31T06:22:41.4828212Z 2023-03-31T06:22:41.4828217Z 2023-03-31T06:22:41.4828311Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4828451Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4828581Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4828795Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4829006Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4829227Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4829356Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4829361Z 2023-03-31T06:22:41.4829366Z 2023-03-31T06:22:41.4829446Z if __name__ == "__main__": 2023-03-31T06:22:41.4829521Z import argparse 2023-03-31T06:22:41.4829640Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4829683Z 2023-03-31T06:22:41.4829782Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4830092Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4830429Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4830707Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4830802Z args = parser.parse_args() 2023-03-31T06:22:41.4830810Z 2023-03-31T06:22:41.4830899Z if args.benchmark_kernels: 2023-03-31T06:22:41.4831088Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4831153Z else: 2023-03-31T06:22:41.4831253Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4831348Z benchmark_compiled_module() 2023-03-31T06:22:41.4831356Z 2023-03-31T06:22:41.4831424Z if p: 2023-03-31T06:22:41.4831574Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4831672Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4831811Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4831817Z 2023-03-31T06:22:41.4832096Z [2023-03-31 06:00:30,489] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 81 2023-03-31T06:22:41.4832386Z [2023-03-31 06:00:30,554] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 82 2023-03-31T06:22:41.4832818Z [2023-03-31 06:00:51,939] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/xh/cxhgiecv2umpkv7lx62dspfedooobnpi77y2jsj5tig2vlfcz664.py 2023-03-31T06:22:41.4833037Z [2023-03-31 06:00:51,940] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4833043Z 2023-03-31T06:22:41.4833139Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4833214Z import torch 2023-03-31T06:22:41.4833284Z import math 2023-03-31T06:22:41.4833355Z import random 2023-03-31T06:22:41.4833423Z import os 2023-03-31T06:22:41.4833499Z import tempfile 2023-03-31T06:22:41.4833605Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4833611Z 2023-03-31T06:22:41.4833731Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4833855Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4833987Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4833995Z 2023-03-31T06:22:41.4834077Z aten = torch.ops.aten 2023-03-31T06:22:41.4834215Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4834309Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4834315Z 2023-03-31T06:22:41.4834406Z async_compile.wait(globals()) 2023-03-31T06:22:41.4834468Z del async_compile 2023-03-31T06:22:41.4834589Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4834663Z wrapper = ( 2023-03-31T06:22:41.4834743Z ''' 2023-03-31T06:22:41.4834819Z #include 2023-03-31T06:22:41.4834894Z #include 2023-03-31T06:22:41.4834926Z 2023-03-31T06:22:41.4835019Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4835024Z 2023-03-31T06:22:41.4835104Z template 2023-03-31T06:22:41.4835226Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4835312Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4835436Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4835534Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4835661Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4835740Z return kernel_cpp; 2023-03-31T06:22:41.4835789Z } 2023-03-31T06:22:41.4835917Z std::vector call_97(std::vector args) { 2023-03-31T06:22:41.4836021Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4836123Z arg0_1 = args[0]; 2023-03-31T06:22:41.4836197Z arg1_1 = args[1]; 2023-03-31T06:22:41.4836268Z arg2_1 = args[2]; 2023-03-31T06:22:41.4836340Z arg3_1 = args[3]; 2023-03-31T06:22:41.4836391Z 2023-03-31T06:22:41.4836493Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4836586Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4836674Z .findSchemaOrThrow( 2023-03-31T06:22:41.4836770Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4836837Z "") 2023-03-31T06:22:41.4836906Z .typed< 2023-03-31T06:22:41.4836970Z at::Tensor( 2023-03-31T06:22:41.4837072Z const at::Tensor& input_t, 2023-03-31T06:22:41.4837171Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4837289Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4837379Z c10::string_view attr, 2023-03-31T06:22:41.4837498Z torch::List> scalars, 2023-03-31T06:22:41.4837621Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4837674Z 2023-03-31T06:22:41.4837838Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4837911Z arg1_1.reset(); 2023-03-31T06:22:41.4837983Z arg2_1.reset(); 2023-03-31T06:22:41.4838056Z arg3_1.reset(); 2023-03-31T06:22:41.4838159Z return std::vector({buf0}); 2023-03-31T06:22:41.4838221Z } 2023-03-31T06:22:41.4838290Z ''' 2023-03-31T06:22:41.4838352Z ) 2023-03-31T06:22:41.4838357Z 2023-03-31T06:22:41.4838461Z module = load_inline( 2023-03-31T06:22:41.4838759Z name='inline_extension_c2hs5opcafpw4goqfmhdn4sfm5mn5hcfq3zgnku6z7hzzbefhon2', 2023-03-31T06:22:41.4838846Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4838960Z functions=['call_97'], 2023-03-31T06:22:41.4839339Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4839495Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4840197Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4840216Z 2023-03-31T06:22:41.4840280Z def _wrap_func(f): 2023-03-31T06:22:41.4840357Z def g(args): 2023-03-31T06:22:41.4840434Z return f(args) 2023-03-31T06:22:41.4840502Z return g 2023-03-31T06:22:41.4840597Z call = _wrap_func(module.call_97) 2023-03-31T06:22:41.4840603Z 2023-03-31T06:22:41.4840608Z 2023-03-31T06:22:41.4840702Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4840820Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4840936Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4841149Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4841389Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4841601Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4841810Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4841946Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4841955Z 2023-03-31T06:22:41.4841961Z 2023-03-31T06:22:41.4842043Z if __name__ == "__main__": 2023-03-31T06:22:41.4842121Z import argparse 2023-03-31T06:22:41.4842254Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4842260Z 2023-03-31T06:22:41.4842357Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4842670Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4843242Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4843523Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4843616Z args = parser.parse_args() 2023-03-31T06:22:41.4843621Z 2023-03-31T06:22:41.4843713Z if args.benchmark_kernels: 2023-03-31T06:22:41.4843907Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4843980Z else: 2023-03-31T06:22:41.4844081Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4844177Z benchmark_compiled_module() 2023-03-31T06:22:41.4844182Z 2023-03-31T06:22:41.4844253Z if p: 2023-03-31T06:22:41.4844408Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4844510Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4844649Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4844655Z 2023-03-31T06:22:41.4844944Z [2023-03-31 06:00:51,940] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 82 2023-03-31T06:22:41.4845216Z [2023-03-31 06:00:51,987] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 83 2023-03-31T06:22:41.4845668Z [2023-03-31 06:01:13,424] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/sk/csk667eqrg4cwry2lv5jfziaohrfwraxpex56epmjru5evyeoilq.py 2023-03-31T06:22:41.4845949Z [2023-03-31 06:01:13,425] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4845955Z 2023-03-31T06:22:41.4846055Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4846129Z import torch 2023-03-31T06:22:41.4846202Z import math 2023-03-31T06:22:41.4846276Z import random 2023-03-31T06:22:41.4846345Z import os 2023-03-31T06:22:41.4846424Z import tempfile 2023-03-31T06:22:41.4846528Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4846534Z 2023-03-31T06:22:41.4846656Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4846783Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4846918Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4846924Z 2023-03-31T06:22:41.4847005Z aten = torch.ops.aten 2023-03-31T06:22:41.4847145Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4847245Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4847251Z 2023-03-31T06:22:41.4847349Z async_compile.wait(globals()) 2023-03-31T06:22:41.4847411Z del async_compile 2023-03-31T06:22:41.4847534Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4847609Z wrapper = ( 2023-03-31T06:22:41.4847691Z ''' 2023-03-31T06:22:41.4847769Z #include 2023-03-31T06:22:41.4847846Z #include 2023-03-31T06:22:41.4847854Z 2023-03-31T06:22:41.4847949Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4847954Z 2023-03-31T06:22:41.4848035Z template 2023-03-31T06:22:41.4848195Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4848285Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4848409Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4848508Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4848632Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4848715Z return kernel_cpp; 2023-03-31T06:22:41.4848768Z } 2023-03-31T06:22:41.4848899Z std::vector call_99(std::vector args) { 2023-03-31T06:22:41.4848995Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4849071Z arg0_1 = args[0]; 2023-03-31T06:22:41.4849146Z arg1_1 = args[1]; 2023-03-31T06:22:41.4849216Z arg2_1 = args[2]; 2023-03-31T06:22:41.4849280Z 2023-03-31T06:22:41.4849404Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4849497Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4849586Z .findSchemaOrThrow( 2023-03-31T06:22:41.4849688Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4849754Z "") 2023-03-31T06:22:41.4849821Z .typed< 2023-03-31T06:22:41.4849896Z at::Tensor( 2023-03-31T06:22:41.4849982Z const at::Tensor& input_t, 2023-03-31T06:22:41.4850081Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4850197Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4850286Z c10::string_view attr, 2023-03-31T06:22:41.4850408Z torch::List> scalars, 2023-03-31T06:22:41.4850530Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4850598Z 2023-03-31T06:22:41.4850768Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "leaky_relu", {0.100000000000000, }, ""); 2023-03-31T06:22:41.4850831Z arg1_1.reset(); 2023-03-31T06:22:41.4850903Z arg2_1.reset(); 2023-03-31T06:22:41.4851006Z return std::vector({buf0}); 2023-03-31T06:22:41.4851071Z } 2023-03-31T06:22:41.4851155Z ''' 2023-03-31T06:22:41.4851221Z ) 2023-03-31T06:22:41.4851227Z 2023-03-31T06:22:41.4851304Z module = load_inline( 2023-03-31T06:22:41.4851555Z name='inline_extension_cabai5zl24epsf4tb5et443eooog5gqm5vfqz6torsz4rh3eskkg', 2023-03-31T06:22:41.4851645Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4851760Z functions=['call_99'], 2023-03-31T06:22:41.4852169Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4852352Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4853060Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4853070Z 2023-03-31T06:22:41.4853146Z def _wrap_func(f): 2023-03-31T06:22:41.4853217Z def g(args): 2023-03-31T06:22:41.4853278Z return f(args) 2023-03-31T06:22:41.4853345Z return g 2023-03-31T06:22:41.4853438Z call = _wrap_func(module.call_99) 2023-03-31T06:22:41.4853444Z 2023-03-31T06:22:41.4853449Z 2023-03-31T06:22:41.4853548Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4853668Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4853795Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4854009Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4854220Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4854418Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4854543Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4854582Z 2023-03-31T06:22:41.4854588Z 2023-03-31T06:22:41.4854666Z if __name__ == "__main__": 2023-03-31T06:22:41.4854742Z import argparse 2023-03-31T06:22:41.4854872Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4854878Z 2023-03-31T06:22:41.4854988Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4855299Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4855639Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4855916Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4856027Z args = parser.parse_args() 2023-03-31T06:22:41.4856033Z 2023-03-31T06:22:41.4856121Z if args.benchmark_kernels: 2023-03-31T06:22:41.4856312Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4856379Z else: 2023-03-31T06:22:41.4856491Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4856585Z benchmark_compiled_module() 2023-03-31T06:22:41.4856591Z 2023-03-31T06:22:41.4856658Z if p: 2023-03-31T06:22:41.4856807Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4856893Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4857033Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4857038Z 2023-03-31T06:22:41.4857318Z [2023-03-31 06:01:13,425] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 83 2023-03-31T06:22:41.4857585Z [2023-03-31 06:01:13,481] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 84 2023-03-31T06:22:41.4858002Z [2023-03-31 06:01:34,621] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/rn/crncpnvwdrfjtjpp2qqrdx5o3z5yekoc2jrmsjk6oyynl57vioc6.py 2023-03-31T06:22:41.4858220Z [2023-03-31 06:01:34,621] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4858227Z 2023-03-31T06:22:41.4858323Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4858394Z import torch 2023-03-31T06:22:41.4858453Z import math 2023-03-31T06:22:41.4858524Z import random 2023-03-31T06:22:41.4858627Z import os 2023-03-31T06:22:41.4858728Z import tempfile 2023-03-31T06:22:41.4858845Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4858851Z 2023-03-31T06:22:41.4858970Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4859095Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4859216Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4859235Z 2023-03-31T06:22:41.4859304Z aten = torch.ops.aten 2023-03-31T06:22:41.4859445Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4859542Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4859548Z 2023-03-31T06:22:41.4859639Z async_compile.wait(globals()) 2023-03-31T06:22:41.4859713Z del async_compile 2023-03-31T06:22:41.4859834Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4859908Z wrapper = ( 2023-03-31T06:22:41.4859979Z ''' 2023-03-31T06:22:41.4860054Z #include 2023-03-31T06:22:41.4860131Z #include 2023-03-31T06:22:41.4860138Z 2023-03-31T06:22:41.4860232Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4860237Z 2023-03-31T06:22:41.4860329Z template 2023-03-31T06:22:41.4860452Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4860538Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4860650Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4860750Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4860874Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4860981Z return kernel_cpp; 2023-03-31T06:22:41.4861045Z } 2023-03-31T06:22:41.4861175Z std::vector call_101(std::vector args) { 2023-03-31T06:22:41.4861278Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4861339Z arg0_1 = args[0]; 2023-03-31T06:22:41.4861411Z arg1_1 = args[1]; 2023-03-31T06:22:41.4861483Z arg2_1 = args[2]; 2023-03-31T06:22:41.4861557Z arg3_1 = args[3]; 2023-03-31T06:22:41.4861620Z 2023-03-31T06:22:41.4861719Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4861812Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4861887Z .findSchemaOrThrow( 2023-03-31T06:22:41.4861985Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4862052Z "") 2023-03-31T06:22:41.4862120Z .typed< 2023-03-31T06:22:41.4862226Z at::Tensor( 2023-03-31T06:22:41.4862326Z const at::Tensor& input_t, 2023-03-31T06:22:41.4862526Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4862636Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4862726Z c10::string_view attr, 2023-03-31T06:22:41.4862849Z torch::List> scalars, 2023-03-31T06:22:41.4862971Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4863038Z 2023-03-31T06:22:41.4863301Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4863377Z arg1_1.reset(); 2023-03-31T06:22:41.4863449Z arg2_1.reset(); 2023-03-31T06:22:41.4863505Z arg3_1.reset(); 2023-03-31T06:22:41.4863612Z return std::vector({buf0}); 2023-03-31T06:22:41.4863677Z } 2023-03-31T06:22:41.4863759Z ''' 2023-03-31T06:22:41.4863821Z ) 2023-03-31T06:22:41.4863829Z 2023-03-31T06:22:41.4863908Z module = load_inline( 2023-03-31T06:22:41.4864180Z name='inline_extension_cow6quriqr7bswivrijfisjdrcn3dcra3pq6fb32ucqf4e4yfsyt', 2023-03-31T06:22:41.4864256Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4864373Z functions=['call_101'], 2023-03-31T06:22:41.4864757Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4864912Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4865668Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4865701Z 2023-03-31T06:22:41.4865778Z def _wrap_func(f): 2023-03-31T06:22:41.4865853Z def g(args): 2023-03-31T06:22:41.4865927Z return f(args) 2023-03-31T06:22:41.4865983Z return g 2023-03-31T06:22:41.4866080Z call = _wrap_func(module.call_101) 2023-03-31T06:22:41.4866086Z 2023-03-31T06:22:41.4866091Z 2023-03-31T06:22:41.4866186Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4866304Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4866431Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4866646Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4866856Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4867067Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4867269Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4867405Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4867414Z 2023-03-31T06:22:41.4867418Z 2023-03-31T06:22:41.4867496Z if __name__ == "__main__": 2023-03-31T06:22:41.4867575Z import argparse 2023-03-31T06:22:41.4867745Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4867752Z 2023-03-31T06:22:41.4867863Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4868171Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4868511Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4868789Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4868869Z args = parser.parse_args() 2023-03-31T06:22:41.4868874Z 2023-03-31T06:22:41.4868963Z if args.benchmark_kernels: 2023-03-31T06:22:41.4869154Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4869263Z else: 2023-03-31T06:22:41.4869378Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4869478Z benchmark_compiled_module() 2023-03-31T06:22:41.4869484Z 2023-03-31T06:22:41.4869556Z if p: 2023-03-31T06:22:41.4869707Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4869790Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4869929Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4869937Z 2023-03-31T06:22:41.4870220Z [2023-03-31 06:01:34,621] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 84 2023-03-31T06:22:41.4870486Z [2023-03-31 06:01:34,665] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 85 2023-03-31T06:22:41.4870886Z [2023-03-31 06:01:55,876] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/23/c237je7zvbskuywwhcijqdil4zqly6a7tfep4e5g6pgueeek7pwp.py 2023-03-31T06:22:41.4871106Z [2023-03-31 06:01:55,876] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4871115Z 2023-03-31T06:22:41.4871213Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4871287Z import torch 2023-03-31T06:22:41.4871346Z import math 2023-03-31T06:22:41.4871419Z import random 2023-03-31T06:22:41.4871487Z import os 2023-03-31T06:22:41.4871561Z import tempfile 2023-03-31T06:22:41.4871680Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4871685Z 2023-03-31T06:22:41.4871841Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4871995Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4872116Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4872133Z 2023-03-31T06:22:41.4872201Z aten = torch.ops.aten 2023-03-31T06:22:41.4872340Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4872438Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4872443Z 2023-03-31T06:22:41.4872535Z async_compile.wait(globals()) 2023-03-31T06:22:41.4872611Z del async_compile 2023-03-31T06:22:41.4872735Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4872809Z wrapper = ( 2023-03-31T06:22:41.4872878Z ''' 2023-03-31T06:22:41.4872954Z #include 2023-03-31T06:22:41.4873030Z #include 2023-03-31T06:22:41.4873035Z 2023-03-31T06:22:41.4873128Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4873134Z 2023-03-31T06:22:41.4873228Z template 2023-03-31T06:22:41.4873353Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4873438Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4873551Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4873647Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4873771Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4873851Z return kernel_cpp; 2023-03-31T06:22:41.4873914Z } 2023-03-31T06:22:41.4874044Z std::vector call_103(std::vector args) { 2023-03-31T06:22:41.4874168Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4874230Z arg0_1 = args[0]; 2023-03-31T06:22:41.4874300Z arg1_1 = args[1]; 2023-03-31T06:22:41.4874370Z arg2_1 = args[2]; 2023-03-31T06:22:41.4874432Z 2023-03-31T06:22:41.4874531Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4874624Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4874711Z .findSchemaOrThrow( 2023-03-31T06:22:41.4874801Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4874871Z "") 2023-03-31T06:22:41.4874938Z .typed< 2023-03-31T06:22:41.4875012Z at::Tensor( 2023-03-31T06:22:41.4875110Z const at::Tensor& input_t, 2023-03-31T06:22:41.4875212Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4875329Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4875433Z c10::string_view attr, 2023-03-31T06:22:41.4875554Z torch::List> scalars, 2023-03-31T06:22:41.4875678Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4875744Z 2023-03-31T06:22:41.4876015Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4876088Z arg1_1.reset(); 2023-03-31T06:22:41.4876159Z arg2_1.reset(); 2023-03-31T06:22:41.4876265Z return std::vector({buf0}); 2023-03-31T06:22:41.4876317Z } 2023-03-31T06:22:41.4876399Z ''' 2023-03-31T06:22:41.4876460Z ) 2023-03-31T06:22:41.4876466Z 2023-03-31T06:22:41.4876543Z module = load_inline( 2023-03-31T06:22:41.4876814Z name='inline_extension_cgzzjygjdtyaeq5ee6eazo4mwcm6w5k4l5xxnkhipg3d5ktdbcnm', 2023-03-31T06:22:41.4876899Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4877016Z functions=['call_103'], 2023-03-31T06:22:41.4877387Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4877544Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4878331Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4878362Z 2023-03-31T06:22:41.4878439Z def _wrap_func(f): 2023-03-31T06:22:41.4878513Z def g(args): 2023-03-31T06:22:41.4878588Z return f(args) 2023-03-31T06:22:41.4878657Z return g 2023-03-31T06:22:41.4878752Z call = _wrap_func(module.call_103) 2023-03-31T06:22:41.4878759Z 2023-03-31T06:22:41.4878763Z 2023-03-31T06:22:41.4878864Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4878970Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4879105Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4879321Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4879534Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4879754Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4879886Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4879893Z 2023-03-31T06:22:41.4879897Z 2023-03-31T06:22:41.4879976Z if __name__ == "__main__": 2023-03-31T06:22:41.4880054Z import argparse 2023-03-31T06:22:41.4880173Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4880193Z 2023-03-31T06:22:41.4880292Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4880603Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4880971Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4881255Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4881350Z args = parser.parse_args() 2023-03-31T06:22:41.4881355Z 2023-03-31T06:22:41.4881446Z if args.benchmark_kernels: 2023-03-31T06:22:41.4881641Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4881711Z else: 2023-03-31T06:22:41.4881814Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4881911Z benchmark_compiled_module() 2023-03-31T06:22:41.4881916Z 2023-03-31T06:22:41.4881986Z if p: 2023-03-31T06:22:41.4882138Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4882264Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4882406Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4882415Z 2023-03-31T06:22:41.4882697Z [2023-03-31 06:01:55,876] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 85 2023-03-31T06:22:41.4882963Z [2023-03-31 06:01:55,931] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 86 2023-03-31T06:22:41.4883538Z [2023-03-31 06:02:17,181] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ut/cutwjszpjy7zvt45mfxsoohxgjx5zsuw5oapslxzzfp3fswq2lzy.py 2023-03-31T06:22:41.4883763Z [2023-03-31 06:02:17,182] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4883768Z 2023-03-31T06:22:41.4883866Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4883940Z import torch 2023-03-31T06:22:41.4884016Z import math 2023-03-31T06:22:41.4884090Z import random 2023-03-31T06:22:41.4884162Z import os 2023-03-31T06:22:41.4884224Z import tempfile 2023-03-31T06:22:41.4884344Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4884350Z 2023-03-31T06:22:41.4884473Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4884600Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4884734Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4884740Z 2023-03-31T06:22:41.4884822Z aten = torch.ops.aten 2023-03-31T06:22:41.4885030Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4885161Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4885165Z 2023-03-31T06:22:41.4885244Z async_compile.wait(globals()) 2023-03-31T06:22:41.4885320Z del async_compile 2023-03-31T06:22:41.4885442Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4885516Z wrapper = ( 2023-03-31T06:22:41.4885601Z ''' 2023-03-31T06:22:41.4885678Z #include 2023-03-31T06:22:41.4885759Z #include 2023-03-31T06:22:41.4885764Z 2023-03-31T06:22:41.4885843Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4885866Z 2023-03-31T06:22:41.4885947Z template 2023-03-31T06:22:41.4886073Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4886160Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4886290Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4886389Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4886520Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4886602Z return kernel_cpp; 2023-03-31T06:22:41.4886653Z } 2023-03-31T06:22:41.4886782Z std::vector call_105(std::vector args) { 2023-03-31T06:22:41.4886890Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4886963Z arg0_1 = args[0]; 2023-03-31T06:22:41.4887040Z arg1_1 = args[1]; 2023-03-31T06:22:41.4887110Z arg2_1 = args[2]; 2023-03-31T06:22:41.4887171Z arg3_1 = args[3]; 2023-03-31T06:22:41.4887232Z 2023-03-31T06:22:41.4887330Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4887463Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4887552Z .findSchemaOrThrow( 2023-03-31T06:22:41.4887652Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4887718Z "") 2023-03-31T06:22:41.4887774Z .typed< 2023-03-31T06:22:41.4887848Z at::Tensor( 2023-03-31T06:22:41.4887945Z const at::Tensor& input_t, 2023-03-31T06:22:41.4888048Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4888167Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4888255Z c10::string_view attr, 2023-03-31T06:22:41.4888375Z torch::List> scalars, 2023-03-31T06:22:41.4888495Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4888548Z 2023-03-31T06:22:41.4888846Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4888921Z arg1_1.reset(); 2023-03-31T06:22:41.4888995Z arg2_1.reset(); 2023-03-31T06:22:41.4889065Z arg3_1.reset(); 2023-03-31T06:22:41.4889170Z return std::vector({buf0}); 2023-03-31T06:22:41.4889233Z } 2023-03-31T06:22:41.4889300Z ''' 2023-03-31T06:22:41.4889361Z ) 2023-03-31T06:22:41.4889366Z 2023-03-31T06:22:41.4889444Z module = load_inline( 2023-03-31T06:22:41.4889719Z name='inline_extension_cow6quriqr7bswivrijfisjdrcn3dcra3pq6fb32ucqf4e4yfsyt', 2023-03-31T06:22:41.4889808Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4889926Z functions=['call_105'], 2023-03-31T06:22:41.4890305Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4890449Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4891166Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4891186Z 2023-03-31T06:22:41.4891249Z def _wrap_func(f): 2023-03-31T06:22:41.4891320Z def g(args): 2023-03-31T06:22:41.4891399Z return f(args) 2023-03-31T06:22:41.4891468Z return g 2023-03-31T06:22:41.4891592Z call = _wrap_func(module.call_105) 2023-03-31T06:22:41.4891636Z 2023-03-31T06:22:41.4891641Z 2023-03-31T06:22:41.4891737Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4891855Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4891970Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4892185Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4892396Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4892606Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4892812Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4892946Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4892952Z 2023-03-31T06:22:41.4892957Z 2023-03-31T06:22:41.4893035Z if __name__ == "__main__": 2023-03-31T06:22:41.4893115Z import argparse 2023-03-31T06:22:41.4893234Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4893255Z 2023-03-31T06:22:41.4893352Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4893664Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4894002Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4894314Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4894408Z args = parser.parse_args() 2023-03-31T06:22:41.4894413Z 2023-03-31T06:22:41.4894502Z if args.benchmark_kernels: 2023-03-31T06:22:41.4894692Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4894758Z else: 2023-03-31T06:22:41.4894860Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4894957Z benchmark_compiled_module() 2023-03-31T06:22:41.4894963Z 2023-03-31T06:22:41.4895030Z if p: 2023-03-31T06:22:41.4895178Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4895273Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4895410Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4895450Z 2023-03-31T06:22:41.4895730Z [2023-03-31 06:02:17,182] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 86 2023-03-31T06:22:41.4896002Z [2023-03-31 06:02:17,226] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 87 2023-03-31T06:22:41.4896395Z [2023-03-31 06:02:38,480] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/n3/cn3qpi7skpp57cskiu5lbhpaw376mvcdcozlufdrumyhmyazdx4h.py 2023-03-31T06:22:41.4896620Z [2023-03-31 06:02:38,480] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4896628Z 2023-03-31T06:22:41.4896725Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4896799Z import torch 2023-03-31T06:22:41.4896870Z import math 2023-03-31T06:22:41.4896945Z import random 2023-03-31T06:22:41.4897013Z import os 2023-03-31T06:22:41.4897075Z import tempfile 2023-03-31T06:22:41.4897195Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4897203Z 2023-03-31T06:22:41.4897321Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4897446Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4897581Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4897587Z 2023-03-31T06:22:41.4897666Z aten = torch.ops.aten 2023-03-31T06:22:41.4897806Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4897900Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4897905Z 2023-03-31T06:22:41.4897983Z async_compile.wait(globals()) 2023-03-31T06:22:41.4898084Z del async_compile 2023-03-31T06:22:41.4898243Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4898315Z wrapper = ( 2023-03-31T06:22:41.4898396Z ''' 2023-03-31T06:22:41.4898471Z #include 2023-03-31T06:22:41.4898548Z #include 2023-03-31T06:22:41.4898553Z 2023-03-31T06:22:41.4898631Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4898648Z 2023-03-31T06:22:41.4898730Z template 2023-03-31T06:22:41.4898854Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4898940Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4899067Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4899164Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4899288Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4899366Z return kernel_cpp; 2023-03-31T06:22:41.4899415Z } 2023-03-31T06:22:41.4899544Z std::vector call_107(std::vector args) { 2023-03-31T06:22:41.4899639Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4899716Z arg0_1 = args[0]; 2023-03-31T06:22:41.4899791Z arg1_1 = args[1]; 2023-03-31T06:22:41.4899864Z arg2_1 = args[2]; 2023-03-31T06:22:41.4899913Z 2023-03-31T06:22:41.4900011Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4900105Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4900194Z .findSchemaOrThrow( 2023-03-31T06:22:41.4900297Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4900364Z "") 2023-03-31T06:22:41.4900432Z .typed< 2023-03-31T06:22:41.4900526Z at::Tensor( 2023-03-31T06:22:41.4900624Z const at::Tensor& input_t, 2023-03-31T06:22:41.4900725Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4900843Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4900930Z c10::string_view attr, 2023-03-31T06:22:41.4901053Z torch::List> scalars, 2023-03-31T06:22:41.4901176Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4901241Z 2023-03-31T06:22:41.4901498Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {-0.500000000000000, 4}, ""); 2023-03-31T06:22:41.4901570Z arg1_1.reset(); 2023-03-31T06:22:41.4901641Z arg2_1.reset(); 2023-03-31T06:22:41.4901776Z return std::vector({buf0}); 2023-03-31T06:22:41.4901838Z } 2023-03-31T06:22:41.4901918Z ''' 2023-03-31T06:22:41.4901979Z ) 2023-03-31T06:22:41.4901984Z 2023-03-31T06:22:41.4902051Z module = load_inline( 2023-03-31T06:22:41.4902326Z name='inline_extension_cgzzjygjdtyaeq5ee6eazo4mwcm6w5k4l5xxnkhipg3d5ktdbcnm', 2023-03-31T06:22:41.4902513Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4902637Z functions=['call_107'], 2023-03-31T06:22:41.4903022Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4903182Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4903893Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4903903Z 2023-03-31T06:22:41.4903979Z def _wrap_func(f): 2023-03-31T06:22:41.4904054Z def g(args): 2023-03-31T06:22:41.4904116Z return f(args) 2023-03-31T06:22:41.4904184Z return g 2023-03-31T06:22:41.4904282Z call = _wrap_func(module.call_107) 2023-03-31T06:22:41.4904287Z 2023-03-31T06:22:41.4904292Z 2023-03-31T06:22:41.4904386Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4904505Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4904670Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4904915Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4905130Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4905366Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4905534Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4905546Z 2023-03-31T06:22:41.4905551Z 2023-03-31T06:22:41.4905670Z if __name__ == "__main__": 2023-03-31T06:22:41.4905770Z import argparse 2023-03-31T06:22:41.4905956Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4905966Z 2023-03-31T06:22:41.4906114Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4906713Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4907156Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4907440Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4907520Z args = parser.parse_args() 2023-03-31T06:22:41.4907526Z 2023-03-31T06:22:41.4907616Z if args.benchmark_kernels: 2023-03-31T06:22:41.4907816Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4907887Z else: 2023-03-31T06:22:41.4908048Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4908144Z benchmark_compiled_module() 2023-03-31T06:22:41.4908150Z 2023-03-31T06:22:41.4908220Z if p: 2023-03-31T06:22:41.4908358Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4908454Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4908593Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4908601Z 2023-03-31T06:22:41.4908884Z [2023-03-31 06:02:38,481] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 87 2023-03-31T06:22:41.4909151Z [2023-03-31 06:02:38,541] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 88 2023-03-31T06:22:41.4909559Z [2023-03-31 06:02:59,587] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/2w/c2wncvaaspbi5nwsfawyqhwuowypja4rzqigxsds365r2cnqfzdu.py 2023-03-31T06:22:41.4909823Z [2023-03-31 06:02:59,588] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4909829Z 2023-03-31T06:22:41.4909927Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4909998Z import torch 2023-03-31T06:22:41.4910055Z import math 2023-03-31T06:22:41.4910127Z import random 2023-03-31T06:22:41.4910197Z import os 2023-03-31T06:22:41.4910270Z import tempfile 2023-03-31T06:22:41.4910388Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4910397Z 2023-03-31T06:22:41.4910518Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4910643Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4910763Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4910769Z 2023-03-31T06:22:41.4910849Z aten = torch.ops.aten 2023-03-31T06:22:41.4910987Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4911086Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4911091Z 2023-03-31T06:22:41.4911183Z async_compile.wait(globals()) 2023-03-31T06:22:41.4911258Z del async_compile 2023-03-31T06:22:41.4911379Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4911453Z wrapper = ( 2023-03-31T06:22:41.4911523Z ''' 2023-03-31T06:22:41.4911600Z #include 2023-03-31T06:22:41.4911675Z #include 2023-03-31T06:22:41.4911680Z 2023-03-31T06:22:41.4911778Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4911783Z 2023-03-31T06:22:41.4911905Z template 2023-03-31T06:22:41.4912056Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4912142Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4912254Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4912352Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4912477Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4912558Z return kernel_cpp; 2023-03-31T06:22:41.4912621Z } 2023-03-31T06:22:41.4912750Z std::vector call_108(std::vector args) { 2023-03-31T06:22:41.4912856Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4912918Z arg0_1 = args[0]; 2023-03-31T06:22:41.4912990Z arg1_1 = args[1]; 2023-03-31T06:22:41.4913059Z arg2_1 = args[2]; 2023-03-31T06:22:41.4913129Z arg3_1 = args[3]; 2023-03-31T06:22:41.4913192Z 2023-03-31T06:22:41.4913293Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4913387Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4913464Z .findSchemaOrThrow( 2023-03-31T06:22:41.4913562Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4913629Z "") 2023-03-31T06:22:41.4913697Z .typed< 2023-03-31T06:22:41.4913773Z at::Tensor( 2023-03-31T06:22:41.4913870Z const at::Tensor& input_t, 2023-03-31T06:22:41.4913970Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4914077Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4914165Z c10::string_view attr, 2023-03-31T06:22:41.4914317Z torch::List> scalars, 2023-03-31T06:22:41.4914443Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4914510Z 2023-03-31T06:22:41.4914736Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "none"); 2023-03-31T06:22:41.4914808Z arg1_1.reset(); 2023-03-31T06:22:41.4914869Z arg2_1.reset(); 2023-03-31T06:22:41.4914941Z arg3_1.reset(); 2023-03-31T06:22:41.4915050Z return std::vector({buf0}); 2023-03-31T06:22:41.4915112Z } 2023-03-31T06:22:41.4915193Z ''' 2023-03-31T06:22:41.4915254Z ) 2023-03-31T06:22:41.4915260Z 2023-03-31T06:22:41.4915338Z module = load_inline( 2023-03-31T06:22:41.4915596Z name='inline_extension_cwidvkkn5ycr44b4b5libljko5pjbaxwy2dstyiheyapho2qbgzi', 2023-03-31T06:22:41.4915735Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4915854Z functions=['call_108'], 2023-03-31T06:22:41.4916242Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4916400Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4917110Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4917119Z 2023-03-31T06:22:41.4917195Z def _wrap_func(f): 2023-03-31T06:22:41.4917266Z def g(args): 2023-03-31T06:22:41.4917344Z return f(args) 2023-03-31T06:22:41.4917399Z return g 2023-03-31T06:22:41.4917499Z call = _wrap_func(module.call_108) 2023-03-31T06:22:41.4917504Z 2023-03-31T06:22:41.4917509Z 2023-03-31T06:22:41.4917605Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4917728Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4917856Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4918070Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4918280Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4918524Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4918757Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4918896Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4918902Z 2023-03-31T06:22:41.4918907Z 2023-03-31T06:22:41.4918985Z if __name__ == "__main__": 2023-03-31T06:22:41.4919064Z import argparse 2023-03-31T06:22:41.4919200Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4919206Z 2023-03-31T06:22:41.4919317Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4919629Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4919969Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4920247Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4920330Z args = parser.parse_args() 2023-03-31T06:22:41.4920335Z 2023-03-31T06:22:41.4920425Z if args.benchmark_kernels: 2023-03-31T06:22:41.4920620Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4920690Z else: 2023-03-31T06:22:41.4920805Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4920905Z benchmark_compiled_module() 2023-03-31T06:22:41.4920911Z 2023-03-31T06:22:41.4920981Z if p: 2023-03-31T06:22:41.4921149Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4921247Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4921386Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4921392Z 2023-03-31T06:22:41.4921674Z [2023-03-31 06:02:59,588] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 88 2023-03-31T06:22:41.4921948Z [2023-03-31 06:02:59,630] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 89 2023-03-31T06:22:41.4922357Z [2023-03-31 06:03:20,798] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/g2/cg2cavwcyupbcviwxlprlg6tac4ialb3dpg5zg6gmkuxso7jmq3y.py 2023-03-31T06:22:41.4922580Z [2023-03-31 06:03:20,799] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4922614Z 2023-03-31T06:22:41.4922714Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4922787Z import torch 2023-03-31T06:22:41.4922844Z import math 2023-03-31T06:22:41.4922922Z import random 2023-03-31T06:22:41.4922994Z import os 2023-03-31T06:22:41.4923224Z import tempfile 2023-03-31T06:22:41.4923345Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4923352Z 2023-03-31T06:22:41.4923472Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4923598Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4923721Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4923744Z 2023-03-31T06:22:41.4923812Z aten = torch.ops.aten 2023-03-31T06:22:41.4923954Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4924050Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4924056Z 2023-03-31T06:22:41.4924149Z async_compile.wait(globals()) 2023-03-31T06:22:41.4924227Z del async_compile 2023-03-31T06:22:41.4924350Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4924425Z wrapper = ( 2023-03-31T06:22:41.4924496Z ''' 2023-03-31T06:22:41.4924578Z #include 2023-03-31T06:22:41.4924660Z #include 2023-03-31T06:22:41.4924665Z 2023-03-31T06:22:41.4924759Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4924764Z 2023-03-31T06:22:41.4924859Z template 2023-03-31T06:22:41.4924985Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4925072Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4925250Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4925382Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4925507Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4925590Z return kernel_cpp; 2023-03-31T06:22:41.4925653Z } 2023-03-31T06:22:41.4925787Z std::vector call_109(std::vector args) { 2023-03-31T06:22:41.4925882Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4925943Z arg0_1 = args[0]; 2023-03-31T06:22:41.4926018Z arg1_1 = args[1]; 2023-03-31T06:22:41.4926092Z arg2_1 = args[2]; 2023-03-31T06:22:41.4926156Z 2023-03-31T06:22:41.4926256Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4926352Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4926441Z .findSchemaOrThrow( 2023-03-31T06:22:41.4926526Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4926595Z "") 2023-03-31T06:22:41.4926664Z .typed< 2023-03-31T06:22:41.4926744Z at::Tensor( 2023-03-31T06:22:41.4926844Z const at::Tensor& input_t, 2023-03-31T06:22:41.4926945Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4927063Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4927140Z c10::string_view attr, 2023-03-31T06:22:41.4927259Z torch::List> scalars, 2023-03-31T06:22:41.4927382Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4927448Z 2023-03-31T06:22:41.4927731Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "none"); 2023-03-31T06:22:41.4927805Z arg1_1.reset(); 2023-03-31T06:22:41.4927877Z arg2_1.reset(); 2023-03-31T06:22:41.4927981Z return std::vector({buf0}); 2023-03-31T06:22:41.4928032Z } 2023-03-31T06:22:41.4928115Z ''' 2023-03-31T06:22:41.4928175Z ) 2023-03-31T06:22:41.4928181Z 2023-03-31T06:22:41.4928261Z module = load_inline( 2023-03-31T06:22:41.4928523Z name='inline_extension_cm5x7bscub67egu6kil23wnc2ntd2qsrllom36y3vydcvodxeoao', 2023-03-31T06:22:41.4928613Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4928730Z functions=['call_109'], 2023-03-31T06:22:41.4929099Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4929334Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4930057Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4930064Z 2023-03-31T06:22:41.4930144Z def _wrap_func(f): 2023-03-31T06:22:41.4930220Z def g(args): 2023-03-31T06:22:41.4930298Z return f(args) 2023-03-31T06:22:41.4930366Z return g 2023-03-31T06:22:41.4930464Z call = _wrap_func(module.call_109) 2023-03-31T06:22:41.4930470Z 2023-03-31T06:22:41.4930475Z 2023-03-31T06:22:41.4930571Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4930678Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4930805Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4931020Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4931236Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4931455Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4931583Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4931588Z 2023-03-31T06:22:41.4931593Z 2023-03-31T06:22:41.4931671Z if __name__ == "__main__": 2023-03-31T06:22:41.4931778Z import argparse 2023-03-31T06:22:41.4931922Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4931928Z 2023-03-31T06:22:41.4932039Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4932351Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4932684Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4932965Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4933061Z args = parser.parse_args() 2023-03-31T06:22:41.4933066Z 2023-03-31T06:22:41.4933155Z if args.benchmark_kernels: 2023-03-31T06:22:41.4933344Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4933399Z else: 2023-03-31T06:22:41.4933514Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4933610Z benchmark_compiled_module() 2023-03-31T06:22:41.4933615Z 2023-03-31T06:22:41.4933684Z if p: 2023-03-31T06:22:41.4933836Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4933935Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4934074Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4934082Z 2023-03-31T06:22:41.4934360Z [2023-03-31 06:03:20,799] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 89 2023-03-31T06:22:41.4934643Z [2023-03-31 06:03:20,854] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 90 2023-03-31T06:22:41.4935046Z [2023-03-31 06:03:41,890] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/gg/cggw72w7zzvy2vhys4onpnrfwbogsybxdm6nboeazgyia5ag2y3d.py 2023-03-31T06:22:41.4935270Z [2023-03-31 06:03:41,890] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4935277Z 2023-03-31T06:22:41.4935375Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4935448Z import torch 2023-03-31T06:22:41.4935518Z import math 2023-03-31T06:22:41.4935590Z import random 2023-03-31T06:22:41.4935659Z import os 2023-03-31T06:22:41.4935719Z import tempfile 2023-03-31T06:22:41.4935836Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4935870Z 2023-03-31T06:22:41.4935991Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4936116Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4936250Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4936257Z 2023-03-31T06:22:41.4936338Z aten = torch.ops.aten 2023-03-31T06:22:41.4936477Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4936572Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4936577Z 2023-03-31T06:22:41.4936656Z async_compile.wait(globals()) 2023-03-31T06:22:41.4936734Z del async_compile 2023-03-31T06:22:41.4936858Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4936930Z wrapper = ( 2023-03-31T06:22:41.4937013Z ''' 2023-03-31T06:22:41.4937090Z #include 2023-03-31T06:22:41.4937166Z #include 2023-03-31T06:22:41.4937171Z 2023-03-31T06:22:41.4937250Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4937255Z 2023-03-31T06:22:41.4937349Z template 2023-03-31T06:22:41.4937472Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4937557Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4937684Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4937781Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4937906Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4937972Z return kernel_cpp; 2023-03-31T06:22:41.4938033Z } 2023-03-31T06:22:41.4938191Z std::vector call_110(std::vector args) { 2023-03-31T06:22:41.4938330Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4938404Z arg0_1 = args[0]; 2023-03-31T06:22:41.4938477Z arg1_1 = args[1]; 2023-03-31T06:22:41.4938547Z arg2_1 = args[2]; 2023-03-31T06:22:41.4938607Z arg3_1 = args[3]; 2023-03-31T06:22:41.4938670Z 2023-03-31T06:22:41.4938768Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4938866Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4938955Z .findSchemaOrThrow( 2023-03-31T06:22:41.4939052Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4939119Z "") 2023-03-31T06:22:41.4939176Z .typed< 2023-03-31T06:22:41.4939251Z at::Tensor( 2023-03-31T06:22:41.4939349Z const at::Tensor& input_t, 2023-03-31T06:22:41.4939450Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4939568Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4939658Z c10::string_view attr, 2023-03-31T06:22:41.4939782Z torch::List> scalars, 2023-03-31T06:22:41.4939893Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4939961Z 2023-03-31T06:22:41.4940191Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "none"); 2023-03-31T06:22:41.4940267Z arg1_1.reset(); 2023-03-31T06:22:41.4940340Z arg2_1.reset(); 2023-03-31T06:22:41.4940411Z arg3_1.reset(); 2023-03-31T06:22:41.4940516Z return std::vector({buf0}); 2023-03-31T06:22:41.4940566Z } 2023-03-31T06:22:41.4940678Z ''' 2023-03-31T06:22:41.4940740Z ) 2023-03-31T06:22:41.4940745Z 2023-03-31T06:22:41.4940822Z module = load_inline( 2023-03-31T06:22:41.4941091Z name='inline_extension_cwidvkkn5ycr44b4b5libljko5pjbaxwy2dstyiheyapho2qbgzi', 2023-03-31T06:22:41.4941177Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4941293Z functions=['call_110'], 2023-03-31T06:22:41.4941677Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4941823Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4942639Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4942688Z 2023-03-31T06:22:41.4942769Z def _wrap_func(f): 2023-03-31T06:22:41.4942845Z def g(args): 2023-03-31T06:22:41.4942921Z return f(args) 2023-03-31T06:22:41.4942989Z return g 2023-03-31T06:22:41.4943088Z call = _wrap_func(module.call_110) 2023-03-31T06:22:41.4943094Z 2023-03-31T06:22:41.4943099Z 2023-03-31T06:22:41.4943197Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4943308Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4943441Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4943663Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4943873Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4944082Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4944288Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4944426Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4944432Z 2023-03-31T06:22:41.4944436Z 2023-03-31T06:22:41.4944519Z if __name__ == "__main__": 2023-03-31T06:22:41.4944596Z import argparse 2023-03-31T06:22:41.4944715Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4944721Z 2023-03-31T06:22:41.4944862Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4945205Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4945546Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4945828Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4945925Z args = parser.parse_args() 2023-03-31T06:22:41.4945931Z 2023-03-31T06:22:41.4946019Z if args.benchmark_kernels: 2023-03-31T06:22:41.4946210Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4946265Z else: 2023-03-31T06:22:41.4946377Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4946472Z benchmark_compiled_module() 2023-03-31T06:22:41.4946478Z 2023-03-31T06:22:41.4946545Z if p: 2023-03-31T06:22:41.4946697Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4946797Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4946935Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4946940Z 2023-03-31T06:22:41.4947222Z [2023-03-31 06:03:41,890] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 90 2023-03-31T06:22:41.4947477Z [2023-03-31 06:03:41,930] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 91 2023-03-31T06:22:41.4947912Z [2023-03-31 06:04:03,244] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/6d/c6dlokykqs2ogrzwcl6p5ykmylr6yg6avglarfghzlwjgoqy7ftn.py 2023-03-31T06:22:41.4948136Z [2023-03-31 06:04:03,244] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4948142Z 2023-03-31T06:22:41.4948240Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4948312Z import torch 2023-03-31T06:22:41.4948385Z import math 2023-03-31T06:22:41.4948458Z import random 2023-03-31T06:22:41.4948525Z import os 2023-03-31T06:22:41.4948586Z import tempfile 2023-03-31T06:22:41.4948705Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4948711Z 2023-03-31T06:22:41.4948832Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4948958Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4949121Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4949126Z 2023-03-31T06:22:41.4949207Z aten = torch.ops.aten 2023-03-31T06:22:41.4949349Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4949443Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4949448Z 2023-03-31T06:22:41.4949529Z async_compile.wait(globals()) 2023-03-31T06:22:41.4949603Z del async_compile 2023-03-31T06:22:41.4949724Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4949796Z wrapper = ( 2023-03-31T06:22:41.4949880Z ''' 2023-03-31T06:22:41.4949957Z #include 2023-03-31T06:22:41.4950036Z #include 2023-03-31T06:22:41.4950041Z 2023-03-31T06:22:41.4950122Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4950127Z 2023-03-31T06:22:41.4950219Z template 2023-03-31T06:22:41.4950342Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4950431Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4950560Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4950656Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4950784Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4950851Z return kernel_cpp; 2023-03-31T06:22:41.4950913Z } 2023-03-31T06:22:41.4951042Z std::vector call_111(std::vector args) { 2023-03-31T06:22:41.4951135Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4951209Z arg0_1 = args[0]; 2023-03-31T06:22:41.4951281Z arg1_1 = args[1]; 2023-03-31T06:22:41.4951380Z arg2_1 = args[2]; 2023-03-31T06:22:41.4951464Z 2023-03-31T06:22:41.4951563Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4951657Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4951745Z .findSchemaOrThrow( 2023-03-31T06:22:41.4951843Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4951910Z "") 2023-03-31T06:22:41.4951978Z .typed< 2023-03-31T06:22:41.4952044Z at::Tensor( 2023-03-31T06:22:41.4952142Z const at::Tensor& input_t, 2023-03-31T06:22:41.4952243Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4952362Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4952451Z c10::string_view attr, 2023-03-31T06:22:41.4952571Z torch::List> scalars, 2023-03-31T06:22:41.4952693Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4952746Z 2023-03-31T06:22:41.4952990Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "none"); 2023-03-31T06:22:41.4953067Z arg1_1.reset(); 2023-03-31T06:22:41.4953139Z arg2_1.reset(); 2023-03-31T06:22:41.4953244Z return std::vector({buf0}); 2023-03-31T06:22:41.4953310Z } 2023-03-31T06:22:41.4953392Z ''' 2023-03-31T06:22:41.4953441Z ) 2023-03-31T06:22:41.4953459Z 2023-03-31T06:22:41.4953524Z module = load_inline( 2023-03-31T06:22:41.4953788Z name='inline_extension_cm5x7bscub67egu6kil23wnc2ntd2qsrllom36y3vydcvodxeoao', 2023-03-31T06:22:41.4953876Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4954022Z functions=['call_111'], 2023-03-31T06:22:41.4954402Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4954559Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4955265Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4955274Z 2023-03-31T06:22:41.4955353Z def _wrap_func(f): 2023-03-31T06:22:41.4955412Z def g(args): 2023-03-31T06:22:41.4955519Z return f(args) 2023-03-31T06:22:41.4955586Z return g 2023-03-31T06:22:41.4955682Z call = _wrap_func(module.call_111) 2023-03-31T06:22:41.4955687Z 2023-03-31T06:22:41.4955694Z 2023-03-31T06:22:41.4955789Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4955910Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4956041Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4956257Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4956458Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4956667Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4956794Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4956799Z 2023-03-31T06:22:41.4956804Z 2023-03-31T06:22:41.4956882Z if __name__ == "__main__": 2023-03-31T06:22:41.4956960Z import argparse 2023-03-31T06:22:41.4957091Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4957097Z 2023-03-31T06:22:41.4957211Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4957526Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4957865Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4958167Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4958292Z args = parser.parse_args() 2023-03-31T06:22:41.4958298Z 2023-03-31T06:22:41.4958387Z if args.benchmark_kernels: 2023-03-31T06:22:41.4958579Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4958648Z else: 2023-03-31T06:22:41.4958763Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4958862Z benchmark_compiled_module() 2023-03-31T06:22:41.4958868Z 2023-03-31T06:22:41.4958937Z if p: 2023-03-31T06:22:41.4959076Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4959175Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4959314Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4959319Z 2023-03-31T06:22:41.4959600Z [2023-03-31 06:04:03,244] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 91 2023-03-31T06:22:41.4959872Z [2023-03-31 06:04:03,300] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 92 2023-03-31T06:22:41.4960281Z [2023-03-31 06:04:24,455] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/h6/ch6mfvkktixa5hh6inezei4tlhs6qqhefnh4m4l2rxg4u5vrnl3v.py 2023-03-31T06:22:41.4960503Z [2023-03-31 06:04:24,456] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4960511Z 2023-03-31T06:22:41.4960610Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4960670Z import torch 2023-03-31T06:22:41.4960743Z import math 2023-03-31T06:22:41.4960845Z import random 2023-03-31T06:22:41.4960918Z import os 2023-03-31T06:22:41.4960994Z import tempfile 2023-03-31T06:22:41.4961113Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4961119Z 2023-03-31T06:22:41.4961240Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4961352Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4961492Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4961499Z 2023-03-31T06:22:41.4961581Z aten = torch.ops.aten 2023-03-31T06:22:41.4961721Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4961817Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4961822Z 2023-03-31T06:22:41.4961914Z async_compile.wait(globals()) 2023-03-31T06:22:41.4962034Z del async_compile 2023-03-31T06:22:41.4962156Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4962216Z wrapper = ( 2023-03-31T06:22:41.4962300Z ''' 2023-03-31T06:22:41.4962381Z #include 2023-03-31T06:22:41.4962459Z #include 2023-03-31T06:22:41.4962465Z 2023-03-31T06:22:41.4962558Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4962563Z 2023-03-31T06:22:41.4962657Z template 2023-03-31T06:22:41.4962785Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4962859Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4962988Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4963255Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4963383Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4963463Z return kernel_cpp; 2023-03-31T06:22:41.4963528Z } 2023-03-31T06:22:41.4963661Z std::vector call_112(std::vector args) { 2023-03-31T06:22:41.4963755Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4963833Z arg0_1 = args[0]; 2023-03-31T06:22:41.4963907Z arg1_1 = args[1]; 2023-03-31T06:22:41.4963981Z arg2_1 = args[2]; 2023-03-31T06:22:41.4964054Z arg3_1 = args[3]; 2023-03-31T06:22:41.4964117Z 2023-03-31T06:22:41.4964217Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4964297Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4964390Z .findSchemaOrThrow( 2023-03-31T06:22:41.4964490Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4964558Z "") 2023-03-31T06:22:41.4964685Z .typed< 2023-03-31T06:22:41.4964797Z at::Tensor( 2023-03-31T06:22:41.4964897Z const at::Tensor& input_t, 2023-03-31T06:22:41.4964986Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4965105Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4965194Z c10::string_view attr, 2023-03-31T06:22:41.4965319Z torch::List> scalars, 2023-03-31T06:22:41.4965447Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4965515Z 2023-03-31T06:22:41.4965748Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4965823Z arg1_1.reset(); 2023-03-31T06:22:41.4965882Z arg2_1.reset(); 2023-03-31T06:22:41.4965954Z arg3_1.reset(); 2023-03-31T06:22:41.4966060Z return std::vector({buf0}); 2023-03-31T06:22:41.4966124Z } 2023-03-31T06:22:41.4966205Z ''' 2023-03-31T06:22:41.4966271Z ) 2023-03-31T06:22:41.4966280Z 2023-03-31T06:22:41.4966345Z module = load_inline( 2023-03-31T06:22:41.4966617Z name='inline_extension_ctlllk2n4h2r3nrblvgymuc5rhrnmphwpwykxvktrtxfj2cfcnbv', 2023-03-31T06:22:41.4966706Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4966824Z functions=['call_112'], 2023-03-31T06:22:41.4967208Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4967363Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4968107Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4968115Z 2023-03-31T06:22:41.4968195Z def _wrap_func(f): 2023-03-31T06:22:41.4968267Z def g(args): 2023-03-31T06:22:41.4968330Z return f(args) 2023-03-31T06:22:41.4968397Z return g 2023-03-31T06:22:41.4968493Z call = _wrap_func(module.call_112) 2023-03-31T06:22:41.4968498Z 2023-03-31T06:22:41.4968503Z 2023-03-31T06:22:41.4968599Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4968718Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4968880Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4969097Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4969310Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4969506Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4969721Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4969857Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4969864Z 2023-03-31T06:22:41.4969869Z 2023-03-31T06:22:41.4969947Z if __name__ == "__main__": 2023-03-31T06:22:41.4970023Z import argparse 2023-03-31T06:22:41.4970154Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4970160Z 2023-03-31T06:22:41.4970270Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4970583Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4970925Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4971190Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4971282Z args = parser.parse_args() 2023-03-31T06:22:41.4971288Z 2023-03-31T06:22:41.4971404Z if args.benchmark_kernels: 2023-03-31T06:22:41.4971621Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4971688Z else: 2023-03-31T06:22:41.4971801Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4971896Z benchmark_compiled_module() 2023-03-31T06:22:41.4971902Z 2023-03-31T06:22:41.4971972Z if p: 2023-03-31T06:22:41.4972109Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4972207Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4972345Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4972351Z 2023-03-31T06:22:41.4972629Z [2023-03-31 06:04:24,456] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 92 2023-03-31T06:22:41.4972896Z [2023-03-31 06:04:24,499] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 93 2023-03-31T06:22:41.4973306Z [2023-03-31 06:04:45,593] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/jq/cjqayenhttxcvebiouw4hcqscj5lyi75wf4uo4wcb3qwnaiu75nm.py 2023-03-31T06:22:41.4973526Z [2023-03-31 06:04:45,593] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4973532Z 2023-03-31T06:22:41.4973628Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4973688Z import torch 2023-03-31T06:22:41.4973757Z import math 2023-03-31T06:22:41.4973831Z import random 2023-03-31T06:22:41.4973898Z import os 2023-03-31T06:22:41.4973972Z import tempfile 2023-03-31T06:22:41.4974088Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4974129Z 2023-03-31T06:22:41.4974248Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4974361Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4974496Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4974501Z 2023-03-31T06:22:41.4974581Z aten = torch.ops.aten 2023-03-31T06:22:41.4974721Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4974819Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4974825Z 2023-03-31T06:22:41.4974915Z async_compile.wait(globals()) 2023-03-31T06:22:41.4974989Z del async_compile 2023-03-31T06:22:41.4975110Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4975169Z wrapper = ( 2023-03-31T06:22:41.4975251Z ''' 2023-03-31T06:22:41.4975370Z #include 2023-03-31T06:22:41.4975446Z #include 2023-03-31T06:22:41.4975452Z 2023-03-31T06:22:41.4975543Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4975548Z 2023-03-31T06:22:41.4975643Z template 2023-03-31T06:22:41.4975767Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4975840Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4975964Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4976060Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4976188Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4976268Z return kernel_cpp; 2023-03-31T06:22:41.4976331Z } 2023-03-31T06:22:41.4976461Z std::vector call_113(std::vector args) { 2023-03-31T06:22:41.4976542Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.4976615Z arg0_1 = args[0]; 2023-03-31T06:22:41.4976687Z arg1_1 = args[1]; 2023-03-31T06:22:41.4976757Z arg2_1 = args[2]; 2023-03-31T06:22:41.4976822Z 2023-03-31T06:22:41.4976920Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4977012Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4977089Z .findSchemaOrThrow( 2023-03-31T06:22:41.4977186Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4977253Z "") 2023-03-31T06:22:41.4977321Z .typed< 2023-03-31T06:22:41.4977395Z at::Tensor( 2023-03-31T06:22:41.4977492Z const at::Tensor& input_t, 2023-03-31T06:22:41.4977595Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4977729Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4977845Z c10::string_view attr, 2023-03-31T06:22:41.4977965Z torch::List> scalars, 2023-03-31T06:22:41.4978086Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4978152Z 2023-03-31T06:22:41.4978394Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4978470Z arg1_1.reset(); 2023-03-31T06:22:41.4978541Z arg2_1.reset(); 2023-03-31T06:22:41.4978636Z return std::vector({buf0}); 2023-03-31T06:22:41.4978698Z } 2023-03-31T06:22:41.4978778Z ''' 2023-03-31T06:22:41.4978840Z ) 2023-03-31T06:22:41.4978846Z 2023-03-31T06:22:41.4978923Z module = load_inline( 2023-03-31T06:22:41.4979191Z name='inline_extension_crz2jjgr4amu57dazhb33qd7qxxdtpn2teqtf6b3zcjdkld77yyo', 2023-03-31T06:22:41.4979278Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4979384Z functions=['call_113'], 2023-03-31T06:22:41.4979766Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4979921Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4980699Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4980709Z 2023-03-31T06:22:41.4980785Z def _wrap_func(f): 2023-03-31T06:22:41.4980856Z def g(args): 2023-03-31T06:22:41.4980932Z return f(args) 2023-03-31T06:22:41.4980999Z return g 2023-03-31T06:22:41.4981082Z call = _wrap_func(module.call_113) 2023-03-31T06:22:41.4981100Z 2023-03-31T06:22:41.4981107Z 2023-03-31T06:22:41.4981191Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4981310Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4981440Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4981655Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4981866Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4982111Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4982244Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.4982249Z 2023-03-31T06:22:41.4982254Z 2023-03-31T06:22:41.4982333Z if __name__ == "__main__": 2023-03-31T06:22:41.4982496Z import argparse 2023-03-31T06:22:41.4982633Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4982639Z 2023-03-31T06:22:41.4982750Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4983066Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4983406Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4983683Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4983779Z args = parser.parse_args() 2023-03-31T06:22:41.4983784Z 2023-03-31T06:22:41.4983874Z if args.benchmark_kernels: 2023-03-31T06:22:41.4984067Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4984121Z else: 2023-03-31T06:22:41.4984233Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4984330Z benchmark_compiled_module() 2023-03-31T06:22:41.4984336Z 2023-03-31T06:22:41.4984405Z if p: 2023-03-31T06:22:41.4984593Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4984724Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4984862Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4984868Z 2023-03-31T06:22:41.4985133Z [2023-03-31 06:04:45,593] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 93 2023-03-31T06:22:41.4985403Z [2023-03-31 06:04:45,659] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 94 2023-03-31T06:22:41.4985810Z [2023-03-31 06:05:06,720] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ba/cbapn3zz5pmhjosm7xxk6osez3qnifgaas53ercbyovfqbbwbkb5.py 2023-03-31T06:22:41.4986028Z [2023-03-31 06:05:06,720] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4986034Z 2023-03-31T06:22:41.4986130Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4986202Z import torch 2023-03-31T06:22:41.4986273Z import math 2023-03-31T06:22:41.4986348Z import random 2023-03-31T06:22:41.4986406Z import os 2023-03-31T06:22:41.4986481Z import tempfile 2023-03-31T06:22:41.4986600Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4986606Z 2023-03-31T06:22:41.4986725Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4986849Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4986985Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4986991Z 2023-03-31T06:22:41.4987072Z aten = torch.ops.aten 2023-03-31T06:22:41.4987238Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.4987322Z async_compile = AsyncCompile() 2023-03-31T06:22:41.4987328Z 2023-03-31T06:22:41.4987419Z async_compile.wait(globals()) 2023-03-31T06:22:41.4987494Z del async_compile 2023-03-31T06:22:41.4987615Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.4987687Z wrapper = ( 2023-03-31T06:22:41.4987769Z ''' 2023-03-31T06:22:41.4987848Z #include 2023-03-31T06:22:41.4987914Z #include 2023-03-31T06:22:41.4987920Z 2023-03-31T06:22:41.4988013Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.4988018Z 2023-03-31T06:22:41.4988110Z template 2023-03-31T06:22:41.4988233Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.4988320Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.4988472Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.4988569Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.4988684Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.4988763Z return kernel_cpp; 2023-03-31T06:22:41.4988824Z } 2023-03-31T06:22:41.4988954Z std::vector call_114(std::vector args) { 2023-03-31T06:22:41.4989057Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.4989130Z arg0_1 = args[0]; 2023-03-31T06:22:41.4989204Z arg1_1 = args[1]; 2023-03-31T06:22:41.4989266Z arg2_1 = args[2]; 2023-03-31T06:22:41.4989342Z arg3_1 = args[3]; 2023-03-31T06:22:41.4989403Z 2023-03-31T06:22:41.4989501Z static auto op_linear_pointwise = 2023-03-31T06:22:41.4989594Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.4989682Z .findSchemaOrThrow( 2023-03-31T06:22:41.4989780Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.4989834Z "") 2023-03-31T06:22:41.4989903Z .typed< 2023-03-31T06:22:41.4989977Z at::Tensor( 2023-03-31T06:22:41.4990075Z const at::Tensor& input_t, 2023-03-31T06:22:41.4990178Z const at::Tensor& weight_t, 2023-03-31T06:22:41.4990294Z const c10::optional& bias_opt, 2023-03-31T06:22:41.4990383Z c10::string_view attr, 2023-03-31T06:22:41.4990503Z torch::List> scalars, 2023-03-31T06:22:41.4990612Z c10::optional algorithm)>(); 2023-03-31T06:22:41.4990707Z 2023-03-31T06:22:41.4990959Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.4991032Z arg1_1.reset(); 2023-03-31T06:22:41.4991104Z arg2_1.reset(); 2023-03-31T06:22:41.4991173Z arg3_1.reset(); 2023-03-31T06:22:41.4991277Z return std::vector({buf0}); 2023-03-31T06:22:41.4991327Z } 2023-03-31T06:22:41.4991408Z ''' 2023-03-31T06:22:41.4991473Z ) 2023-03-31T06:22:41.4991478Z 2023-03-31T06:22:41.4991562Z module = load_inline( 2023-03-31T06:22:41.4991833Z name='inline_extension_ctlllk2n4h2r3nrblvgymuc5rhrnmphwpwykxvktrtxfj2cfcnbv', 2023-03-31T06:22:41.4991919Z cpp_sources=[wrapper], 2023-03-31T06:22:41.4992035Z functions=['call_114'], 2023-03-31T06:22:41.4992399Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.4992556Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.4993269Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.4993278Z 2023-03-31T06:22:41.4993357Z def _wrap_func(f): 2023-03-31T06:22:41.4993427Z def g(args): 2023-03-31T06:22:41.4993502Z return f(args) 2023-03-31T06:22:41.4993569Z return g 2023-03-31T06:22:41.4993692Z call = _wrap_func(module.call_114) 2023-03-31T06:22:41.4993699Z 2023-03-31T06:22:41.4993704Z 2023-03-31T06:22:41.4993798Z def benchmark_compiled_module(): 2023-03-31T06:22:41.4993904Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.4994031Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.4994249Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4994458Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4994666Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4994874Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.4995009Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.4995046Z 2023-03-31T06:22:41.4995050Z 2023-03-31T06:22:41.4995131Z if __name__ == "__main__": 2023-03-31T06:22:41.4995198Z import argparse 2023-03-31T06:22:41.4995329Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.4995334Z 2023-03-31T06:22:41.4995445Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.4995756Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.4996094Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.4996372Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.4996466Z args = parser.parse_args() 2023-03-31T06:22:41.4996472Z 2023-03-31T06:22:41.4996562Z if args.benchmark_kernels: 2023-03-31T06:22:41.4996757Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.4996811Z else: 2023-03-31T06:22:41.4996926Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.4997026Z benchmark_compiled_module() 2023-03-31T06:22:41.4997031Z 2023-03-31T06:22:41.4997105Z if p: 2023-03-31T06:22:41.4997259Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.4997358Z p.export_chrome_trace(path) 2023-03-31T06:22:41.4997525Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.4997557Z 2023-03-31T06:22:41.4997825Z [2023-03-31 06:05:06,721] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 94 2023-03-31T06:22:41.4998093Z [2023-03-31 06:05:06,767] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 95 2023-03-31T06:22:41.4998502Z [2023-03-31 06:05:27,797] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zp/czpp2ey5aemuhvhls2sejwcjbugip2fjwfr2cghyqaqpwefwukfx.py 2023-03-31T06:22:41.4998731Z [2023-03-31 06:05:27,797] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.4998737Z 2023-03-31T06:22:41.4998837Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.4998912Z import torch 2023-03-31T06:22:41.4998985Z import math 2023-03-31T06:22:41.4999060Z import random 2023-03-31T06:22:41.4999115Z import os 2023-03-31T06:22:41.4999190Z import tempfile 2023-03-31T06:22:41.4999312Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.4999320Z 2023-03-31T06:22:41.4999446Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.4999572Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.4999709Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.4999715Z 2023-03-31T06:22:41.4999797Z aten = torch.ops.aten 2023-03-31T06:22:41.4999939Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5000025Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5000030Z 2023-03-31T06:22:41.5000122Z async_compile.wait(globals()) 2023-03-31T06:22:41.5000229Z del async_compile 2023-03-31T06:22:41.5000356Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5000431Z wrapper = ( 2023-03-31T06:22:41.5000515Z ''' 2023-03-31T06:22:41.5000593Z #include 2023-03-31T06:22:41.5000657Z #include 2023-03-31T06:22:41.5000662Z 2023-03-31T06:22:41.5000755Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5000760Z 2023-03-31T06:22:41.5000860Z template 2023-03-31T06:22:41.5000987Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5001075Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5001202Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5001301Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5001415Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5001529Z return kernel_cpp; 2023-03-31T06:22:41.5001593Z } 2023-03-31T06:22:41.5001726Z std::vector call_115(std::vector args) { 2023-03-31T06:22:41.5001823Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5006787Z arg0_1 = args[0]; 2023-03-31T06:22:41.5006891Z arg1_1 = args[1]; 2023-03-31T06:22:41.5006953Z arg2_1 = args[2]; 2023-03-31T06:22:41.5007018Z 2023-03-31T06:22:41.5007122Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5007229Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5007325Z .findSchemaOrThrow( 2023-03-31T06:22:41.5007431Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5007501Z "") 2023-03-31T06:22:41.5007560Z .typed< 2023-03-31T06:22:41.5007637Z at::Tensor( 2023-03-31T06:22:41.5007741Z const at::Tensor& input_t, 2023-03-31T06:22:41.5007845Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5007966Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5008060Z c10::string_view attr, 2023-03-31T06:22:41.5008188Z torch::List> scalars, 2023-03-31T06:22:41.5008315Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5008368Z 2023-03-31T06:22:41.5008652Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "gelu", {-1}, "tanh"); 2023-03-31T06:22:41.5008728Z arg1_1.reset(); 2023-03-31T06:22:41.5008801Z arg2_1.reset(); 2023-03-31T06:22:41.5009034Z return std::vector({buf0}); 2023-03-31T06:22:41.5009147Z } 2023-03-31T06:22:41.5009235Z ''' 2023-03-31T06:22:41.5009284Z ) 2023-03-31T06:22:41.5009292Z 2023-03-31T06:22:41.5009372Z module = load_inline( 2023-03-31T06:22:41.5009641Z name='inline_extension_crz2jjgr4amu57dazhb33qd7qxxdtpn2teqtf6b3zcjdkld77yyo', 2023-03-31T06:22:41.5009730Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5009850Z functions=['call_115'], 2023-03-31T06:22:41.5010244Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5010406Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5011123Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5011132Z 2023-03-31T06:22:41.5011213Z def _wrap_func(f): 2023-03-31T06:22:41.5011272Z def g(args): 2023-03-31T06:22:41.5011349Z return f(args) 2023-03-31T06:22:41.5011421Z return g 2023-03-31T06:22:41.5011521Z call = _wrap_func(module.call_115) 2023-03-31T06:22:41.5011526Z 2023-03-31T06:22:41.5011534Z 2023-03-31T06:22:41.5011630Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5011751Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5011925Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5012129Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5012342Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5012548Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5012681Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5012688Z 2023-03-31T06:22:41.5012692Z 2023-03-31T06:22:41.5012773Z if __name__ == "__main__": 2023-03-31T06:22:41.5012851Z import argparse 2023-03-31T06:22:41.5012981Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5012986Z 2023-03-31T06:22:41.5013100Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5013479Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5013807Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5014083Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5014178Z args = parser.parse_args() 2023-03-31T06:22:41.5014183Z 2023-03-31T06:22:41.5014274Z if args.benchmark_kernels: 2023-03-31T06:22:41.5014468Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5014538Z else: 2023-03-31T06:22:41.5014653Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5014750Z benchmark_compiled_module() 2023-03-31T06:22:41.5014756Z 2023-03-31T06:22:41.5014812Z if p: 2023-03-31T06:22:41.5014963Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5015063Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5015201Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5015209Z 2023-03-31T06:22:41.5015489Z [2023-03-31 06:05:27,798] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 95 2023-03-31T06:22:41.5015758Z [2023-03-31 06:05:27,854] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 96 2023-03-31T06:22:41.5016198Z [2023-03-31 06:05:49,102] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3r/c3rkuhb5deox55bqzxwrafvryfwuwffl7symgyzpozycpzh7e7uj.py 2023-03-31T06:22:41.5016446Z [2023-03-31 06:05:49,102] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5016452Z 2023-03-31T06:22:41.5016549Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5016608Z import torch 2023-03-31T06:22:41.5016677Z import math 2023-03-31T06:22:41.5016753Z import random 2023-03-31T06:22:41.5016824Z import os 2023-03-31T06:22:41.5016897Z import tempfile 2023-03-31T06:22:41.5017014Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5017023Z 2023-03-31T06:22:41.5017139Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5017252Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5017385Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5017390Z 2023-03-31T06:22:41.5017471Z aten = torch.ops.aten 2023-03-31T06:22:41.5017612Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5017709Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5017715Z 2023-03-31T06:22:41.5017810Z async_compile.wait(globals()) 2023-03-31T06:22:41.5017884Z del async_compile 2023-03-31T06:22:41.5017992Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5018064Z wrapper = ( 2023-03-31T06:22:41.5018145Z ''' 2023-03-31T06:22:41.5018222Z #include 2023-03-31T06:22:41.5018302Z #include 2023-03-31T06:22:41.5018307Z 2023-03-31T06:22:41.5018399Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5018404Z 2023-03-31T06:22:41.5018523Z template 2023-03-31T06:22:41.5018637Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5018721Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5018848Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5018945Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5019073Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5019154Z return kernel_cpp; 2023-03-31T06:22:41.5019216Z } 2023-03-31T06:22:41.5019336Z std::vector call_117(std::vector args) { 2023-03-31T06:22:41.5019440Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5019514Z arg0_1 = args[0]; 2023-03-31T06:22:41.5019586Z arg1_1 = args[1]; 2023-03-31T06:22:41.5019657Z arg2_1 = args[2]; 2023-03-31T06:22:41.5019758Z arg3_1 = args[3]; 2023-03-31T06:22:41.5019820Z 2023-03-31T06:22:41.5019906Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5020001Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5020090Z .findSchemaOrThrow( 2023-03-31T06:22:41.5020188Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5020255Z "") 2023-03-31T06:22:41.5020322Z .typed< 2023-03-31T06:22:41.5020398Z at::Tensor( 2023-03-31T06:22:41.5020483Z const at::Tensor& input_t, 2023-03-31T06:22:41.5020586Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5020704Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5020793Z c10::string_view attr, 2023-03-31T06:22:41.5020913Z torch::List> scalars, 2023-03-31T06:22:41.5021033Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5021098Z 2023-03-31T06:22:41.5021251Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.5021312Z arg1_1.reset(); 2023-03-31T06:22:41.5021385Z arg2_1.reset(); 2023-03-31T06:22:41.5021455Z arg3_1.reset(); 2023-03-31T06:22:41.5021559Z return std::vector({buf0}); 2023-03-31T06:22:41.5021623Z } 2023-03-31T06:22:41.5021705Z ''' 2023-03-31T06:22:41.5021767Z ) 2023-03-31T06:22:41.5021773Z 2023-03-31T06:22:41.5021839Z module = load_inline( 2023-03-31T06:22:41.5022134Z name='inline_extension_cchmpo3apapn7nym7i52nb22x4wy2ymilgftavlve4qpxwqdrpqg', 2023-03-31T06:22:41.5022247Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5022471Z functions=['call_117'], 2023-03-31T06:22:41.5022868Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5023026Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5023742Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5023751Z 2023-03-31T06:22:41.5023828Z def _wrap_func(f): 2023-03-31T06:22:41.5023902Z def g(args): 2023-03-31T06:22:41.5023966Z return f(args) 2023-03-31T06:22:41.5024034Z return g 2023-03-31T06:22:41.5024133Z call = _wrap_func(module.call_117) 2023-03-31T06:22:41.5024141Z 2023-03-31T06:22:41.5024146Z 2023-03-31T06:22:41.5024241Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5024361Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5024489Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5024703Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5024902Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5025145Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5025362Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5025495Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5025501Z 2023-03-31T06:22:41.5025506Z 2023-03-31T06:22:41.5025586Z if __name__ == "__main__": 2023-03-31T06:22:41.5025667Z import argparse 2023-03-31T06:22:41.5025800Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5025805Z 2023-03-31T06:22:41.5025915Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5026223Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5026549Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5026868Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5026961Z args = parser.parse_args() 2023-03-31T06:22:41.5026966Z 2023-03-31T06:22:41.5027056Z if args.benchmark_kernels: 2023-03-31T06:22:41.5027245Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5027312Z else: 2023-03-31T06:22:41.5027424Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5027522Z benchmark_compiled_module() 2023-03-31T06:22:41.5027529Z 2023-03-31T06:22:41.5027584Z if p: 2023-03-31T06:22:41.5027735Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5027831Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5027968Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5027976Z 2023-03-31T06:22:41.5028255Z [2023-03-31 06:05:49,103] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 96 2023-03-31T06:22:41.5028525Z [2023-03-31 06:05:49,146] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 97 2023-03-31T06:22:41.5028929Z [2023-03-31 06:06:10,175] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/fz/cfzga23irwaaqkun7uvkrtoepcl3yfbfs5kqfeme63amfgocmocz.py 2023-03-31T06:22:41.5029179Z [2023-03-31 06:06:10,176] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5029215Z 2023-03-31T06:22:41.5029314Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5029374Z import torch 2023-03-31T06:22:41.5029445Z import math 2023-03-31T06:22:41.5029516Z import random 2023-03-31T06:22:41.5029584Z import os 2023-03-31T06:22:41.5029658Z import tempfile 2023-03-31T06:22:41.5029775Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5029783Z 2023-03-31T06:22:41.5029901Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5030014Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5030148Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5030154Z 2023-03-31T06:22:41.5030234Z aten = torch.ops.aten 2023-03-31T06:22:41.5030372Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5030467Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5030472Z 2023-03-31T06:22:41.5030563Z async_compile.wait(globals()) 2023-03-31T06:22:41.5030640Z del async_compile 2023-03-31T06:22:41.5030751Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5030822Z wrapper = ( 2023-03-31T06:22:41.5030904Z ''' 2023-03-31T06:22:41.5030981Z #include 2023-03-31T06:22:41.5031056Z #include 2023-03-31T06:22:41.5031062Z 2023-03-31T06:22:41.5031153Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5031158Z 2023-03-31T06:22:41.5031253Z template 2023-03-31T06:22:41.5031374Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5031448Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5031602Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5031701Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5031825Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5031905Z return kernel_cpp; 2023-03-31T06:22:41.5031968Z } 2023-03-31T06:22:41.5032098Z std::vector call_119(std::vector args) { 2023-03-31T06:22:41.5032181Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5032257Z arg0_1 = args[0]; 2023-03-31T06:22:41.5032330Z arg1_1 = args[1]; 2023-03-31T06:22:41.5032400Z arg2_1 = args[2]; 2023-03-31T06:22:41.5032463Z 2023-03-31T06:22:41.5032560Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5032641Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5032727Z .findSchemaOrThrow( 2023-03-31T06:22:41.5032856Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5032923Z "") 2023-03-31T06:22:41.5032989Z .typed< 2023-03-31T06:22:41.5033067Z at::Tensor( 2023-03-31T06:22:41.5033164Z const at::Tensor& input_t, 2023-03-31T06:22:41.5033254Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5033370Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5033458Z c10::string_view attr, 2023-03-31T06:22:41.5033580Z torch::List> scalars, 2023-03-31T06:22:41.5033702Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5033768Z 2023-03-31T06:22:41.5033929Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.5034002Z arg1_1.reset(); 2023-03-31T06:22:41.5034060Z arg2_1.reset(); 2023-03-31T06:22:41.5034164Z return std::vector({buf0}); 2023-03-31T06:22:41.5034228Z } 2023-03-31T06:22:41.5034310Z ''' 2023-03-31T06:22:41.5034548Z ) 2023-03-31T06:22:41.5034553Z 2023-03-31T06:22:41.5034635Z module = load_inline( 2023-03-31T06:22:41.5034907Z name='inline_extension_cxmtqkjphkfev7ejgp2tmn6rjh6v7wt5i32fvaq375ug5fuqtged', 2023-03-31T06:22:41.5034981Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5035099Z functions=['call_119'], 2023-03-31T06:22:41.5035520Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5035760Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5036482Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5036492Z 2023-03-31T06:22:41.5036568Z def _wrap_func(f): 2023-03-31T06:22:41.5036641Z def g(args): 2023-03-31T06:22:41.5036719Z return f(args) 2023-03-31T06:22:41.5036788Z return g 2023-03-31T06:22:41.5036871Z call = _wrap_func(module.call_119) 2023-03-31T06:22:41.5036876Z 2023-03-31T06:22:41.5036881Z 2023-03-31T06:22:41.5036975Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5037094Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5037224Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5037442Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5037653Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5037871Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5037998Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5038006Z 2023-03-31T06:22:41.5038011Z 2023-03-31T06:22:41.5038077Z if __name__ == "__main__": 2023-03-31T06:22:41.5038152Z import argparse 2023-03-31T06:22:41.5038311Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5038317Z 2023-03-31T06:22:41.5038430Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5038736Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5039077Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5039355Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5039448Z args = parser.parse_args() 2023-03-31T06:22:41.5039454Z 2023-03-31T06:22:41.5039544Z if args.benchmark_kernels: 2023-03-31T06:22:41.5039748Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5039814Z else: 2023-03-31T06:22:41.5039925Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5040023Z benchmark_compiled_module() 2023-03-31T06:22:41.5040029Z 2023-03-31T06:22:41.5040097Z if p: 2023-03-31T06:22:41.5040246Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5040343Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5040470Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5040489Z 2023-03-31T06:22:41.5040756Z [2023-03-31 06:06:10,176] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 97 2023-03-31T06:22:41.5041023Z [2023-03-31 06:06:10,230] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 98 2023-03-31T06:22:41.5041424Z [2023-03-31 06:06:30,991] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/lv/clvom7mg446x2udou5joihf3kroplvahxipikgnffr7pouzu5brp.py 2023-03-31T06:22:41.5041646Z [2023-03-31 06:06:30,991] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5041652Z 2023-03-31T06:22:41.5041749Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5041821Z import torch 2023-03-31T06:22:41.5041892Z import math 2023-03-31T06:22:41.5041963Z import random 2023-03-31T06:22:41.5042018Z import os 2023-03-31T06:22:41.5042092Z import tempfile 2023-03-31T06:22:41.5042210Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5042249Z 2023-03-31T06:22:41.5042392Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5042518Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5042652Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5042658Z 2023-03-31T06:22:41.5042738Z aten = torch.ops.aten 2023-03-31T06:22:41.5042865Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5042962Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5042968Z 2023-03-31T06:22:41.5043200Z async_compile.wait(globals()) 2023-03-31T06:22:41.5043280Z del async_compile 2023-03-31T06:22:41.5043402Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5043477Z wrapper = ( 2023-03-31T06:22:41.5043564Z ''' 2023-03-31T06:22:41.5043627Z #include 2023-03-31T06:22:41.5043704Z #include 2023-03-31T06:22:41.5043709Z 2023-03-31T06:22:41.5043803Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5043808Z 2023-03-31T06:22:41.5043903Z template 2023-03-31T06:22:41.5044029Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5044119Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5044246Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5044344Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5044457Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5044539Z return kernel_cpp; 2023-03-31T06:22:41.5044603Z } 2023-03-31T06:22:41.5044734Z std::vector call_121(std::vector args) { 2023-03-31T06:22:41.5044896Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5044972Z arg0_1 = args[0]; 2023-03-31T06:22:41.5045046Z arg1_1 = args[1]; 2023-03-31T06:22:41.5045104Z arg2_1 = args[2]; 2023-03-31T06:22:41.5045177Z arg3_1 = args[3]; 2023-03-31T06:22:41.5045241Z 2023-03-31T06:22:41.5045340Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5045438Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5045528Z .findSchemaOrThrow( 2023-03-31T06:22:41.5045628Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5045682Z "") 2023-03-31T06:22:41.5045751Z .typed< 2023-03-31T06:22:41.5045828Z at::Tensor( 2023-03-31T06:22:41.5045930Z const at::Tensor& input_t, 2023-03-31T06:22:41.5046032Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5046186Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5046276Z c10::string_view attr, 2023-03-31T06:22:41.5046386Z torch::List> scalars, 2023-03-31T06:22:41.5046512Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5046580Z 2023-03-31T06:22:41.5046733Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.5046808Z arg1_1.reset(); 2023-03-31T06:22:41.5046883Z arg2_1.reset(); 2023-03-31T06:22:41.5046957Z arg3_1.reset(); 2023-03-31T06:22:41.5047051Z return std::vector({buf0}); 2023-03-31T06:22:41.5047115Z } 2023-03-31T06:22:41.5047201Z ''' 2023-03-31T06:22:41.5047265Z ) 2023-03-31T06:22:41.5047270Z 2023-03-31T06:22:41.5047349Z module = load_inline( 2023-03-31T06:22:41.5047617Z name='inline_extension_cchmpo3apapn7nym7i52nb22x4wy2ymilgftavlve4qpxwqdrpqg', 2023-03-31T06:22:41.5047707Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5047814Z functions=['call_121'], 2023-03-31T06:22:41.5048199Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5048356Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5049100Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5049139Z 2023-03-31T06:22:41.5049217Z def _wrap_func(f): 2023-03-31T06:22:41.5049288Z def g(args): 2023-03-31T06:22:41.5049365Z return f(args) 2023-03-31T06:22:41.5049435Z return g 2023-03-31T06:22:41.5049535Z call = _wrap_func(module.call_121) 2023-03-31T06:22:41.5049541Z 2023-03-31T06:22:41.5049546Z 2023-03-31T06:22:41.5049629Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5049752Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5049882Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5050099Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5050307Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5050519Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5050729Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5050866Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5050871Z 2023-03-31T06:22:41.5050876Z 2023-03-31T06:22:41.5050941Z if __name__ == "__main__": 2023-03-31T06:22:41.5051018Z import argparse 2023-03-31T06:22:41.5051152Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5051158Z 2023-03-31T06:22:41.5051270Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5051623Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5051964Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5052244Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5052340Z args = parser.parse_args() 2023-03-31T06:22:41.5052346Z 2023-03-31T06:22:41.5052438Z if args.benchmark_kernels: 2023-03-31T06:22:41.5052616Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5052685Z else: 2023-03-31T06:22:41.5052800Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5052925Z benchmark_compiled_module() 2023-03-31T06:22:41.5052931Z 2023-03-31T06:22:41.5053002Z if p: 2023-03-31T06:22:41.5053155Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5053251Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5053376Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5053393Z 2023-03-31T06:22:41.5053659Z [2023-03-31 06:06:30,992] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 98 2023-03-31T06:22:41.5053929Z [2023-03-31 06:06:31,034] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 99 2023-03-31T06:22:41.5054334Z [2023-03-31 06:06:52,313] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/nb/cnb4jh64qiie4hoyhprnmqp5xarivycdp54kmyluc5sxkfu6opku.py 2023-03-31T06:22:41.5054556Z [2023-03-31 06:06:52,313] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5054565Z 2023-03-31T06:22:41.5054662Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5054734Z import torch 2023-03-31T06:22:41.5054805Z import math 2023-03-31T06:22:41.5054882Z import random 2023-03-31T06:22:41.5054939Z import os 2023-03-31T06:22:41.5055012Z import tempfile 2023-03-31T06:22:41.5055129Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5055135Z 2023-03-31T06:22:41.5055256Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5055381Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5055551Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5055583Z 2023-03-31T06:22:41.5055664Z aten = torch.ops.aten 2023-03-31T06:22:41.5055791Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5055886Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5055891Z 2023-03-31T06:22:41.5055981Z async_compile.wait(globals()) 2023-03-31T06:22:41.5056058Z del async_compile 2023-03-31T06:22:41.5056178Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5056251Z wrapper = ( 2023-03-31T06:22:41.5056333Z ''' 2023-03-31T06:22:41.5056398Z #include 2023-03-31T06:22:41.5056473Z #include 2023-03-31T06:22:41.5056478Z 2023-03-31T06:22:41.5056570Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5056575Z 2023-03-31T06:22:41.5056666Z template 2023-03-31T06:22:41.5056789Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5056874Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5057000Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5057098Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5057211Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5057289Z return kernel_cpp; 2023-03-31T06:22:41.5057353Z } 2023-03-31T06:22:41.5057483Z std::vector call_123(std::vector args) { 2023-03-31T06:22:41.5057579Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5057652Z arg0_1 = args[0]; 2023-03-31T06:22:41.5057725Z arg1_1 = args[1]; 2023-03-31T06:22:41.5057811Z arg2_1 = args[2]; 2023-03-31T06:22:41.5057874Z 2023-03-31T06:22:41.5057972Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5058066Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5058154Z .findSchemaOrThrow( 2023-03-31T06:22:41.5058253Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5058319Z "") 2023-03-31T06:22:41.5058375Z .typed< 2023-03-31T06:22:41.5058451Z at::Tensor( 2023-03-31T06:22:41.5058550Z const at::Tensor& input_t, 2023-03-31T06:22:41.5058650Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5058766Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5058854Z c10::string_view attr, 2023-03-31T06:22:41.5058975Z torch::List> scalars, 2023-03-31T06:22:41.5059127Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5059192Z 2023-03-31T06:22:41.5059356Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardtanh", {0, 6}, ""); 2023-03-31T06:22:41.5059429Z arg1_1.reset(); 2023-03-31T06:22:41.5059500Z arg2_1.reset(); 2023-03-31T06:22:41.5059604Z return std::vector({buf0}); 2023-03-31T06:22:41.5059667Z } 2023-03-31T06:22:41.5059736Z ''' 2023-03-31T06:22:41.5059796Z ) 2023-03-31T06:22:41.5059802Z 2023-03-31T06:22:41.5059883Z module = load_inline( 2023-03-31T06:22:41.5060154Z name='inline_extension_cxmtqkjphkfev7ejgp2tmn6rjh6v7wt5i32fvaq375ug5fuqtged', 2023-03-31T06:22:41.5060239Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5060356Z functions=['call_123'], 2023-03-31T06:22:41.5060736Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5060894Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5061603Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5061623Z 2023-03-31T06:22:41.5061687Z def _wrap_func(f): 2023-03-31T06:22:41.5061758Z def g(args): 2023-03-31T06:22:41.5061879Z return f(args) 2023-03-31T06:22:41.5061982Z return g 2023-03-31T06:22:41.5062078Z call = _wrap_func(module.call_123) 2023-03-31T06:22:41.5062083Z 2023-03-31T06:22:41.5062089Z 2023-03-31T06:22:41.5062183Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5062302Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5062525Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5062737Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5062949Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5063156Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5063283Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5063289Z 2023-03-31T06:22:41.5063293Z 2023-03-31T06:22:41.5063371Z if __name__ == "__main__": 2023-03-31T06:22:41.5063447Z import argparse 2023-03-31T06:22:41.5063581Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5063589Z 2023-03-31T06:22:41.5063699Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5063996Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5064333Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5064646Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5064741Z args = parser.parse_args() 2023-03-31T06:22:41.5064746Z 2023-03-31T06:22:41.5064837Z if args.benchmark_kernels: 2023-03-31T06:22:41.5065031Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5065099Z else: 2023-03-31T06:22:41.5065212Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5065297Z benchmark_compiled_module() 2023-03-31T06:22:41.5065305Z 2023-03-31T06:22:41.5065377Z if p: 2023-03-31T06:22:41.5065527Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5065623Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5065759Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5065793Z 2023-03-31T06:22:41.5066073Z [2023-03-31 06:06:52,314] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 99 2023-03-31T06:22:41.5066343Z [2023-03-31 06:06:52,373] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 100 2023-03-31T06:22:41.5066748Z [2023-03-31 06:07:13,437] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/b3/cb3qcmcg2w3wmkbvyjzrehpvby365hdlr7gldwvgdakardlamzeo.py 2023-03-31T06:22:41.5066970Z [2023-03-31 06:07:13,437] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5066977Z 2023-03-31T06:22:41.5067062Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5067134Z import torch 2023-03-31T06:22:41.5067203Z import math 2023-03-31T06:22:41.5067275Z import random 2023-03-31T06:22:41.5067344Z import os 2023-03-31T06:22:41.5067417Z import tempfile 2023-03-31T06:22:41.5067535Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5067543Z 2023-03-31T06:22:41.5067649Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5067776Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5067911Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5067916Z 2023-03-31T06:22:41.5067999Z aten = torch.ops.aten 2023-03-31T06:22:41.5068138Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5068232Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5068237Z 2023-03-31T06:22:41.5068328Z async_compile.wait(globals()) 2023-03-31T06:22:41.5068431Z del async_compile 2023-03-31T06:22:41.5068568Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5068639Z wrapper = ( 2023-03-31T06:22:41.5068725Z ''' 2023-03-31T06:22:41.5068800Z #include 2023-03-31T06:22:41.5068877Z #include 2023-03-31T06:22:41.5068882Z 2023-03-31T06:22:41.5068976Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5068981Z 2023-03-31T06:22:41.5069073Z template 2023-03-31T06:22:41.5069186Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5069271Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5069398Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5069495Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5069623Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5069702Z return kernel_cpp; 2023-03-31T06:22:41.5069767Z } 2023-03-31T06:22:41.5069886Z std::vector call_124(std::vector args) { 2023-03-31T06:22:41.5069993Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5070069Z arg0_1 = args[0]; 2023-03-31T06:22:41.5070140Z arg1_1 = args[1]; 2023-03-31T06:22:41.5070210Z arg2_1 = args[2]; 2023-03-31T06:22:41.5070282Z arg3_1 = args[3]; 2023-03-31T06:22:41.5070345Z 2023-03-31T06:22:41.5070430Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5070523Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5070613Z .findSchemaOrThrow( 2023-03-31T06:22:41.5070710Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5070776Z "") 2023-03-31T06:22:41.5070878Z .typed< 2023-03-31T06:22:41.5070954Z at::Tensor( 2023-03-31T06:22:41.5071039Z const at::Tensor& input_t, 2023-03-31T06:22:41.5071139Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5071258Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5071347Z c10::string_view attr, 2023-03-31T06:22:41.5071472Z torch::List> scalars, 2023-03-31T06:22:41.5071594Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5071662Z 2023-03-31T06:22:41.5071874Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "swish", {-1}, ""); 2023-03-31T06:22:41.5071947Z arg1_1.reset(); 2023-03-31T06:22:41.5072017Z arg2_1.reset(); 2023-03-31T06:22:41.5072115Z arg3_1.reset(); 2023-03-31T06:22:41.5072219Z return std::vector({buf0}); 2023-03-31T06:22:41.5072282Z } 2023-03-31T06:22:41.5072362Z ''' 2023-03-31T06:22:41.5072414Z ) 2023-03-31T06:22:41.5072420Z 2023-03-31T06:22:41.5072498Z module = load_inline( 2023-03-31T06:22:41.5072766Z name='inline_extension_co2j6sdzmzk3o6xkazchr6wfthup3h4hy5pmjbkj5ysafin6kr5b', 2023-03-31T06:22:41.5072853Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5072969Z functions=['call_124'], 2023-03-31T06:22:41.5073355Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5073514Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5074226Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5074235Z 2023-03-31T06:22:41.5074314Z def _wrap_func(f): 2023-03-31T06:22:41.5074373Z def g(args): 2023-03-31T06:22:41.5074448Z return f(args) 2023-03-31T06:22:41.5074514Z return g 2023-03-31T06:22:41.5074609Z call = _wrap_func(module.call_124) 2023-03-31T06:22:41.5074615Z 2023-03-31T06:22:41.5074620Z 2023-03-31T06:22:41.5074714Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5074866Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5075019Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5075223Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5075433Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5075642Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5075861Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5075997Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5076003Z 2023-03-31T06:22:41.5076007Z 2023-03-31T06:22:41.5076085Z if __name__ == "__main__": 2023-03-31T06:22:41.5076161Z import argparse 2023-03-31T06:22:41.5076290Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5076296Z 2023-03-31T06:22:41.5076406Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5076709Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5077051Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5077330Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5077428Z args = parser.parse_args() 2023-03-31T06:22:41.5077433Z 2023-03-31T06:22:41.5077522Z if args.benchmark_kernels: 2023-03-31T06:22:41.5077747Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5077815Z else: 2023-03-31T06:22:41.5077927Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5078009Z benchmark_compiled_module() 2023-03-31T06:22:41.5078015Z 2023-03-31T06:22:41.5078083Z if p: 2023-03-31T06:22:41.5078233Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5078331Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5078467Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5078473Z 2023-03-31T06:22:41.5078752Z [2023-03-31 06:07:13,438] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 100 2023-03-31T06:22:41.5079022Z [2023-03-31 06:07:13,483] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 101 2023-03-31T06:22:41.5079455Z [2023-03-31 06:07:34,655] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/wu/cwuystiaw6pmujd6wb5zrkzxin7stn2lep3jszdakr2two4vfyc6.py 2023-03-31T06:22:41.5079675Z [2023-03-31 06:07:34,655] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5079681Z 2023-03-31T06:22:41.5079766Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5079837Z import torch 2023-03-31T06:22:41.5079908Z import math 2023-03-31T06:22:41.5079982Z import random 2023-03-31T06:22:41.5080052Z import os 2023-03-31T06:22:41.5080127Z import tempfile 2023-03-31T06:22:41.5080245Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5080250Z 2023-03-31T06:22:41.5080356Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5080482Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5080615Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5080623Z 2023-03-31T06:22:41.5080702Z aten = torch.ops.aten 2023-03-31T06:22:41.5080842Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5080938Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5080943Z 2023-03-31T06:22:41.5081035Z async_compile.wait(globals()) 2023-03-31T06:22:41.5081108Z del async_compile 2023-03-31T06:22:41.5081216Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5081288Z wrapper = ( 2023-03-31T06:22:41.5081370Z ''' 2023-03-31T06:22:41.5081471Z #include 2023-03-31T06:22:41.5081584Z #include 2023-03-31T06:22:41.5081590Z 2023-03-31T06:22:41.5081681Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5081686Z 2023-03-31T06:22:41.5081777Z template 2023-03-31T06:22:41.5081890Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5081981Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5082106Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5082205Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5082330Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5082409Z return kernel_cpp; 2023-03-31T06:22:41.5082471Z } 2023-03-31T06:22:41.5082588Z std::vector call_125(std::vector args) { 2023-03-31T06:22:41.5082679Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5082752Z arg0_1 = args[0]; 2023-03-31T06:22:41.5082830Z arg1_1 = args[1]; 2023-03-31T06:22:41.5082903Z arg2_1 = args[2]; 2023-03-31T06:22:41.5082967Z 2023-03-31T06:22:41.5083244Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5083326Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5083414Z .findSchemaOrThrow( 2023-03-31T06:22:41.5083514Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5083582Z "") 2023-03-31T06:22:41.5083651Z .typed< 2023-03-31T06:22:41.5083730Z at::Tensor( 2023-03-31T06:22:41.5083830Z const at::Tensor& input_t, 2023-03-31T06:22:41.5083917Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5084093Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5084184Z c10::string_view attr, 2023-03-31T06:22:41.5084307Z torch::List> scalars, 2023-03-31T06:22:41.5084429Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5084497Z 2023-03-31T06:22:41.5084750Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "swish", {-1}, ""); 2023-03-31T06:22:41.5084827Z arg1_1.reset(); 2023-03-31T06:22:41.5084885Z arg2_1.reset(); 2023-03-31T06:22:41.5084994Z return std::vector({buf0}); 2023-03-31T06:22:41.5085059Z } 2023-03-31T06:22:41.5085141Z ''' 2023-03-31T06:22:41.5085204Z ) 2023-03-31T06:22:41.5085210Z 2023-03-31T06:22:41.5085289Z module = load_inline( 2023-03-31T06:22:41.5085601Z name='inline_extension_cqvj3c7evbaptsjxt2y7hhph6qcdvq6y4ukqluourdvjmnmle7j7', 2023-03-31T06:22:41.5085674Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5085794Z functions=['call_125'], 2023-03-31T06:22:41.5086174Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5086331Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5087041Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5087050Z 2023-03-31T06:22:41.5087129Z def _wrap_func(f): 2023-03-31T06:22:41.5087203Z def g(args): 2023-03-31T06:22:41.5087284Z return f(args) 2023-03-31T06:22:41.5087339Z return g 2023-03-31T06:22:41.5087438Z call = _wrap_func(module.call_125) 2023-03-31T06:22:41.5087443Z 2023-03-31T06:22:41.5087448Z 2023-03-31T06:22:41.5087548Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5087668Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5087797Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5088012Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5088304Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5088557Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5088672Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5088692Z 2023-03-31T06:22:41.5088697Z 2023-03-31T06:22:41.5088761Z if __name__ == "__main__": 2023-03-31T06:22:41.5088843Z import argparse 2023-03-31T06:22:41.5088977Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5088982Z 2023-03-31T06:22:41.5089093Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5089408Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5089749Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5090028Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5090125Z args = parser.parse_args() 2023-03-31T06:22:41.5090130Z 2023-03-31T06:22:41.5090208Z if args.benchmark_kernels: 2023-03-31T06:22:41.5090402Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5090474Z else: 2023-03-31T06:22:41.5090587Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5090689Z benchmark_compiled_module() 2023-03-31T06:22:41.5090694Z 2023-03-31T06:22:41.5090765Z if p: 2023-03-31T06:22:41.5090944Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5091044Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5091168Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5091174Z 2023-03-31T06:22:41.5091457Z [2023-03-31 06:07:34,655] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 101 2023-03-31T06:22:41.5091732Z [2023-03-31 06:07:34,713] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 102 2023-03-31T06:22:41.5092140Z [2023-03-31 06:07:55,660] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/77/c77quysssmlqpl6duv3gp5t5swlihnlwsz2rsrwih54ypr35thnj.py 2023-03-31T06:22:41.5092360Z [2023-03-31 06:07:55,660] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5092394Z 2023-03-31T06:22:41.5092494Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5092567Z import torch 2023-03-31T06:22:41.5092637Z import math 2023-03-31T06:22:41.5092699Z import random 2023-03-31T06:22:41.5092766Z import os 2023-03-31T06:22:41.5092841Z import tempfile 2023-03-31T06:22:41.5092958Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5092964Z 2023-03-31T06:22:41.5093082Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5093208Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5093341Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5093349Z 2023-03-31T06:22:41.5093429Z aten = torch.ops.aten 2023-03-31T06:22:41.5093556Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5093651Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5093656Z 2023-03-31T06:22:41.5093747Z async_compile.wait(globals()) 2023-03-31T06:22:41.5093825Z del async_compile 2023-03-31T06:22:41.5093951Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5094026Z wrapper = ( 2023-03-31T06:22:41.5094109Z ''' 2023-03-31T06:22:41.5094174Z #include 2023-03-31T06:22:41.5094250Z #include 2023-03-31T06:22:41.5094255Z 2023-03-31T06:22:41.5094346Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5094351Z 2023-03-31T06:22:41.5094443Z template 2023-03-31T06:22:41.5094566Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5094654Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5094812Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5094922Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5095047Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5095126Z return kernel_cpp; 2023-03-31T06:22:41.5095189Z } 2023-03-31T06:22:41.5095318Z std::vector call_126(std::vector args) { 2023-03-31T06:22:41.5095424Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5095500Z arg0_1 = args[0]; 2023-03-31T06:22:41.5095561Z arg1_1 = args[1]; 2023-03-31T06:22:41.5095636Z arg2_1 = args[2]; 2023-03-31T06:22:41.5095707Z arg3_1 = args[3]; 2023-03-31T06:22:41.5095770Z 2023-03-31T06:22:41.5095869Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5095960Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5096047Z .findSchemaOrThrow( 2023-03-31T06:22:41.5096133Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5096198Z "") 2023-03-31T06:22:41.5096269Z .typed< 2023-03-31T06:22:41.5096346Z at::Tensor( 2023-03-31T06:22:41.5096443Z const at::Tensor& input_t, 2023-03-31T06:22:41.5096542Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5096658Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5096734Z c10::string_view attr, 2023-03-31T06:22:41.5096854Z torch::List> scalars, 2023-03-31T06:22:41.5096979Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5097045Z 2023-03-31T06:22:41.5097300Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "swish", {-1}, ""); 2023-03-31T06:22:41.5097374Z arg1_1.reset(); 2023-03-31T06:22:41.5097445Z arg2_1.reset(); 2023-03-31T06:22:41.5097502Z arg3_1.reset(); 2023-03-31T06:22:41.5097608Z return std::vector({buf0}); 2023-03-31T06:22:41.5097671Z } 2023-03-31T06:22:41.5099554Z ''' 2023-03-31T06:22:41.5099619Z ) 2023-03-31T06:22:41.5099627Z 2023-03-31T06:22:41.5099707Z module = load_inline( 2023-03-31T06:22:41.5099982Z name='inline_extension_co2j6sdzmzk3o6xkazchr6wfthup3h4hy5pmjbkj5ysafin6kr5b', 2023-03-31T06:22:41.5100069Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5100174Z functions=['call_126'], 2023-03-31T06:22:41.5100556Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5100765Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5101487Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5101494Z 2023-03-31T06:22:41.5101573Z def _wrap_func(f): 2023-03-31T06:22:41.5101646Z def g(args): 2023-03-31T06:22:41.5101723Z return f(args) 2023-03-31T06:22:41.5101790Z return g 2023-03-31T06:22:41.5101873Z call = _wrap_func(module.call_126) 2023-03-31T06:22:41.5101892Z 2023-03-31T06:22:41.5101897Z 2023-03-31T06:22:41.5101979Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5102099Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5102231Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5102549Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5102766Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5102976Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5103183Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5103354Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5103389Z 2023-03-31T06:22:41.5103394Z 2023-03-31T06:22:41.5103461Z if __name__ == "__main__": 2023-03-31T06:22:41.5103539Z import argparse 2023-03-31T06:22:41.5103668Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5103674Z 2023-03-31T06:22:41.5103781Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5104092Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5104437Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5104716Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5104809Z args = parser.parse_args() 2023-03-31T06:22:41.5104814Z 2023-03-31T06:22:41.5104903Z if args.benchmark_kernels: 2023-03-31T06:22:41.5105082Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5105151Z else: 2023-03-31T06:22:41.5105265Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5105361Z benchmark_compiled_module() 2023-03-31T06:22:41.5105367Z 2023-03-31T06:22:41.5105436Z if p: 2023-03-31T06:22:41.5105585Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5105683Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5105807Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5105841Z 2023-03-31T06:22:41.5106123Z [2023-03-31 06:07:55,661] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 102 2023-03-31T06:22:41.5106389Z [2023-03-31 06:07:55,706] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 103 2023-03-31T06:22:41.5106790Z [2023-03-31 06:08:16,898] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/46/c46ad272nkisfh4t7ticcbohbj7xmruomkmvmdhdbtajn2e3mysd.py 2023-03-31T06:22:41.5107012Z [2023-03-31 06:08:16,899] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5107018Z 2023-03-31T06:22:41.5107115Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5107186Z import torch 2023-03-31T06:22:41.5107257Z import math 2023-03-31T06:22:41.5107363Z import random 2023-03-31T06:22:41.5107431Z import os 2023-03-31T06:22:41.5107505Z import tempfile 2023-03-31T06:22:41.5107623Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5107632Z 2023-03-31T06:22:41.5107752Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5107877Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5108014Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5108019Z 2023-03-31T06:22:41.5108099Z aten = torch.ops.aten 2023-03-31T06:22:41.5108229Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5108326Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5108331Z 2023-03-31T06:22:41.5108421Z async_compile.wait(globals()) 2023-03-31T06:22:41.5108496Z del async_compile 2023-03-31T06:22:41.5108617Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5108689Z wrapper = ( 2023-03-31T06:22:41.5108772Z ''' 2023-03-31T06:22:41.5108836Z #include 2023-03-31T06:22:41.5108914Z #include 2023-03-31T06:22:41.5108920Z 2023-03-31T06:22:41.5109016Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5109022Z 2023-03-31T06:22:41.5109116Z template 2023-03-31T06:22:41.5109238Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5109325Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5109448Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5109533Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5109687Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5109793Z return kernel_cpp; 2023-03-31T06:22:41.5109856Z } 2023-03-31T06:22:41.5109987Z std::vector call_127(std::vector args) { 2023-03-31T06:22:41.5110079Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5110151Z arg0_1 = args[0]; 2023-03-31T06:22:41.5110210Z arg1_1 = args[1]; 2023-03-31T06:22:41.5110281Z arg2_1 = args[2]; 2023-03-31T06:22:41.5110345Z 2023-03-31T06:22:41.5110442Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5110535Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5110625Z .findSchemaOrThrow( 2023-03-31T06:22:41.5110722Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5110776Z "") 2023-03-31T06:22:41.5110843Z .typed< 2023-03-31T06:22:41.5110917Z at::Tensor( 2023-03-31T06:22:41.5111016Z const at::Tensor& input_t, 2023-03-31T06:22:41.5111120Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5111241Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5111331Z c10::string_view attr, 2023-03-31T06:22:41.5111451Z torch::List> scalars, 2023-03-31T06:22:41.5111560Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5111626Z 2023-03-31T06:22:41.5111864Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "swish", {-1}, ""); 2023-03-31T06:22:41.5111940Z arg1_1.reset(); 2023-03-31T06:22:41.5112011Z arg2_1.reset(); 2023-03-31T06:22:41.5112143Z return std::vector({buf0}); 2023-03-31T06:22:41.5112206Z } 2023-03-31T06:22:41.5112274Z ''' 2023-03-31T06:22:41.5112335Z ) 2023-03-31T06:22:41.5112341Z 2023-03-31T06:22:41.5112417Z module = load_inline( 2023-03-31T06:22:41.5112687Z name='inline_extension_cqvj3c7evbaptsjxt2y7hhph6qcdvq6y4ukqluourdvjmnmle7j7', 2023-03-31T06:22:41.5112773Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5112892Z functions=['call_127'], 2023-03-31T06:22:41.5113274Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5113418Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5114127Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5114177Z 2023-03-31T06:22:41.5114242Z def _wrap_func(f): 2023-03-31T06:22:41.5114314Z def g(args): 2023-03-31T06:22:41.5114389Z return f(args) 2023-03-31T06:22:41.5114455Z return g 2023-03-31T06:22:41.5114553Z call = _wrap_func(module.call_127) 2023-03-31T06:22:41.5114559Z 2023-03-31T06:22:41.5114563Z 2023-03-31T06:22:41.5114661Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5114782Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5114897Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5115111Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5115321Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5115531Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5115659Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5115664Z 2023-03-31T06:22:41.5115669Z 2023-03-31T06:22:41.5115746Z if __name__ == "__main__": 2023-03-31T06:22:41.5115823Z import argparse 2023-03-31T06:22:41.5115952Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5115957Z 2023-03-31T06:22:41.5116067Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5116392Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5116764Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5117040Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5117136Z args = parser.parse_args() 2023-03-31T06:22:41.5117141Z 2023-03-31T06:22:41.5117230Z if args.benchmark_kernels: 2023-03-31T06:22:41.5117422Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5117489Z else: 2023-03-31T06:22:41.5117600Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5117682Z benchmark_compiled_module() 2023-03-31T06:22:41.5117687Z 2023-03-31T06:22:41.5117754Z if p: 2023-03-31T06:22:41.5117904Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5118002Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5118139Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5118144Z 2023-03-31T06:22:41.5118426Z [2023-03-31 06:08:16,899] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 103 2023-03-31T06:22:41.5118699Z [2023-03-31 06:08:16,959] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 104 2023-03-31T06:22:41.5119143Z [2023-03-31 06:08:37,550] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/i6/ci6ci4ih6elckk2ifrfvp5dmwy7pberahxm27fumslhbnhb7qgmb.py 2023-03-31T06:22:41.5119350Z [2023-03-31 06:08:37,550] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5119368Z 2023-03-31T06:22:41.5119452Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5119525Z import torch 2023-03-31T06:22:41.5119596Z import math 2023-03-31T06:22:41.5119670Z import random 2023-03-31T06:22:41.5119740Z import os 2023-03-31T06:22:41.5119814Z import tempfile 2023-03-31T06:22:41.5119920Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5119937Z 2023-03-31T06:22:41.5120043Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5120168Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5120299Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5120333Z 2023-03-31T06:22:41.5120416Z aten = torch.ops.aten 2023-03-31T06:22:41.5120557Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5120652Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5120658Z 2023-03-31T06:22:41.5120749Z async_compile.wait(globals()) 2023-03-31T06:22:41.5120811Z del async_compile 2023-03-31T06:22:41.5120932Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5121004Z wrapper = ( 2023-03-31T06:22:41.5121086Z ''' 2023-03-31T06:22:41.5121162Z #include 2023-03-31T06:22:41.5121242Z #include 2023-03-31T06:22:41.5121249Z 2023-03-31T06:22:41.5121343Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5121348Z 2023-03-31T06:22:41.5121439Z template 2023-03-31T06:22:41.5121549Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5121638Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5121763Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5121864Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5121991Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5122074Z return kernel_cpp; 2023-03-31T06:22:41.5122139Z } 2023-03-31T06:22:41.5122256Z std::vector call_128(std::vector args) { 2023-03-31T06:22:41.5122362Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5122438Z arg0_1 = args[0]; 2023-03-31T06:22:41.5122513Z arg1_1 = args[1]; 2023-03-31T06:22:41.5122585Z arg2_1 = args[2]; 2023-03-31T06:22:41.5122691Z arg3_1 = args[3]; 2023-03-31T06:22:41.5122768Z 2023-03-31T06:22:41.5122867Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5122962Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5123206Z .findSchemaOrThrow( 2023-03-31T06:22:41.5123312Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5123384Z "") 2023-03-31T06:22:41.5123454Z .typed< 2023-03-31T06:22:41.5123519Z at::Tensor( 2023-03-31T06:22:41.5123618Z const at::Tensor& input_t, 2023-03-31T06:22:41.5123723Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5123841Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5123931Z c10::string_view attr, 2023-03-31T06:22:41.5124054Z torch::List> scalars, 2023-03-31T06:22:41.5124176Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5124243Z 2023-03-31T06:22:41.5124469Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.5124545Z arg1_1.reset(); 2023-03-31T06:22:41.5124617Z arg2_1.reset(); 2023-03-31T06:22:41.5124688Z arg3_1.reset(); 2023-03-31T06:22:41.5124796Z return std::vector({buf0}); 2023-03-31T06:22:41.5124860Z } 2023-03-31T06:22:41.5124942Z ''' 2023-03-31T06:22:41.5124993Z ) 2023-03-31T06:22:41.5124999Z 2023-03-31T06:22:41.5125078Z module = load_inline( 2023-03-31T06:22:41.5125344Z name='inline_extension_cet5mw6ckpn5sdinqb23taoug7nnqbjxmhb6i27zj645y6mt3saj', 2023-03-31T06:22:41.5125498Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5125619Z functions=['call_128'], 2023-03-31T06:22:41.5126003Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5126160Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5126873Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5126882Z 2023-03-31T06:22:41.5126996Z def _wrap_func(f): 2023-03-31T06:22:41.5127055Z def g(args): 2023-03-31T06:22:41.5127131Z return f(args) 2023-03-31T06:22:41.5127199Z return g 2023-03-31T06:22:41.5127299Z call = _wrap_func(module.call_128) 2023-03-31T06:22:41.5127306Z 2023-03-31T06:22:41.5127310Z 2023-03-31T06:22:41.5127406Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5127525Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5127655Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5127860Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5128069Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5128279Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5128493Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5128631Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5128640Z 2023-03-31T06:22:41.5128644Z 2023-03-31T06:22:41.5128723Z if __name__ == "__main__": 2023-03-31T06:22:41.5128801Z import argparse 2023-03-31T06:22:41.5128932Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5128938Z 2023-03-31T06:22:41.5129049Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5129349Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5129723Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5130053Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5130145Z args = parser.parse_args() 2023-03-31T06:22:41.5130150Z 2023-03-31T06:22:41.5130240Z if args.benchmark_kernels: 2023-03-31T06:22:41.5130430Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5130498Z else: 2023-03-31T06:22:41.5130610Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5130695Z benchmark_compiled_module() 2023-03-31T06:22:41.5130701Z 2023-03-31T06:22:41.5130771Z if p: 2023-03-31T06:22:41.5130919Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5131014Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5131151Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5131159Z 2023-03-31T06:22:41.5131442Z [2023-03-31 06:08:37,550] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 104 2023-03-31T06:22:41.5131711Z [2023-03-31 06:08:37,599] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 105 2023-03-31T06:22:41.5132124Z [2023-03-31 06:08:58,357] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/hq/chquo5nogmty7igklsihijwhp5tyc6vbsfpugrdrgd3chlupclt3.py 2023-03-31T06:22:41.5132332Z [2023-03-31 06:08:58,357] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5132381Z 2023-03-31T06:22:41.5132468Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5132539Z import torch 2023-03-31T06:22:41.5132610Z import math 2023-03-31T06:22:41.5132682Z import random 2023-03-31T06:22:41.5132749Z import os 2023-03-31T06:22:41.5132824Z import tempfile 2023-03-31T06:22:41.5132928Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5132951Z 2023-03-31T06:22:41.5133061Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5133185Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5133317Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5133323Z 2023-03-31T06:22:41.5133404Z aten = torch.ops.aten 2023-03-31T06:22:41.5133542Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5133666Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5133672Z 2023-03-31T06:22:41.5133765Z async_compile.wait(globals()) 2023-03-31T06:22:41.5133829Z del async_compile 2023-03-31T06:22:41.5133950Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5134021Z wrapper = ( 2023-03-31T06:22:41.5134104Z ''' 2023-03-31T06:22:41.5134179Z #include 2023-03-31T06:22:41.5134255Z #include 2023-03-31T06:22:41.5134260Z 2023-03-31T06:22:41.5134351Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5134356Z 2023-03-31T06:22:41.5134451Z template 2023-03-31T06:22:41.5134563Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5134648Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5134772Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5134868Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5134993Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5135074Z return kernel_cpp; 2023-03-31T06:22:41.5135136Z } 2023-03-31T06:22:41.5135252Z std::vector call_129(std::vector args) { 2023-03-31T06:22:41.5135349Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5135423Z arg0_1 = args[0]; 2023-03-31T06:22:41.5135494Z arg1_1 = args[1]; 2023-03-31T06:22:41.5135565Z arg2_1 = args[2]; 2023-03-31T06:22:41.5135630Z 2023-03-31T06:22:41.5135713Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5135805Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5135921Z .findSchemaOrThrow( 2023-03-31T06:22:41.5136045Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5136112Z "") 2023-03-31T06:22:41.5136179Z .typed< 2023-03-31T06:22:41.5136253Z at::Tensor( 2023-03-31T06:22:41.5136338Z const at::Tensor& input_t, 2023-03-31T06:22:41.5136437Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5136554Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5136646Z c10::string_view attr, 2023-03-31T06:22:41.5136766Z torch::List> scalars, 2023-03-31T06:22:41.5136887Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5136954Z 2023-03-31T06:22:41.5137206Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.5137267Z arg1_1.reset(); 2023-03-31T06:22:41.5137337Z arg2_1.reset(); 2023-03-31T06:22:41.5137444Z return std::vector({buf0}); 2023-03-31T06:22:41.5137511Z } 2023-03-31T06:22:41.5137591Z ''' 2023-03-31T06:22:41.5137658Z ) 2023-03-31T06:22:41.5137663Z 2023-03-31T06:22:41.5137742Z module = load_inline( 2023-03-31T06:22:41.5137997Z name='inline_extension_ceyanris47ohp6ch643vxttxgfgln7ral5h3xnryxnbufbtgql2n', 2023-03-31T06:22:41.5138082Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5138200Z functions=['call_129'], 2023-03-31T06:22:41.5138579Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5138764Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5139472Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5139481Z 2023-03-31T06:22:41.5139556Z def _wrap_func(f): 2023-03-31T06:22:41.5139626Z def g(args): 2023-03-31T06:22:41.5139699Z return f(args) 2023-03-31T06:22:41.5139754Z return g 2023-03-31T06:22:41.5139848Z call = _wrap_func(module.call_129) 2023-03-31T06:22:41.5139854Z 2023-03-31T06:22:41.5139858Z 2023-03-31T06:22:41.5140021Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5140139Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5140268Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5140482Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5140693Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5140909Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5141025Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5141033Z 2023-03-31T06:22:41.5141051Z 2023-03-31T06:22:41.5141116Z if __name__ == "__main__": 2023-03-31T06:22:41.5141193Z import argparse 2023-03-31T06:22:41.5141322Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5141328Z 2023-03-31T06:22:41.5141438Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5141749Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5142090Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5142448Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5142562Z args = parser.parse_args() 2023-03-31T06:22:41.5142570Z 2023-03-31T06:22:41.5142646Z if args.benchmark_kernels: 2023-03-31T06:22:41.5142876Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5142973Z else: 2023-03-31T06:22:41.5143086Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5143181Z benchmark_compiled_module() 2023-03-31T06:22:41.5143187Z 2023-03-31T06:22:41.5143256Z if p: 2023-03-31T06:22:41.5143409Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5143495Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5143632Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5143640Z 2023-03-31T06:22:41.5143922Z [2023-03-31 06:08:58,357] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 105 2023-03-31T06:22:41.5144191Z [2023-03-31 06:08:58,419] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 106 2023-03-31T06:22:41.5144592Z [2023-03-31 06:09:19,032] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/iq/ciq2ctz4fynnakkmlbe4c5ys6nv765crlufj54wnglay6uhejtwz.py 2023-03-31T06:22:41.5144816Z [2023-03-31 06:09:19,032] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5144822Z 2023-03-31T06:22:41.5144918Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5144991Z import torch 2023-03-31T06:22:41.5145065Z import math 2023-03-31T06:22:41.5145127Z import random 2023-03-31T06:22:41.5145196Z import os 2023-03-31T06:22:41.5145271Z import tempfile 2023-03-31T06:22:41.5145393Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5145433Z 2023-03-31T06:22:41.5145553Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5145678Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5145809Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5145815Z 2023-03-31T06:22:41.5145883Z aten = torch.ops.aten 2023-03-31T06:22:41.5146025Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5146120Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5146126Z 2023-03-31T06:22:41.5146216Z async_compile.wait(globals()) 2023-03-31T06:22:41.5146290Z del async_compile 2023-03-31T06:22:41.5146412Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5146484Z wrapper = ( 2023-03-31T06:22:41.5146554Z ''' 2023-03-31T06:22:41.5146628Z #include 2023-03-31T06:22:41.5146732Z #include 2023-03-31T06:22:41.5146737Z 2023-03-31T06:22:41.5146830Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5146836Z 2023-03-31T06:22:41.5146933Z template 2023-03-31T06:22:41.5147058Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5147144Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5147270Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5147355Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5147486Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5147568Z return kernel_cpp; 2023-03-31T06:22:41.5147633Z } 2023-03-31T06:22:41.5147763Z std::vector call_130(std::vector args) { 2023-03-31T06:22:41.5147868Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5147943Z arg0_1 = args[0]; 2023-03-31T06:22:41.5148003Z arg1_1 = args[1]; 2023-03-31T06:22:41.5148074Z arg2_1 = args[2]; 2023-03-31T06:22:41.5148147Z arg3_1 = args[3]; 2023-03-31T06:22:41.5148213Z 2023-03-31T06:22:41.5148310Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5148406Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5148492Z .findSchemaOrThrow( 2023-03-31T06:22:41.5148578Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5148647Z "") 2023-03-31T06:22:41.5148718Z .typed< 2023-03-31T06:22:41.5148792Z at::Tensor( 2023-03-31T06:22:41.5148890Z const at::Tensor& input_t, 2023-03-31T06:22:41.5149019Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5149160Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5149237Z c10::string_view attr, 2023-03-31T06:22:41.5149358Z torch::List> scalars, 2023-03-31T06:22:41.5149479Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5149544Z 2023-03-31T06:22:41.5149783Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.5149856Z arg1_1.reset(); 2023-03-31T06:22:41.5149930Z arg2_1.reset(); 2023-03-31T06:22:41.5149987Z arg3_1.reset(); 2023-03-31T06:22:41.5150091Z return std::vector({buf0}); 2023-03-31T06:22:41.5150154Z } 2023-03-31T06:22:41.5150233Z ''' 2023-03-31T06:22:41.5150294Z ) 2023-03-31T06:22:41.5150300Z 2023-03-31T06:22:41.5150377Z module = load_inline( 2023-03-31T06:22:41.5150649Z name='inline_extension_cet5mw6ckpn5sdinqb23taoug7nnqbjxmhb6i27zj645y6mt3saj', 2023-03-31T06:22:41.5150724Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5150840Z functions=['call_130'], 2023-03-31T06:22:41.5151219Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5151373Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5152106Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5152115Z 2023-03-31T06:22:41.5152192Z def _wrap_func(f): 2023-03-31T06:22:41.5152263Z def g(args): 2023-03-31T06:22:41.5152338Z return f(args) 2023-03-31T06:22:41.5152405Z return g 2023-03-31T06:22:41.5152490Z call = _wrap_func(module.call_130) 2023-03-31T06:22:41.5152498Z 2023-03-31T06:22:41.5152503Z 2023-03-31T06:22:41.5152597Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5152716Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5152843Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5153055Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5153303Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5153513Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5153716Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5153837Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5153843Z 2023-03-31T06:22:41.5153860Z 2023-03-31T06:22:41.5153926Z if __name__ == "__main__": 2023-03-31T06:22:41.5154005Z import argparse 2023-03-31T06:22:41.5154136Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5154142Z 2023-03-31T06:22:41.5154250Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5154559Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5154892Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5155172Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5155266Z args = parser.parse_args() 2023-03-31T06:22:41.5155271Z 2023-03-31T06:22:41.5155349Z if args.benchmark_kernels: 2023-03-31T06:22:41.5155537Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5155605Z else: 2023-03-31T06:22:41.5155717Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5155839Z benchmark_compiled_module() 2023-03-31T06:22:41.5155870Z 2023-03-31T06:22:41.5155939Z if p: 2023-03-31T06:22:41.5156088Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5156171Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5156310Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5156318Z 2023-03-31T06:22:41.5156599Z [2023-03-31 06:09:19,032] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 106 2023-03-31T06:22:41.5156868Z [2023-03-31 06:09:19,079] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 107 2023-03-31T06:22:41.5157276Z [2023-03-31 06:09:39,464] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/gf/cgfbwq5mteaifg6vyywwc5yqpjdvhak3nbtisiq7waovrbmklwy3.py 2023-03-31T06:22:41.5157498Z [2023-03-31 06:09:39,464] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5157506Z 2023-03-31T06:22:41.5157605Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5157677Z import torch 2023-03-31T06:22:41.5157747Z import math 2023-03-31T06:22:41.5157806Z import random 2023-03-31T06:22:41.5157873Z import os 2023-03-31T06:22:41.5157948Z import tempfile 2023-03-31T06:22:41.5158065Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5158073Z 2023-03-31T06:22:41.5158191Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5158320Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5158482Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5158488Z 2023-03-31T06:22:41.5158556Z aten = torch.ops.aten 2023-03-31T06:22:41.5158695Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5158788Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5158794Z 2023-03-31T06:22:41.5158885Z async_compile.wait(globals()) 2023-03-31T06:22:41.5158963Z del async_compile 2023-03-31T06:22:41.5159085Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5159160Z wrapper = ( 2023-03-31T06:22:41.5159229Z ''' 2023-03-31T06:22:41.5159305Z #include 2023-03-31T06:22:41.5159380Z #include 2023-03-31T06:22:41.5159385Z 2023-03-31T06:22:41.5159476Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5159481Z 2023-03-31T06:22:41.5159605Z template 2023-03-31T06:22:41.5159728Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5159814Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5159940Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5160025Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5160152Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5160233Z return kernel_cpp; 2023-03-31T06:22:41.5160297Z } 2023-03-31T06:22:41.5160428Z std::vector call_131(std::vector args) { 2023-03-31T06:22:41.5160526Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5160604Z arg0_1 = args[0]; 2023-03-31T06:22:41.5160665Z arg1_1 = args[1]; 2023-03-31T06:22:41.5160739Z arg2_1 = args[2]; 2023-03-31T06:22:41.5160803Z 2023-03-31T06:22:41.5160904Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5161001Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5161091Z .findSchemaOrThrow( 2023-03-31T06:22:41.5161194Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5161248Z "") 2023-03-31T06:22:41.5161318Z .typed< 2023-03-31T06:22:41.5161401Z at::Tensor( 2023-03-31T06:22:41.5161502Z const at::Tensor& input_t, 2023-03-31T06:22:41.5161607Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5161727Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5161817Z c10::string_view attr, 2023-03-31T06:22:41.5161954Z torch::List> scalars, 2023-03-31T06:22:41.5162103Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5162168Z 2023-03-31T06:22:41.5162421Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "hardsigmoid", {-1}, ""); 2023-03-31T06:22:41.5162496Z arg1_1.reset(); 2023-03-31T06:22:41.5162569Z arg2_1.reset(); 2023-03-31T06:22:41.5162678Z return std::vector({buf0}); 2023-03-31T06:22:41.5162731Z } 2023-03-31T06:22:41.5162813Z ''' 2023-03-31T06:22:41.5162876Z ) 2023-03-31T06:22:41.5162883Z 2023-03-31T06:22:41.5162963Z module = load_inline( 2023-03-31T06:22:41.5163411Z name='inline_extension_ceyanris47ohp6ch643vxttxgfgln7ral5h3xnryxnbufbtgql2n', 2023-03-31T06:22:41.5163502Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5163622Z functions=['call_131'], 2023-03-31T06:22:41.5164007Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5164153Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5164864Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5164874Z 2023-03-31T06:22:41.5164952Z def _wrap_func(f): 2023-03-31T06:22:41.5165025Z def g(args): 2023-03-31T06:22:41.5165163Z return f(args) 2023-03-31T06:22:41.5165233Z return g 2023-03-31T06:22:41.5165331Z call = _wrap_func(module.call_131) 2023-03-31T06:22:41.5165337Z 2023-03-31T06:22:41.5165342Z 2023-03-31T06:22:41.5165439Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5165562Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5165679Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5165898Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5166110Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5166319Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5166449Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5166491Z 2023-03-31T06:22:41.5166496Z 2023-03-31T06:22:41.5166582Z if __name__ == "__main__": 2023-03-31T06:22:41.5166662Z import argparse 2023-03-31T06:22:41.5166796Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5166803Z 2023-03-31T06:22:41.5166900Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5167215Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5167554Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5167832Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5167927Z args = parser.parse_args() 2023-03-31T06:22:41.5167932Z 2023-03-31T06:22:41.5168022Z if args.benchmark_kernels: 2023-03-31T06:22:41.5168215Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5168283Z else: 2023-03-31T06:22:41.5168382Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5168479Z benchmark_compiled_module() 2023-03-31T06:22:41.5168485Z 2023-03-31T06:22:41.5168555Z if p: 2023-03-31T06:22:41.5168704Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5168800Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5168939Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5168983Z 2023-03-31T06:22:41.5169310Z [2023-03-31 06:09:39,464] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 107 2023-03-31T06:22:41.5169576Z [2023-03-31 06:09:39,519] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 108 2023-03-31T06:22:41.5169981Z [2023-03-31 06:09:59,788] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/lj/cljxntdw6daggps5j7zyz7af2hkyv72hk4pfkgrflo6bbxvd7ojq.py 2023-03-31T06:22:41.5170192Z [2023-03-31 06:09:59,788] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5170201Z 2023-03-31T06:22:41.5170297Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5170370Z import torch 2023-03-31T06:22:41.5170441Z import math 2023-03-31T06:22:41.5170513Z import random 2023-03-31T06:22:41.5170582Z import os 2023-03-31T06:22:41.5170655Z import tempfile 2023-03-31T06:22:41.5170760Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5170768Z 2023-03-31T06:22:41.5170889Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5171014Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5171148Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5171154Z 2023-03-31T06:22:41.5171235Z aten = torch.ops.aten 2023-03-31T06:22:41.5171374Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5171471Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5171476Z 2023-03-31T06:22:41.5171566Z async_compile.wait(globals()) 2023-03-31T06:22:41.5171657Z del async_compile 2023-03-31T06:22:41.5171782Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5171854Z wrapper = ( 2023-03-31T06:22:41.5171936Z ''' 2023-03-31T06:22:41.5172013Z #include 2023-03-31T06:22:41.5172089Z #include 2023-03-31T06:22:41.5172094Z 2023-03-31T06:22:41.5172186Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5172191Z 2023-03-31T06:22:41.5172273Z template 2023-03-31T06:22:41.5172399Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5172484Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5172609Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5172706Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5172829Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5172935Z return kernel_cpp; 2023-03-31T06:22:41.5172985Z } 2023-03-31T06:22:41.5173116Z std::vector call_132(std::vector args) { 2023-03-31T06:22:41.5173224Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5173298Z arg0_1 = args[0]; 2023-03-31T06:22:41.5173369Z arg1_1 = args[1]; 2023-03-31T06:22:41.5173440Z arg2_1 = args[2]; 2023-03-31T06:22:41.5173511Z arg3_1 = args[3]; 2023-03-31T06:22:41.5173560Z 2023-03-31T06:22:41.5173657Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5173753Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5173842Z .findSchemaOrThrow( 2023-03-31T06:22:41.5173940Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5174007Z "") 2023-03-31T06:22:41.5174074Z .typed< 2023-03-31T06:22:41.5174135Z at::Tensor( 2023-03-31T06:22:41.5174233Z const at::Tensor& input_t, 2023-03-31T06:22:41.5174333Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5174452Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5174540Z c10::string_view attr, 2023-03-31T06:22:41.5174664Z torch::List> scalars, 2023-03-31T06:22:41.5174785Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5174838Z 2023-03-31T06:22:41.5175057Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.5175129Z arg1_1.reset(); 2023-03-31T06:22:41.5175202Z arg2_1.reset(); 2023-03-31T06:22:41.5175302Z arg3_1.reset(); 2023-03-31T06:22:41.5175443Z return std::vector({buf0}); 2023-03-31T06:22:41.5175507Z } 2023-03-31T06:22:41.5175576Z ''' 2023-03-31T06:22:41.5175636Z ) 2023-03-31T06:22:41.5175641Z 2023-03-31T06:22:41.5175718Z module = load_inline( 2023-03-31T06:22:41.5175980Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.5176069Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5176190Z functions=['call_132'], 2023-03-31T06:22:41.5176578Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5176732Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5177432Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5177453Z 2023-03-31T06:22:41.5177517Z def _wrap_func(f): 2023-03-31T06:22:41.5177588Z def g(args): 2023-03-31T06:22:41.5177664Z return f(args) 2023-03-31T06:22:41.5177732Z return g 2023-03-31T06:22:41.5177830Z call = _wrap_func(module.call_132) 2023-03-31T06:22:41.5177837Z 2023-03-31T06:22:41.5177842Z 2023-03-31T06:22:41.5177936Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5178087Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5178202Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5178416Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5178622Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5178831Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5179048Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5179182Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5179189Z 2023-03-31T06:22:41.5179193Z 2023-03-31T06:22:41.5179272Z if __name__ == "__main__": 2023-03-31T06:22:41.5179348Z import argparse 2023-03-31T06:22:41.5179512Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5179517Z 2023-03-31T06:22:41.5179614Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5179926Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5180263Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5180540Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5180637Z args = parser.parse_args() 2023-03-31T06:22:41.5180642Z 2023-03-31T06:22:41.5180732Z if args.benchmark_kernels: 2023-03-31T06:22:41.5180919Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5180987Z else: 2023-03-31T06:22:41.5181088Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5181184Z benchmark_compiled_module() 2023-03-31T06:22:41.5181189Z 2023-03-31T06:22:41.5181257Z if p: 2023-03-31T06:22:41.5181407Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5181503Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5181639Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5181644Z 2023-03-31T06:22:41.5181924Z [2023-03-31 06:09:59,788] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 108 2023-03-31T06:22:41.5182224Z [2023-03-31 06:09:59,828] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 109 2023-03-31T06:22:41.5182760Z [2023-03-31 06:10:20,190] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/hb/chbwlzsj3v4btf3pgkgbbegqhwj2pmkpxq25kc3ip54eobkp7acs.py 2023-03-31T06:22:41.5182974Z [2023-03-31 06:10:20,190] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5182985Z 2023-03-31T06:22:41.5183082Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5183158Z import torch 2023-03-31T06:22:41.5183231Z import math 2023-03-31T06:22:41.5183307Z import random 2023-03-31T06:22:41.5183376Z import os 2023-03-31T06:22:41.5183452Z import tempfile 2023-03-31T06:22:41.5183557Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5183562Z 2023-03-31T06:22:41.5183682Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5183808Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5183944Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5183952Z 2023-03-31T06:22:41.5184034Z aten = torch.ops.aten 2023-03-31T06:22:41.5184173Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5184268Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5184273Z 2023-03-31T06:22:41.5184364Z async_compile.wait(globals()) 2023-03-31T06:22:41.5184426Z del async_compile 2023-03-31T06:22:41.5184551Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5184624Z wrapper = ( 2023-03-31T06:22:41.5184707Z ''' 2023-03-31T06:22:41.5184818Z #include 2023-03-31T06:22:41.5184895Z #include 2023-03-31T06:22:41.5184900Z 2023-03-31T06:22:41.5184992Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5184998Z 2023-03-31T06:22:41.5185076Z template 2023-03-31T06:22:41.5185199Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5185284Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5185410Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5185509Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5185633Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5185713Z return kernel_cpp; 2023-03-31T06:22:41.5185763Z } 2023-03-31T06:22:41.5185891Z std::vector call_133(std::vector args) { 2023-03-31T06:22:41.5186017Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5186092Z arg0_1 = args[0]; 2023-03-31T06:22:41.5186165Z arg1_1 = args[1]; 2023-03-31T06:22:41.5186237Z arg2_1 = args[2]; 2023-03-31T06:22:41.5186302Z 2023-03-31T06:22:41.5186387Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5186480Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5186567Z .findSchemaOrThrow( 2023-03-31T06:22:41.5186666Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5186732Z "") 2023-03-31T06:22:41.5186801Z .typed< 2023-03-31T06:22:41.5186880Z at::Tensor( 2023-03-31T06:22:41.5186967Z const at::Tensor& input_t, 2023-03-31T06:22:41.5187066Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5187184Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5187276Z c10::string_view attr, 2023-03-31T06:22:41.5187399Z torch::List> scalars, 2023-03-31T06:22:41.5187521Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5187586Z 2023-03-31T06:22:41.5187821Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.5187882Z arg1_1.reset(); 2023-03-31T06:22:41.5187953Z arg2_1.reset(); 2023-03-31T06:22:41.5188059Z return std::vector({buf0}); 2023-03-31T06:22:41.5188121Z } 2023-03-31T06:22:41.5188202Z ''' 2023-03-31T06:22:41.5188263Z ) 2023-03-31T06:22:41.5188268Z 2023-03-31T06:22:41.5188334Z module = load_inline( 2023-03-31T06:22:41.5188637Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.5188748Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5188869Z functions=['call_133'], 2023-03-31T06:22:41.5189248Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5189407Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5190116Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5190123Z 2023-03-31T06:22:41.5190199Z def _wrap_func(f): 2023-03-31T06:22:41.5190270Z def g(args): 2023-03-31T06:22:41.5190335Z return f(args) 2023-03-31T06:22:41.5190404Z return g 2023-03-31T06:22:41.5190500Z call = _wrap_func(module.call_133) 2023-03-31T06:22:41.5190505Z 2023-03-31T06:22:41.5190510Z 2023-03-31T06:22:41.5190605Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5190723Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5190852Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5191067Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5191344Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5191547Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5191674Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5191680Z 2023-03-31T06:22:41.5191685Z 2023-03-31T06:22:41.5191763Z if __name__ == "__main__": 2023-03-31T06:22:41.5191840Z import argparse 2023-03-31T06:22:41.5191976Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5191984Z 2023-03-31T06:22:41.5192093Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5192404Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5192743Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5193055Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5193136Z args = parser.parse_args() 2023-03-31T06:22:41.5193141Z 2023-03-31T06:22:41.5193232Z if args.benchmark_kernels: 2023-03-31T06:22:41.5193422Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5193489Z else: 2023-03-31T06:22:41.5193602Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5193700Z benchmark_compiled_module() 2023-03-31T06:22:41.5193707Z 2023-03-31T06:22:41.5193774Z if p: 2023-03-31T06:22:41.5193923Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5194007Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5194144Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5194152Z 2023-03-31T06:22:41.5194434Z [2023-03-31 06:10:20,191] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 109 2023-03-31T06:22:41.5194706Z [2023-03-31 06:10:20,242] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 110 2023-03-31T06:22:41.5195103Z [2023-03-31 06:10:40,868] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/2a/c2adk3h26dlne2qc2dgt3g2cgb2c2k7qekbli5nudmnqgee3mprt.py 2023-03-31T06:22:41.5195354Z [2023-03-31 06:10:40,868] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5195387Z 2023-03-31T06:22:41.5195486Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5195559Z import torch 2023-03-31T06:22:41.5195618Z import math 2023-03-31T06:22:41.5195690Z import random 2023-03-31T06:22:41.5195757Z import os 2023-03-31T06:22:41.5195831Z import tempfile 2023-03-31T06:22:41.5195949Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5195955Z 2023-03-31T06:22:41.5196077Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5196203Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5196328Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5196346Z 2023-03-31T06:22:41.5196414Z aten = torch.ops.aten 2023-03-31T06:22:41.5196552Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5196647Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5196652Z 2023-03-31T06:22:41.5196742Z async_compile.wait(globals()) 2023-03-31T06:22:41.5196818Z del async_compile 2023-03-31T06:22:41.5196942Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5197013Z wrapper = ( 2023-03-31T06:22:41.5197082Z ''' 2023-03-31T06:22:41.5197159Z #include 2023-03-31T06:22:41.5197236Z #include 2023-03-31T06:22:41.5197242Z 2023-03-31T06:22:41.5197333Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5197338Z 2023-03-31T06:22:41.5197430Z template 2023-03-31T06:22:41.5197556Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5197641Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5197782Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5197880Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5198004Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5198083Z return kernel_cpp; 2023-03-31T06:22:41.5198146Z } 2023-03-31T06:22:41.5198279Z std::vector call_134(std::vector args) { 2023-03-31T06:22:41.5198388Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5198451Z arg0_1 = args[0]; 2023-03-31T06:22:41.5198524Z arg1_1 = args[1]; 2023-03-31T06:22:41.5198600Z arg2_1 = args[2]; 2023-03-31T06:22:41.5198674Z arg3_1 = args[3]; 2023-03-31T06:22:41.5198739Z 2023-03-31T06:22:41.5198840Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5198935Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5199048Z .findSchemaOrThrow( 2023-03-31T06:22:41.5199148Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5199215Z "") 2023-03-31T06:22:41.5199288Z .typed< 2023-03-31T06:22:41.5199366Z at::Tensor( 2023-03-31T06:22:41.5199466Z const at::Tensor& input_t, 2023-03-31T06:22:41.5199572Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5199676Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5199766Z c10::string_view attr, 2023-03-31T06:22:41.5199890Z torch::List> scalars, 2023-03-31T06:22:41.5200015Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5200082Z 2023-03-31T06:22:41.5200308Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.5200384Z arg1_1.reset(); 2023-03-31T06:22:41.5200441Z arg2_1.reset(); 2023-03-31T06:22:41.5200515Z arg3_1.reset(); 2023-03-31T06:22:41.5200623Z return std::vector({buf0}); 2023-03-31T06:22:41.5200689Z } 2023-03-31T06:22:41.5200771Z ''' 2023-03-31T06:22:41.5200838Z ) 2023-03-31T06:22:41.5200843Z 2023-03-31T06:22:41.5200922Z module = load_inline( 2023-03-31T06:22:41.5201173Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.5201264Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5201383Z functions=['call_134'], 2023-03-31T06:22:41.5201796Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5201979Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5202681Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5202695Z 2023-03-31T06:22:41.5202775Z def _wrap_func(f): 2023-03-31T06:22:41.5202851Z def g(args): 2023-03-31T06:22:41.5202928Z return f(args) 2023-03-31T06:22:41.5202986Z return g 2023-03-31T06:22:41.5203231Z call = _wrap_func(module.call_134) 2023-03-31T06:22:41.5203237Z 2023-03-31T06:22:41.5203242Z 2023-03-31T06:22:41.5203341Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5203464Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5203595Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5203814Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5204023Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5204230Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5204429Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5204621Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5204626Z 2023-03-31T06:22:41.5204631Z 2023-03-31T06:22:41.5204710Z if __name__ == "__main__": 2023-03-31T06:22:41.5204786Z import argparse 2023-03-31T06:22:41.5204917Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5204923Z 2023-03-31T06:22:41.5205035Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5205350Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5205696Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5205974Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5206091Z args = parser.parse_args() 2023-03-31T06:22:41.5206095Z 2023-03-31T06:22:41.5206187Z if args.benchmark_kernels: 2023-03-31T06:22:41.5206381Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5206451Z else: 2023-03-31T06:22:41.5206565Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5206663Z benchmark_compiled_module() 2023-03-31T06:22:41.5206668Z 2023-03-31T06:22:41.5206736Z if p: 2023-03-31T06:22:41.5206891Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5206977Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5207115Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5207121Z 2023-03-31T06:22:41.5207401Z [2023-03-31 06:10:40,869] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 110 2023-03-31T06:22:41.5207669Z [2023-03-31 06:10:40,914] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 111 2023-03-31T06:22:41.5208073Z [2023-03-31 06:11:01,818] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/wq/cwqy345lwz24fupi6pzx4zbqchcpkfotsphymnbep7dcplzese6x.py 2023-03-31T06:22:41.5208293Z [2023-03-31 06:11:01,818] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5208299Z 2023-03-31T06:22:41.5208398Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5208474Z import torch 2023-03-31T06:22:41.5208533Z import math 2023-03-31T06:22:41.5208641Z import random 2023-03-31T06:22:41.5208741Z import os 2023-03-31T06:22:41.5208814Z import tempfile 2023-03-31T06:22:41.5208931Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5208937Z 2023-03-31T06:22:41.5209054Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5209179Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5209300Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5209321Z 2023-03-31T06:22:41.5209390Z aten = torch.ops.aten 2023-03-31T06:22:41.5209531Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5209626Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5209631Z 2023-03-31T06:22:41.5209722Z async_compile.wait(globals()) 2023-03-31T06:22:41.5209796Z del async_compile 2023-03-31T06:22:41.5209917Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5209988Z wrapper = ( 2023-03-31T06:22:41.5210060Z ''' 2023-03-31T06:22:41.5210138Z #include 2023-03-31T06:22:41.5210217Z #include 2023-03-31T06:22:41.5210222Z 2023-03-31T06:22:41.5210316Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5210322Z 2023-03-31T06:22:41.5210417Z template 2023-03-31T06:22:41.5210540Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5210626Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5210738Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5210840Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5210964Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5211084Z return kernel_cpp; 2023-03-31T06:22:41.5211147Z } 2023-03-31T06:22:41.5211276Z std::vector call_135(std::vector args) { 2023-03-31T06:22:41.5211368Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5211429Z arg0_1 = args[0]; 2023-03-31T06:22:41.5211501Z arg1_1 = args[1]; 2023-03-31T06:22:41.5211572Z arg2_1 = args[2]; 2023-03-31T06:22:41.5211635Z 2023-03-31T06:22:41.5211735Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5211827Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5211913Z .findSchemaOrThrow( 2023-03-31T06:22:41.5211999Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5212065Z "") 2023-03-31T06:22:41.5212132Z .typed< 2023-03-31T06:22:41.5212206Z at::Tensor( 2023-03-31T06:22:41.5212330Z const at::Tensor& input_t, 2023-03-31T06:22:41.5212432Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5212551Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5212627Z c10::string_view attr, 2023-03-31T06:22:41.5212748Z torch::List> scalars, 2023-03-31T06:22:41.5212868Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5212934Z 2023-03-31T06:22:41.5213171Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.5213246Z arg1_1.reset(); 2023-03-31T06:22:41.5213317Z arg2_1.reset(); 2023-03-31T06:22:41.5213420Z return std::vector({buf0}); 2023-03-31T06:22:41.5213470Z } 2023-03-31T06:22:41.5213549Z ''' 2023-03-31T06:22:41.5213609Z ) 2023-03-31T06:22:41.5213615Z 2023-03-31T06:22:41.5213694Z module = load_inline( 2023-03-31T06:22:41.5213968Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.5214054Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5214172Z functions=['call_135'], 2023-03-31T06:22:41.5214542Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5214695Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5215426Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5215457Z 2023-03-31T06:22:41.5215536Z def _wrap_func(f): 2023-03-31T06:22:41.5215607Z def g(args): 2023-03-31T06:22:41.5215681Z return f(args) 2023-03-31T06:22:41.5215751Z return g 2023-03-31T06:22:41.5215847Z call = _wrap_func(module.call_135) 2023-03-31T06:22:41.5215852Z 2023-03-31T06:22:41.5215857Z 2023-03-31T06:22:41.5215955Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5216062Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5216188Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5216401Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5216616Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5216824Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5216950Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5216955Z 2023-03-31T06:22:41.5216959Z 2023-03-31T06:22:41.5217037Z if __name__ == "__main__": 2023-03-31T06:22:41.5217115Z import argparse 2023-03-31T06:22:41.5217235Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5217241Z 2023-03-31T06:22:41.5217349Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5217687Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5218024Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5218302Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5218395Z args = parser.parse_args() 2023-03-31T06:22:41.5218400Z 2023-03-31T06:22:41.5218489Z if args.benchmark_kernels: 2023-03-31T06:22:41.5218682Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5218736Z else: 2023-03-31T06:22:41.5218848Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5218942Z benchmark_compiled_module() 2023-03-31T06:22:41.5218974Z 2023-03-31T06:22:41.5219044Z if p: 2023-03-31T06:22:41.5219194Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5219290Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5219425Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5219431Z 2023-03-31T06:22:41.5219714Z [2023-03-31 06:11:01,819] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 111 2023-03-31T06:22:41.5219972Z [2023-03-31 06:11:01,874] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 112 2023-03-31T06:22:41.5220375Z [2023-03-31 06:11:23,032] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/6q/c6qr4vd6tq7pc2wxcrm3crbyizrbv24olqlq2ielsjlrnhpbcpjo.py 2023-03-31T06:22:41.5220593Z [2023-03-31 06:11:23,032] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5220601Z 2023-03-31T06:22:41.5220697Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5220768Z import torch 2023-03-31T06:22:41.5220839Z import math 2023-03-31T06:22:41.5220910Z import random 2023-03-31T06:22:41.5220980Z import os 2023-03-31T06:22:41.5221042Z import tempfile 2023-03-31T06:22:41.5221158Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5221163Z 2023-03-31T06:22:41.5221282Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5221407Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5221568Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5221627Z 2023-03-31T06:22:41.5221709Z aten = torch.ops.aten 2023-03-31T06:22:41.5221846Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5221941Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5221946Z 2023-03-31T06:22:41.5222023Z async_compile.wait(globals()) 2023-03-31T06:22:41.5222098Z del async_compile 2023-03-31T06:22:41.5222223Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5222297Z wrapper = ( 2023-03-31T06:22:41.5222466Z ''' 2023-03-31T06:22:41.5222546Z #include 2023-03-31T06:22:41.5222625Z #include 2023-03-31T06:22:41.5222630Z 2023-03-31T06:22:41.5222712Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5222717Z 2023-03-31T06:22:41.5222809Z template 2023-03-31T06:22:41.5222934Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5223019Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5223146Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5223247Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5223372Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5223437Z return kernel_cpp; 2023-03-31T06:22:41.5223499Z } 2023-03-31T06:22:41.5223629Z std::vector call_136(std::vector args) { 2023-03-31T06:22:41.5223735Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5223812Z arg0_1 = args[0]; 2023-03-31T06:22:41.5223884Z arg1_1 = args[1]; 2023-03-31T06:22:41.5223954Z arg2_1 = args[2]; 2023-03-31T06:22:41.5224046Z arg3_1 = args[3]; 2023-03-31T06:22:41.5224110Z 2023-03-31T06:22:41.5224208Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5224301Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5224391Z .findSchemaOrThrow( 2023-03-31T06:22:41.5224489Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5224555Z "") 2023-03-31T06:22:41.5224613Z .typed< 2023-03-31T06:22:41.5224689Z at::Tensor( 2023-03-31T06:22:41.5224786Z const at::Tensor& input_t, 2023-03-31T06:22:41.5224888Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5225004Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5225093Z c10::string_view attr, 2023-03-31T06:22:41.5225212Z torch::List> scalars, 2023-03-31T06:22:41.5225365Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5225417Z 2023-03-31T06:22:41.5225653Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5225726Z arg1_1.reset(); 2023-03-31T06:22:41.5225798Z arg2_1.reset(); 2023-03-31T06:22:41.5225868Z arg3_1.reset(); 2023-03-31T06:22:41.5225972Z return std::vector({buf0}); 2023-03-31T06:22:41.5226034Z } 2023-03-31T06:22:41.5226101Z ''' 2023-03-31T06:22:41.5226162Z ) 2023-03-31T06:22:41.5226170Z 2023-03-31T06:22:41.5226249Z module = load_inline( 2023-03-31T06:22:41.5226522Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.5226607Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5226723Z functions=['call_136'], 2023-03-31T06:22:41.5227106Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5227252Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5227965Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5227972Z 2023-03-31T06:22:41.5228074Z def _wrap_func(f): 2023-03-31T06:22:41.5228173Z def g(args): 2023-03-31T06:22:41.5228248Z return f(args) 2023-03-31T06:22:41.5228316Z return g 2023-03-31T06:22:41.5228412Z call = _wrap_func(module.call_136) 2023-03-31T06:22:41.5228418Z 2023-03-31T06:22:41.5228423Z 2023-03-31T06:22:41.5228517Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5228635Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5228752Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5228964Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5229172Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5229376Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5229591Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5229728Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5229736Z 2023-03-31T06:22:41.5229741Z 2023-03-31T06:22:41.5229820Z if __name__ == "__main__": 2023-03-31T06:22:41.5229896Z import argparse 2023-03-31T06:22:41.5230013Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5230019Z 2023-03-31T06:22:41.5230127Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5230433Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5230795Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5231074Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5231167Z args = parser.parse_args() 2023-03-31T06:22:41.5231172Z 2023-03-31T06:22:41.5231259Z if args.benchmark_kernels: 2023-03-31T06:22:41.5231452Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5231509Z else: 2023-03-31T06:22:41.5231623Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5231717Z benchmark_compiled_module() 2023-03-31T06:22:41.5231723Z 2023-03-31T06:22:41.5231791Z if p: 2023-03-31T06:22:41.5231940Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5232065Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5232201Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5232209Z 2023-03-31T06:22:41.5232492Z [2023-03-31 06:11:23,032] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 112 2023-03-31T06:22:41.5232760Z [2023-03-31 06:11:23,073] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 113 2023-03-31T06:22:41.5233148Z [2023-03-31 06:11:44,302] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zp/czpza2oltuj76wu6hsl6aq7h72up742kwnlgudnvxtah56mrxazj.py 2023-03-31T06:22:41.5233368Z [2023-03-31 06:11:44,303] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5233374Z 2023-03-31T06:22:41.5233470Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5233541Z import torch 2023-03-31T06:22:41.5233611Z import math 2023-03-31T06:22:41.5233683Z import random 2023-03-31T06:22:41.5233753Z import os 2023-03-31T06:22:41.5233814Z import tempfile 2023-03-31T06:22:41.5233931Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5233936Z 2023-03-31T06:22:41.5234056Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5234183Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5234316Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5234322Z 2023-03-31T06:22:41.5234402Z aten = torch.ops.aten 2023-03-31T06:22:41.5234571Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5234694Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5234699Z 2023-03-31T06:22:41.5234778Z async_compile.wait(globals()) 2023-03-31T06:22:41.5234851Z del async_compile 2023-03-31T06:22:41.5234974Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5235047Z wrapper = ( 2023-03-31T06:22:41.5235130Z ''' 2023-03-31T06:22:41.5235205Z #include 2023-03-31T06:22:41.5235288Z #include 2023-03-31T06:22:41.5235293Z 2023-03-31T06:22:41.5235372Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5235377Z 2023-03-31T06:22:41.5235471Z template 2023-03-31T06:22:41.5235594Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5235680Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5235806Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5235902Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5236026Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5236096Z return kernel_cpp; 2023-03-31T06:22:41.5236160Z } 2023-03-31T06:22:41.5236295Z std::vector call_137(std::vector args) { 2023-03-31T06:22:41.5236388Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5236464Z arg0_1 = args[0]; 2023-03-31T06:22:41.5236539Z arg1_1 = args[1]; 2023-03-31T06:22:41.5236612Z arg2_1 = args[2]; 2023-03-31T06:22:41.5236664Z 2023-03-31T06:22:41.5236763Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5236858Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5236947Z .findSchemaOrThrow( 2023-03-31T06:22:41.5237080Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5237150Z "") 2023-03-31T06:22:41.5237220Z .typed< 2023-03-31T06:22:41.5237284Z at::Tensor( 2023-03-31T06:22:41.5237383Z const at::Tensor& input_t, 2023-03-31T06:22:41.5237486Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5237610Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5237703Z c10::string_view attr, 2023-03-31T06:22:41.5237827Z torch::List> scalars, 2023-03-31T06:22:41.5237948Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5238017Z 2023-03-31T06:22:41.5238246Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5238349Z arg1_1.reset(); 2023-03-31T06:22:41.5238420Z arg2_1.reset(); 2023-03-31T06:22:41.5238529Z return std::vector({buf0}); 2023-03-31T06:22:41.5238593Z } 2023-03-31T06:22:41.5238676Z ''' 2023-03-31T06:22:41.5238740Z ) 2023-03-31T06:22:41.5238745Z 2023-03-31T06:22:41.5238811Z module = load_inline( 2023-03-31T06:22:41.5239076Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.5239167Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5239292Z functions=['call_137'], 2023-03-31T06:22:41.5239674Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5239830Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5240540Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5240549Z 2023-03-31T06:22:41.5240627Z def _wrap_func(f): 2023-03-31T06:22:41.5240701Z def g(args): 2023-03-31T06:22:41.5240763Z return f(args) 2023-03-31T06:22:41.5240833Z return g 2023-03-31T06:22:41.5240931Z call = _wrap_func(module.call_137) 2023-03-31T06:22:41.5240937Z 2023-03-31T06:22:41.5240941Z 2023-03-31T06:22:41.5241063Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5241242Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5241371Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5241587Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5241781Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5241998Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5242129Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5242134Z 2023-03-31T06:22:41.5242140Z 2023-03-31T06:22:41.5242219Z if __name__ == "__main__": 2023-03-31T06:22:41.5242296Z import argparse 2023-03-31T06:22:41.5242427Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5242433Z 2023-03-31T06:22:41.5242543Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5242858Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5248449Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5248782Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5248886Z args = parser.parse_args() 2023-03-31T06:22:41.5248894Z 2023-03-31T06:22:41.5248992Z if args.benchmark_kernels: 2023-03-31T06:22:41.5249294Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5249366Z else: 2023-03-31T06:22:41.5249483Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5249583Z benchmark_compiled_module() 2023-03-31T06:22:41.5249589Z 2023-03-31T06:22:41.5249662Z if p: 2023-03-31T06:22:41.5249805Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5249907Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5250047Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5250053Z 2023-03-31T06:22:41.5250348Z [2023-03-31 06:11:44,303] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 113 2023-03-31T06:22:41.5250621Z [2023-03-31 06:11:44,365] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 114 2023-03-31T06:22:41.5251093Z [2023-03-31 06:12:05,384] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/5g/c5gnd2innl2d2gmodm7gtf4bbmsa3xelsz7xpqilx6obmwo4jsx6.py 2023-03-31T06:22:41.5251319Z [2023-03-31 06:12:05,384] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5251326Z 2023-03-31T06:22:41.5251427Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5251503Z import torch 2023-03-31T06:22:41.5251561Z import math 2023-03-31T06:22:41.5251638Z import random 2023-03-31T06:22:41.5251714Z import os 2023-03-31T06:22:41.5251790Z import tempfile 2023-03-31T06:22:41.5251911Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5251917Z 2023-03-31T06:22:41.5252037Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5252166Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5252287Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5252296Z 2023-03-31T06:22:41.5252378Z aten = torch.ops.aten 2023-03-31T06:22:41.5252521Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5252620Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5252625Z 2023-03-31T06:22:41.5252718Z async_compile.wait(globals()) 2023-03-31T06:22:41.5252794Z del async_compile 2023-03-31T06:22:41.5252918Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5252979Z wrapper = ( 2023-03-31T06:22:41.5253063Z ''' 2023-03-31T06:22:41.5253141Z #include 2023-03-31T06:22:41.5253259Z #include 2023-03-31T06:22:41.5253297Z 2023-03-31T06:22:41.5253400Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5253405Z 2023-03-31T06:22:41.5253500Z template 2023-03-31T06:22:41.5253627Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5253715Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5253828Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5253931Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5254056Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5254139Z return kernel_cpp; 2023-03-31T06:22:41.5254205Z } 2023-03-31T06:22:41.5254337Z std::vector call_138(std::vector args) { 2023-03-31T06:22:41.5254442Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5254503Z arg0_1 = args[0]; 2023-03-31T06:22:41.5254578Z arg1_1 = args[1]; 2023-03-31T06:22:41.5254652Z arg2_1 = args[2]; 2023-03-31T06:22:41.5254726Z arg3_1 = args[3]; 2023-03-31T06:22:41.5254790Z 2023-03-31T06:22:41.5254889Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5254973Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5255059Z .findSchemaOrThrow( 2023-03-31T06:22:41.5255156Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5255226Z "") 2023-03-31T06:22:41.5255294Z .typed< 2023-03-31T06:22:41.5255373Z at::Tensor( 2023-03-31T06:22:41.5255469Z const at::Tensor& input_t, 2023-03-31T06:22:41.5255556Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5255700Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5255790Z c10::string_view attr, 2023-03-31T06:22:41.5255910Z torch::List> scalars, 2023-03-31T06:22:41.5256031Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5256096Z 2023-03-31T06:22:41.5256329Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5256403Z arg1_1.reset(); 2023-03-31T06:22:41.5256461Z arg2_1.reset(); 2023-03-31T06:22:41.5256530Z arg3_1.reset(); 2023-03-31T06:22:41.5256636Z return std::vector({buf0}); 2023-03-31T06:22:41.5256698Z } 2023-03-31T06:22:41.5256781Z ''' 2023-03-31T06:22:41.5256844Z ) 2023-03-31T06:22:41.5256879Z 2023-03-31T06:22:41.5256956Z module = load_inline( 2023-03-31T06:22:41.5257217Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.5257306Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5257423Z functions=['call_138'], 2023-03-31T06:22:41.5257806Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5257963Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5258678Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5258687Z 2023-03-31T06:22:41.5258762Z def _wrap_func(f): 2023-03-31T06:22:41.5258836Z def g(args): 2023-03-31T06:22:41.5258911Z return f(args) 2023-03-31T06:22:41.5258966Z return g 2023-03-31T06:22:41.5259065Z call = _wrap_func(module.call_138) 2023-03-31T06:22:41.5259071Z 2023-03-31T06:22:41.5259075Z 2023-03-31T06:22:41.5259170Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5259287Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5259414Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5259656Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5259891Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5260088Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5260294Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5260425Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5260434Z 2023-03-31T06:22:41.5260439Z 2023-03-31T06:22:41.5260517Z if __name__ == "__main__": 2023-03-31T06:22:41.5260593Z import argparse 2023-03-31T06:22:41.5260724Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5260730Z 2023-03-31T06:22:41.5260841Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5261155Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5261496Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5261763Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5261857Z args = parser.parse_args() 2023-03-31T06:22:41.5261862Z 2023-03-31T06:22:41.5261951Z if args.benchmark_kernels: 2023-03-31T06:22:41.5262141Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5262210Z else: 2023-03-31T06:22:41.5262321Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5262566Z benchmark_compiled_module() 2023-03-31T06:22:41.5262573Z 2023-03-31T06:22:41.5262644Z if p: 2023-03-31T06:22:41.5262782Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5262880Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5263019Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5263028Z 2023-03-31T06:22:41.5263314Z [2023-03-31 06:12:05,384] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 114 2023-03-31T06:22:41.5263585Z [2023-03-31 06:12:05,424] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 115 2023-03-31T06:22:41.5263991Z [2023-03-31 06:12:26,499] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/yk/cykhtebec5fzkv54qjs5ijq4wtk3gww2wacqvwh5n5cvqwvyoe5c.py 2023-03-31T06:22:41.5264248Z [2023-03-31 06:12:26,499] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5264256Z 2023-03-31T06:22:41.5264355Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5264428Z import torch 2023-03-31T06:22:41.5264487Z import math 2023-03-31T06:22:41.5264559Z import random 2023-03-31T06:22:41.5264630Z import os 2023-03-31T06:22:41.5264705Z import tempfile 2023-03-31T06:22:41.5264823Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5264829Z 2023-03-31T06:22:41.5264957Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5265085Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5265206Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5265212Z 2023-03-31T06:22:41.5265292Z aten = torch.ops.aten 2023-03-31T06:22:41.5265430Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5265530Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5265535Z 2023-03-31T06:22:41.5265626Z async_compile.wait(globals()) 2023-03-31T06:22:41.5265702Z del async_compile 2023-03-31T06:22:41.5265825Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5265887Z wrapper = ( 2023-03-31T06:22:41.5265968Z ''' 2023-03-31T06:22:41.5266043Z #include 2023-03-31T06:22:41.5266119Z #include 2023-03-31T06:22:41.5266124Z 2023-03-31T06:22:41.5266216Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5266221Z 2023-03-31T06:22:41.5266343Z template 2023-03-31T06:22:41.5266493Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5266583Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5266695Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5266792Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5266918Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5266998Z return kernel_cpp; 2023-03-31T06:22:41.5267060Z } 2023-03-31T06:22:41.5267189Z std::vector call_139(std::vector args) { 2023-03-31T06:22:41.5267283Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5267345Z arg0_1 = args[0]; 2023-03-31T06:22:41.5267417Z arg1_1 = args[1]; 2023-03-31T06:22:41.5267486Z arg2_1 = args[2]; 2023-03-31T06:22:41.5267548Z 2023-03-31T06:22:41.5267644Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5267737Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5267814Z .findSchemaOrThrow( 2023-03-31T06:22:41.5267914Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5267980Z "") 2023-03-31T06:22:41.5268047Z .typed< 2023-03-31T06:22:41.5268122Z at::Tensor( 2023-03-31T06:22:41.5268217Z const at::Tensor& input_t, 2023-03-31T06:22:41.5268319Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5268435Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5268514Z c10::string_view attr, 2023-03-31T06:22:41.5268635Z torch::List> scalars, 2023-03-31T06:22:41.5268780Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5268847Z 2023-03-31T06:22:41.5269091Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5269164Z arg1_1.reset(); 2023-03-31T06:22:41.5269235Z arg2_1.reset(); 2023-03-31T06:22:41.5269330Z return std::vector({buf0}); 2023-03-31T06:22:41.5269394Z } 2023-03-31T06:22:41.5269474Z ''' 2023-03-31T06:22:41.5269535Z ) 2023-03-31T06:22:41.5269543Z 2023-03-31T06:22:41.5269620Z module = load_inline( 2023-03-31T06:22:41.5269887Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.5269973Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5270078Z functions=['call_139'], 2023-03-31T06:22:41.5270494Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5270651Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5271367Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5271376Z 2023-03-31T06:22:41.5271451Z def _wrap_func(f): 2023-03-31T06:22:41.5271522Z def g(args): 2023-03-31T06:22:41.5271596Z return f(args) 2023-03-31T06:22:41.5271664Z return g 2023-03-31T06:22:41.5271759Z call = _wrap_func(module.call_139) 2023-03-31T06:22:41.5271765Z 2023-03-31T06:22:41.5271769Z 2023-03-31T06:22:41.5271853Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5271972Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5272099Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5272317Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5272528Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5272738Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5272899Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5272941Z 2023-03-31T06:22:41.5272946Z 2023-03-31T06:22:41.5273026Z if __name__ == "__main__": 2023-03-31T06:22:41.5273089Z import argparse 2023-03-31T06:22:41.5273219Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5273224Z 2023-03-31T06:22:41.5273333Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5273643Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5273982Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5274257Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5274348Z args = parser.parse_args() 2023-03-31T06:22:41.5274353Z 2023-03-31T06:22:41.5274443Z if args.benchmark_kernels: 2023-03-31T06:22:41.5274632Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5274689Z else: 2023-03-31T06:22:41.5274801Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5274894Z benchmark_compiled_module() 2023-03-31T06:22:41.5274899Z 2023-03-31T06:22:41.5274966Z if p: 2023-03-31T06:22:41.5275116Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5275213Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5275349Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5275355Z 2023-03-31T06:22:41.5275666Z [2023-03-31 06:12:26,499] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 115 2023-03-31T06:22:41.5275929Z [2023-03-31 06:12:26,553] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 116 2023-03-31T06:22:41.5276333Z [2023-03-31 06:12:47,632] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/q3/cq3st6fxpdpqvn2mxt5aubrcjol4zwi5aicn3s2lucnt6bduxlo7.py 2023-03-31T06:22:41.5276554Z [2023-03-31 06:12:47,632] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5276560Z 2023-03-31T06:22:41.5276657Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5276728Z import torch 2023-03-31T06:22:41.5276797Z import math 2023-03-31T06:22:41.5276869Z import random 2023-03-31T06:22:41.5276974Z import os 2023-03-31T06:22:41.5277035Z import tempfile 2023-03-31T06:22:41.5277152Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5277158Z 2023-03-31T06:22:41.5277279Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5277404Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5277538Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5277543Z 2023-03-31T06:22:41.5277623Z aten = torch.ops.aten 2023-03-31T06:22:41.5277761Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5277845Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5277864Z 2023-03-31T06:22:41.5277943Z async_compile.wait(globals()) 2023-03-31T06:22:41.5278017Z del async_compile 2023-03-31T06:22:41.5278140Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5278212Z wrapper = ( 2023-03-31T06:22:41.5278295Z ''' 2023-03-31T06:22:41.5278371Z #include 2023-03-31T06:22:41.5278437Z #include 2023-03-31T06:22:41.5278442Z 2023-03-31T06:22:41.5278534Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5278539Z 2023-03-31T06:22:41.5278633Z template 2023-03-31T06:22:41.5278756Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5278842Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5278966Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5279066Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5279189Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5279281Z return kernel_cpp; 2023-03-31T06:22:41.5279375Z } 2023-03-31T06:22:41.5279502Z std::vector call_140(std::vector args) { 2023-03-31T06:22:41.5279605Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5279680Z arg0_1 = args[0]; 2023-03-31T06:22:41.5279755Z arg1_1 = args[1]; 2023-03-31T06:22:41.5279825Z arg2_1 = args[2]; 2023-03-31T06:22:41.5279886Z arg3_1 = args[3]; 2023-03-31T06:22:41.5279948Z 2023-03-31T06:22:41.5280046Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5280140Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5280229Z .findSchemaOrThrow( 2023-03-31T06:22:41.5280327Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5280380Z "") 2023-03-31T06:22:41.5280448Z .typed< 2023-03-31T06:22:41.5280525Z at::Tensor( 2023-03-31T06:22:41.5280622Z const at::Tensor& input_t, 2023-03-31T06:22:41.5280723Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5280842Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5280933Z c10::string_view attr, 2023-03-31T06:22:41.5281055Z torch::List> scalars, 2023-03-31T06:22:41.5281165Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5281230Z 2023-03-31T06:22:41.5281452Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.5281527Z arg1_1.reset(); 2023-03-31T06:22:41.5281598Z arg2_1.reset(); 2023-03-31T06:22:41.5281696Z arg3_1.reset(); 2023-03-31T06:22:41.5281802Z return std::vector({buf0}); 2023-03-31T06:22:41.5281852Z } 2023-03-31T06:22:41.5281933Z ''' 2023-03-31T06:22:41.5281994Z ) 2023-03-31T06:22:41.5281999Z 2023-03-31T06:22:41.5282076Z module = load_inline( 2023-03-31T06:22:41.5282342Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.5282430Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5282549Z functions=['call_140'], 2023-03-31T06:22:41.5282914Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5283203Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5283919Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5283981Z 2023-03-31T06:22:41.5284057Z def _wrap_func(f): 2023-03-31T06:22:41.5284129Z def g(args): 2023-03-31T06:22:41.5284205Z return f(args) 2023-03-31T06:22:41.5284273Z return g 2023-03-31T06:22:41.5284371Z call = _wrap_func(module.call_140) 2023-03-31T06:22:41.5284379Z 2023-03-31T06:22:41.5284384Z 2023-03-31T06:22:41.5284479Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5284584Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5284711Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5284927Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5285138Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5285346Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5285563Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5285702Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5285708Z 2023-03-31T06:22:41.5285713Z 2023-03-31T06:22:41.5285792Z if __name__ == "__main__": 2023-03-31T06:22:41.5285856Z import argparse 2023-03-31T06:22:41.5286025Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5286063Z 2023-03-31T06:22:41.5286177Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5286489Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5286827Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5287110Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5287205Z args = parser.parse_args() 2023-03-31T06:22:41.5287211Z 2023-03-31T06:22:41.5287302Z if args.benchmark_kernels: 2023-03-31T06:22:41.5287491Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5287546Z else: 2023-03-31T06:22:41.5287658Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5287757Z benchmark_compiled_module() 2023-03-31T06:22:41.5287765Z 2023-03-31T06:22:41.5287835Z if p: 2023-03-31T06:22:41.5287986Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5288082Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5288220Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5288225Z 2023-03-31T06:22:41.5288514Z [2023-03-31 06:12:47,632] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 116 2023-03-31T06:22:41.5288810Z [2023-03-31 06:12:47,681] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 117 2023-03-31T06:22:41.5289211Z [2023-03-31 06:13:08,843] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/y4/cy4xuq4dx55jq5twhab5cv3hz4uwlewybh6dkj32brbkvvzujwkf.py 2023-03-31T06:22:41.5289437Z [2023-03-31 06:13:08,843] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5289446Z 2023-03-31T06:22:41.5289547Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5289621Z import torch 2023-03-31T06:22:41.5289695Z import math 2023-03-31T06:22:41.5289771Z import random 2023-03-31T06:22:41.5289840Z import os 2023-03-31T06:22:41.5289902Z import tempfile 2023-03-31T06:22:41.5290023Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5290029Z 2023-03-31T06:22:41.5290180Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5290306Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5290445Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5290451Z 2023-03-31T06:22:41.5290533Z aten = torch.ops.aten 2023-03-31T06:22:41.5290673Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5290755Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5290775Z 2023-03-31T06:22:41.5290853Z async_compile.wait(globals()) 2023-03-31T06:22:41.5290929Z del async_compile 2023-03-31T06:22:41.5291054Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5291131Z wrapper = ( 2023-03-31T06:22:41.5291217Z ''' 2023-03-31T06:22:41.5291295Z #include 2023-03-31T06:22:41.5291360Z #include 2023-03-31T06:22:41.5291365Z 2023-03-31T06:22:41.5291458Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5291463Z 2023-03-31T06:22:41.5291558Z template 2023-03-31T06:22:41.5291685Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5291772Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5291901Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5292002Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5292128Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5292195Z return kernel_cpp; 2023-03-31T06:22:41.5292259Z } 2023-03-31T06:22:41.5292391Z std::vector call_141(std::vector args) { 2023-03-31T06:22:41.5292511Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5292614Z arg0_1 = args[0]; 2023-03-31T06:22:41.5292687Z arg1_1 = args[1]; 2023-03-31T06:22:41.5292761Z arg2_1 = args[2]; 2023-03-31T06:22:41.5292811Z 2023-03-31T06:22:41.5292907Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5293000Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5293089Z .findSchemaOrThrow( 2023-03-31T06:22:41.5293188Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5293257Z "") 2023-03-31T06:22:41.5293312Z .typed< 2023-03-31T06:22:41.5293386Z at::Tensor( 2023-03-31T06:22:41.5293484Z const at::Tensor& input_t, 2023-03-31T06:22:41.5293584Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5293702Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5293790Z c10::string_view attr, 2023-03-31T06:22:41.5293910Z torch::List> scalars, 2023-03-31T06:22:41.5294033Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5294089Z 2023-03-31T06:22:41.5294321Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.5294395Z arg1_1.reset(); 2023-03-31T06:22:41.5294465Z arg2_1.reset(); 2023-03-31T06:22:41.5294569Z return std::vector({buf0}); 2023-03-31T06:22:41.5294634Z } 2023-03-31T06:22:41.5294714Z ''' 2023-03-31T06:22:41.5294762Z ) 2023-03-31T06:22:41.5294768Z 2023-03-31T06:22:41.5294844Z module = load_inline( 2023-03-31T06:22:41.5295138Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.5295225Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5295343Z functions=['call_141'], 2023-03-31T06:22:41.5295724Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5295881Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5296583Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5296628Z 2023-03-31T06:22:41.5296706Z def _wrap_func(f): 2023-03-31T06:22:41.5296766Z def g(args): 2023-03-31T06:22:41.5296843Z return f(args) 2023-03-31T06:22:41.5296910Z return g 2023-03-31T06:22:41.5297005Z call = _wrap_func(module.call_141) 2023-03-31T06:22:41.5297011Z 2023-03-31T06:22:41.5297016Z 2023-03-31T06:22:41.5297110Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5297228Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5297355Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5297572Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5297771Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5297988Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5298116Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5298124Z 2023-03-31T06:22:41.5298129Z 2023-03-31T06:22:41.5298206Z if __name__ == "__main__": 2023-03-31T06:22:41.5298283Z import argparse 2023-03-31T06:22:41.5298415Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5298421Z 2023-03-31T06:22:41.5298531Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5298840Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5299226Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5299529Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5299622Z args = parser.parse_args() 2023-03-31T06:22:41.5299627Z 2023-03-31T06:22:41.5299716Z if args.benchmark_kernels: 2023-03-31T06:22:41.5299905Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5299975Z else: 2023-03-31T06:22:41.5300088Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5300187Z benchmark_compiled_module() 2023-03-31T06:22:41.5300193Z 2023-03-31T06:22:41.5300249Z if p: 2023-03-31T06:22:41.5300398Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5300495Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5300632Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5300638Z 2023-03-31T06:22:41.5300923Z [2023-03-31 06:13:08,844] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 117 2023-03-31T06:22:41.5301196Z [2023-03-31 06:13:08,906] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 118 2023-03-31T06:22:41.5301594Z [2023-03-31 06:13:29,683] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/n2/cn2kx7civs3plulhbo23eg35mkod2nnsum6vs26mx7gndor4lgpp.py 2023-03-31T06:22:41.5301816Z [2023-03-31 06:13:29,684] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5301848Z 2023-03-31T06:22:41.5301948Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5302007Z import torch 2023-03-31T06:22:41.5302077Z import math 2023-03-31T06:22:41.5302149Z import random 2023-03-31T06:22:41.5302215Z import os 2023-03-31T06:22:41.5302288Z import tempfile 2023-03-31T06:22:41.5302506Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5302512Z 2023-03-31T06:22:41.5302639Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5302755Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5302890Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5302895Z 2023-03-31T06:22:41.5302977Z aten = torch.ops.aten 2023-03-31T06:22:41.5303116Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5303277Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5303283Z 2023-03-31T06:22:41.5303374Z async_compile.wait(globals()) 2023-03-31T06:22:41.5303448Z del async_compile 2023-03-31T06:22:41.5303571Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5303631Z wrapper = ( 2023-03-31T06:22:41.5303721Z ''' 2023-03-31T06:22:41.5303797Z #include 2023-03-31T06:22:41.5303872Z #include 2023-03-31T06:22:41.5303878Z 2023-03-31T06:22:41.5303971Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5303976Z 2023-03-31T06:22:41.5304069Z template 2023-03-31T06:22:41.5304195Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5304268Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5304393Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5304491Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5304615Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5304697Z return kernel_cpp; 2023-03-31T06:22:41.5304760Z } 2023-03-31T06:22:41.5304891Z std::vector call_142(std::vector args) { 2023-03-31T06:22:41.5304984Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5305058Z arg0_1 = args[0]; 2023-03-31T06:22:41.5305131Z arg1_1 = args[1]; 2023-03-31T06:22:41.5305201Z arg2_1 = args[2]; 2023-03-31T06:22:41.5305271Z arg3_1 = args[3]; 2023-03-31T06:22:41.5305333Z 2023-03-31T06:22:41.5305430Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5305511Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5305624Z .findSchemaOrThrow( 2023-03-31T06:22:41.5305749Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5305814Z "") 2023-03-31T06:22:41.5305881Z .typed< 2023-03-31T06:22:41.5305955Z at::Tensor( 2023-03-31T06:22:41.5306052Z const at::Tensor& input_t, 2023-03-31T06:22:41.5306140Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5306259Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5306347Z c10::string_view attr, 2023-03-31T06:22:41.5306471Z torch::List> scalars, 2023-03-31T06:22:41.5306595Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5306660Z 2023-03-31T06:22:41.5306882Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.5306942Z arg1_1.reset(); 2023-03-31T06:22:41.5307011Z arg2_1.reset(); 2023-03-31T06:22:41.5307082Z arg3_1.reset(); 2023-03-31T06:22:41.5307189Z return std::vector({buf0}); 2023-03-31T06:22:41.5307250Z } 2023-03-31T06:22:41.5307331Z ''' 2023-03-31T06:22:41.5307391Z ) 2023-03-31T06:22:41.5307397Z 2023-03-31T06:22:41.5307461Z module = load_inline( 2023-03-31T06:22:41.5307725Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.5307811Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5307932Z functions=['call_142'], 2023-03-31T06:22:41.5308343Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5308500Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5309202Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5309211Z 2023-03-31T06:22:41.5309286Z def _wrap_func(f): 2023-03-31T06:22:41.5309358Z def g(args): 2023-03-31T06:22:41.5309421Z return f(args) 2023-03-31T06:22:41.5309487Z return g 2023-03-31T06:22:41.5309581Z call = _wrap_func(module.call_142) 2023-03-31T06:22:41.5309615Z 2023-03-31T06:22:41.5309620Z 2023-03-31T06:22:41.5309719Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5309837Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5309966Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5310179Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5310385Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5310586Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5310793Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5310927Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5310934Z 2023-03-31T06:22:41.5310939Z 2023-03-31T06:22:41.5311015Z if __name__ == "__main__": 2023-03-31T06:22:41.5311092Z import argparse 2023-03-31T06:22:41.5311220Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5311228Z 2023-03-31T06:22:41.5311337Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5311646Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5311971Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5312281Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5312398Z args = parser.parse_args() 2023-03-31T06:22:41.5312404Z 2023-03-31T06:22:41.5312493Z if args.benchmark_kernels: 2023-03-31T06:22:41.5312683Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5312749Z else: 2023-03-31T06:22:41.5312861Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5312956Z benchmark_compiled_module() 2023-03-31T06:22:41.5312965Z 2023-03-31T06:22:41.5313021Z if p: 2023-03-31T06:22:41.5313169Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5313268Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5313404Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5313410Z 2023-03-31T06:22:41.5313690Z [2023-03-31 06:13:29,684] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 118 2023-03-31T06:22:41.5313961Z [2023-03-31 06:13:29,730] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 119 2023-03-31T06:22:41.5314374Z [2023-03-31 06:13:50,416] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/cb/ccbrwvn2iqpzyux6xyptittmxujvs2wlutybgyxadlrybfa7ho5i.py 2023-03-31T06:22:41.5314593Z [2023-03-31 06:13:50,416] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5314601Z 2023-03-31T06:22:41.5314698Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5314758Z import torch 2023-03-31T06:22:41.5314828Z import math 2023-03-31T06:22:41.5314900Z import random 2023-03-31T06:22:41.5314994Z import os 2023-03-31T06:22:41.5315069Z import tempfile 2023-03-31T06:22:41.5315186Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5315192Z 2023-03-31T06:22:41.5315311Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5315424Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5315559Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5315566Z 2023-03-31T06:22:41.5315646Z aten = torch.ops.aten 2023-03-31T06:22:41.5315785Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5315881Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5315887Z 2023-03-31T06:22:41.5315979Z async_compile.wait(globals()) 2023-03-31T06:22:41.5316054Z del async_compile 2023-03-31T06:22:41.5316202Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5316261Z wrapper = ( 2023-03-31T06:22:41.5316345Z ''' 2023-03-31T06:22:41.5316421Z #include 2023-03-31T06:22:41.5316498Z #include 2023-03-31T06:22:41.5316504Z 2023-03-31T06:22:41.5316594Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5316600Z 2023-03-31T06:22:41.5316692Z template 2023-03-31T06:22:41.5316815Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5316888Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5317013Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5317111Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5317237Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5317316Z return kernel_cpp; 2023-03-31T06:22:41.5317377Z } 2023-03-31T06:22:41.5317506Z std::vector call_143(std::vector args) { 2023-03-31T06:22:41.5317586Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5317661Z arg0_1 = args[0]; 2023-03-31T06:22:41.5317732Z arg1_1 = args[1]; 2023-03-31T06:22:41.5317803Z arg2_1 = args[2]; 2023-03-31T06:22:41.5317863Z 2023-03-31T06:22:41.5317962Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5318054Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5318129Z .findSchemaOrThrow( 2023-03-31T06:22:41.5318227Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5318292Z "") 2023-03-31T06:22:41.5318360Z .typed< 2023-03-31T06:22:41.5318436Z at::Tensor( 2023-03-31T06:22:41.5318560Z const at::Tensor& input_t, 2023-03-31T06:22:41.5318696Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5318801Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5318889Z c10::string_view attr, 2023-03-31T06:22:41.5319011Z torch::List> scalars, 2023-03-31T06:22:41.5319134Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5319200Z 2023-03-31T06:22:41.5319437Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.5319509Z arg1_1.reset(); 2023-03-31T06:22:41.5319568Z arg2_1.reset(); 2023-03-31T06:22:41.5319672Z return std::vector({buf0}); 2023-03-31T06:22:41.5319734Z } 2023-03-31T06:22:41.5319814Z ''' 2023-03-31T06:22:41.5319875Z ) 2023-03-31T06:22:41.5319881Z 2023-03-31T06:22:41.5319958Z module = load_inline( 2023-03-31T06:22:41.5320226Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.5320303Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5320419Z functions=['call_143'], 2023-03-31T06:22:41.5320796Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5320950Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5321692Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5321700Z 2023-03-31T06:22:41.5321777Z def _wrap_func(f): 2023-03-31T06:22:41.5321847Z def g(args): 2023-03-31T06:22:41.5321924Z return f(args) 2023-03-31T06:22:41.5321993Z return g 2023-03-31T06:22:41.5322077Z call = _wrap_func(module.call_143) 2023-03-31T06:22:41.5322082Z 2023-03-31T06:22:41.5322099Z 2023-03-31T06:22:41.5322181Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5322298Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5322425Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5322669Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5322881Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5323244Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5323376Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5323384Z 2023-03-31T06:22:41.5323389Z 2023-03-31T06:22:41.5323467Z if __name__ == "__main__": 2023-03-31T06:22:41.5323530Z import argparse 2023-03-31T06:22:41.5323662Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5323670Z 2023-03-31T06:22:41.5323781Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5324097Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5324437Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5324719Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5324816Z args = parser.parse_args() 2023-03-31T06:22:41.5324822Z 2023-03-31T06:22:41.5324914Z if args.benchmark_kernels: 2023-03-31T06:22:41.5325090Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5325158Z else: 2023-03-31T06:22:41.5325274Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5325423Z benchmark_compiled_module() 2023-03-31T06:22:41.5325462Z 2023-03-31T06:22:41.5325533Z if p: 2023-03-31T06:22:41.5325685Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5325787Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5325911Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5325931Z 2023-03-31T06:22:41.5326206Z [2023-03-31 06:13:50,416] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 119 2023-03-31T06:22:41.5326480Z [2023-03-31 06:13:50,475] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 120 2023-03-31T06:22:41.5326899Z [2023-03-31 06:14:11,354] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/uf/cufcyciqditcnjqa3ryknncsledwisorijxu5n4efmpjuesl75m3.py 2023-03-31T06:22:41.5327119Z [2023-03-31 06:14:11,355] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5327125Z 2023-03-31T06:22:41.5327227Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5327303Z import torch 2023-03-31T06:22:41.5327377Z import math 2023-03-31T06:22:41.5327453Z import random 2023-03-31T06:22:41.5327508Z import os 2023-03-31T06:22:41.5327584Z import tempfile 2023-03-31T06:22:41.5327703Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5327709Z 2023-03-31T06:22:41.5327830Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5327959Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5328129Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5328136Z 2023-03-31T06:22:41.5328220Z aten = torch.ops.aten 2023-03-31T06:22:41.5328346Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5328441Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5328447Z 2023-03-31T06:22:41.5328539Z async_compile.wait(globals()) 2023-03-31T06:22:41.5328616Z del async_compile 2023-03-31T06:22:41.5328742Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5328817Z wrapper = ( 2023-03-31T06:22:41.5328907Z ''' 2023-03-31T06:22:41.5328972Z #include 2023-03-31T06:22:41.5329049Z #include 2023-03-31T06:22:41.5329055Z 2023-03-31T06:22:41.5329150Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5329156Z 2023-03-31T06:22:41.5329252Z template 2023-03-31T06:22:41.5329418Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5329503Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5329633Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5329732Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5329845Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5329925Z return kernel_cpp; 2023-03-31T06:22:41.5329989Z } 2023-03-31T06:22:41.5330120Z std::vector call_144(std::vector args) { 2023-03-31T06:22:41.5330225Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5330305Z arg0_1 = args[0]; 2023-03-31T06:22:41.5330377Z arg1_1 = args[1]; 2023-03-31T06:22:41.5330436Z arg2_1 = args[2]; 2023-03-31T06:22:41.5330508Z arg3_1 = args[3]; 2023-03-31T06:22:41.5330569Z 2023-03-31T06:22:41.5330667Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5330760Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5330849Z .findSchemaOrThrow( 2023-03-31T06:22:41.5330948Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5331003Z "") 2023-03-31T06:22:41.5331072Z .typed< 2023-03-31T06:22:41.5331150Z at::Tensor( 2023-03-31T06:22:41.5331247Z const at::Tensor& input_t, 2023-03-31T06:22:41.5331349Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5331467Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5331558Z c10::string_view attr, 2023-03-31T06:22:41.5331693Z torch::List> scalars, 2023-03-31T06:22:41.5331854Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5331919Z 2023-03-31T06:22:41.5332142Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.5332214Z arg1_1.reset(); 2023-03-31T06:22:41.5332285Z arg2_1.reset(); 2023-03-31T06:22:41.5332353Z arg3_1.reset(); 2023-03-31T06:22:41.5332449Z return std::vector({buf0}); 2023-03-31T06:22:41.5332511Z } 2023-03-31T06:22:41.5332591Z ''' 2023-03-31T06:22:41.5332655Z ) 2023-03-31T06:22:41.5332661Z 2023-03-31T06:22:41.5332743Z module = load_inline( 2023-03-31T06:22:41.5333008Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.5333095Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5333200Z functions=['call_144'], 2023-03-31T06:22:41.5333582Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5333740Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5334446Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5334456Z 2023-03-31T06:22:41.5334531Z def _wrap_func(f): 2023-03-31T06:22:41.5334630Z def g(args): 2023-03-31T06:22:41.5334707Z return f(args) 2023-03-31T06:22:41.5334773Z return g 2023-03-31T06:22:41.5334869Z call = _wrap_func(module.call_144) 2023-03-31T06:22:41.5334875Z 2023-03-31T06:22:41.5334880Z 2023-03-31T06:22:41.5334962Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5335081Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5335210Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5335426Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5335634Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5335843Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5336086Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5336222Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5336231Z 2023-03-31T06:22:41.5336236Z 2023-03-31T06:22:41.5336313Z if __name__ == "__main__": 2023-03-31T06:22:41.5336376Z import argparse 2023-03-31T06:22:41.5336506Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5336512Z 2023-03-31T06:22:41.5336620Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5336935Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5337276Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5337550Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5337645Z args = parser.parse_args() 2023-03-31T06:22:41.5337650Z 2023-03-31T06:22:41.5337738Z if args.benchmark_kernels: 2023-03-31T06:22:41.5337918Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5337985Z else: 2023-03-31T06:22:41.5338097Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5338191Z benchmark_compiled_module() 2023-03-31T06:22:41.5338198Z 2023-03-31T06:22:41.5338268Z if p: 2023-03-31T06:22:41.5338418Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5338543Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5338692Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5338710Z 2023-03-31T06:22:41.5338978Z [2023-03-31 06:14:11,355] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 120 2023-03-31T06:22:41.5339250Z [2023-03-31 06:14:11,396] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 121 2023-03-31T06:22:41.5339661Z [2023-03-31 06:14:32,269] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/l6/cl6b5ymmltxjvczuhfkg32orijdlhq5wwx6c5lhhcgmkoc5ffljf.py 2023-03-31T06:22:41.5339880Z [2023-03-31 06:14:32,269] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5339886Z 2023-03-31T06:22:41.5339983Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5340055Z import torch 2023-03-31T06:22:41.5340130Z import math 2023-03-31T06:22:41.5340201Z import random 2023-03-31T06:22:41.5340259Z import os 2023-03-31T06:22:41.5340335Z import tempfile 2023-03-31T06:22:41.5340451Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5340457Z 2023-03-31T06:22:41.5340576Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5340701Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5340838Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5340847Z 2023-03-31T06:22:41.5340926Z aten = torch.ops.aten 2023-03-31T06:22:41.5341065Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5341174Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5341180Z 2023-03-31T06:22:41.5341273Z async_compile.wait(globals()) 2023-03-31T06:22:41.5341349Z del async_compile 2023-03-31T06:22:41.5341470Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5341545Z wrapper = ( 2023-03-31T06:22:41.5341626Z ''' 2023-03-31T06:22:41.5341700Z #include 2023-03-31T06:22:41.5341765Z #include 2023-03-31T06:22:41.5341773Z 2023-03-31T06:22:41.5341864Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5341870Z 2023-03-31T06:22:41.5341963Z template 2023-03-31T06:22:41.5342086Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5342171Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5342295Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5342513Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5342628Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5342708Z return kernel_cpp; 2023-03-31T06:22:41.5342771Z } 2023-03-31T06:22:41.5342903Z std::vector call_145(std::vector args) { 2023-03-31T06:22:41.5342996Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5343070Z arg0_1 = args[0]; 2023-03-31T06:22:41.5343142Z arg1_1 = args[1]; 2023-03-31T06:22:41.5343200Z arg2_1 = args[2]; 2023-03-31T06:22:41.5343263Z 2023-03-31T06:22:41.5343365Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5343462Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5343547Z .findSchemaOrThrow( 2023-03-31T06:22:41.5343645Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5343711Z "") 2023-03-31T06:22:41.5343767Z .typed< 2023-03-31T06:22:41.5343841Z at::Tensor( 2023-03-31T06:22:41.5343942Z const at::Tensor& input_t, 2023-03-31T06:22:41.5344044Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5344165Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5344253Z c10::string_view attr, 2023-03-31T06:22:41.5344376Z torch::List> scalars, 2023-03-31T06:22:41.5344486Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5344550Z 2023-03-31T06:22:41.5344827Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.5344927Z arg1_1.reset(); 2023-03-31T06:22:41.5344997Z arg2_1.reset(); 2023-03-31T06:22:41.5345101Z return std::vector({buf0}); 2023-03-31T06:22:41.5345164Z } 2023-03-31T06:22:41.5345232Z ''' 2023-03-31T06:22:41.5345292Z ) 2023-03-31T06:22:41.5345299Z 2023-03-31T06:22:41.5345376Z module = load_inline( 2023-03-31T06:22:41.5345650Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.5345737Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5345855Z functions=['call_145'], 2023-03-31T06:22:41.5346239Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5346394Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5347101Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5347110Z 2023-03-31T06:22:41.5347173Z def _wrap_func(f): 2023-03-31T06:22:41.5347244Z def g(args): 2023-03-31T06:22:41.5347323Z return f(args) 2023-03-31T06:22:41.5347393Z return g 2023-03-31T06:22:41.5347489Z call = _wrap_func(module.call_145) 2023-03-31T06:22:41.5347495Z 2023-03-31T06:22:41.5347500Z 2023-03-31T06:22:41.5347618Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5347738Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5347865Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5348065Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5348275Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5348499Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5348631Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5348637Z 2023-03-31T06:22:41.5348642Z 2023-03-31T06:22:41.5348720Z if __name__ == "__main__": 2023-03-31T06:22:41.5348798Z import argparse 2023-03-31T06:22:41.5348929Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5348995Z 2023-03-31T06:22:41.5349106Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5349407Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5349745Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5350021Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5350116Z args = parser.parse_args() 2023-03-31T06:22:41.5350122Z 2023-03-31T06:22:41.5350217Z if args.benchmark_kernels: 2023-03-31T06:22:41.5350405Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5350473Z else: 2023-03-31T06:22:41.5350589Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5350671Z benchmark_compiled_module() 2023-03-31T06:22:41.5350691Z 2023-03-31T06:22:41.5350747Z if p: 2023-03-31T06:22:41.5350896Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5350994Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5351131Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5351137Z 2023-03-31T06:22:41.5351419Z [2023-03-31 06:14:32,270] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 121 2023-03-31T06:22:41.5351714Z [2023-03-31 06:14:32,322] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 122 2023-03-31T06:22:41.5352134Z [2023-03-31 06:14:53,166] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/5t/c5tc63j65okd5yt6msbttzletiiw6dikcfl6jtheaasd2gowged5.py 2023-03-31T06:22:41.5352353Z [2023-03-31 06:14:53,167] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5352359Z 2023-03-31T06:22:41.5352446Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5352518Z import torch 2023-03-31T06:22:41.5352589Z import math 2023-03-31T06:22:41.5352661Z import random 2023-03-31T06:22:41.5352731Z import os 2023-03-31T06:22:41.5352805Z import tempfile 2023-03-31T06:22:41.5352922Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5352927Z 2023-03-31T06:22:41.5353034Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5353158Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5353293Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5353300Z 2023-03-31T06:22:41.5353381Z aten = torch.ops.aten 2023-03-31T06:22:41.5353520Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5353614Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5353620Z 2023-03-31T06:22:41.5353710Z async_compile.wait(globals()) 2023-03-31T06:22:41.5353784Z del async_compile 2023-03-31T06:22:41.5353893Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5353967Z wrapper = ( 2023-03-31T06:22:41.5354048Z ''' 2023-03-31T06:22:41.5354123Z #include 2023-03-31T06:22:41.5354232Z #include 2023-03-31T06:22:41.5354237Z 2023-03-31T06:22:41.5354329Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5354334Z 2023-03-31T06:22:41.5354425Z template 2023-03-31T06:22:41.5354537Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5354624Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5354752Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5354851Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5354977Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5355055Z return kernel_cpp; 2023-03-31T06:22:41.5355118Z } 2023-03-31T06:22:41.5355235Z std::vector call_146(std::vector args) { 2023-03-31T06:22:41.5355338Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5355438Z arg0_1 = args[0]; 2023-03-31T06:22:41.5355511Z arg1_1 = args[1]; 2023-03-31T06:22:41.5355582Z arg2_1 = args[2]; 2023-03-31T06:22:41.5355656Z arg3_1 = args[3]; 2023-03-31T06:22:41.5355718Z 2023-03-31T06:22:41.5355804Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5355896Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5355983Z .findSchemaOrThrow( 2023-03-31T06:22:41.5356080Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5356146Z "") 2023-03-31T06:22:41.5356213Z .typed< 2023-03-31T06:22:41.5356290Z at::Tensor( 2023-03-31T06:22:41.5356378Z const at::Tensor& input_t, 2023-03-31T06:22:41.5356478Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5356595Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5356684Z c10::string_view attr, 2023-03-31T06:22:41.5356805Z torch::List> scalars, 2023-03-31T06:22:41.5356928Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5356994Z 2023-03-31T06:22:41.5357218Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "relu", {-1}, ""); 2023-03-31T06:22:41.5357279Z arg1_1.reset(); 2023-03-31T06:22:41.5357349Z arg2_1.reset(); 2023-03-31T06:22:41.5357418Z arg3_1.reset(); 2023-03-31T06:22:41.5357521Z return std::vector({buf0}); 2023-03-31T06:22:41.5357584Z } 2023-03-31T06:22:41.5357664Z ''' 2023-03-31T06:22:41.5357713Z ) 2023-03-31T06:22:41.5357746Z 2023-03-31T06:22:41.5357852Z module = load_inline( 2023-03-31T06:22:41.5358116Z name='inline_extension_cfp32nu5nvaa2is25xsr35s2d77fogwzkiwwgsdellnpkybk6deo', 2023-03-31T06:22:41.5358201Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5358317Z functions=['call_146'], 2023-03-31T06:22:41.5358696Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5358860Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5359568Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5359575Z 2023-03-31T06:22:41.5359649Z def _wrap_func(f): 2023-03-31T06:22:41.5359709Z def g(args): 2023-03-31T06:22:41.5359784Z return f(args) 2023-03-31T06:22:41.5359851Z return g 2023-03-31T06:22:41.5359946Z call = _wrap_func(module.call_146) 2023-03-31T06:22:41.5359951Z 2023-03-31T06:22:41.5359955Z 2023-03-31T06:22:41.5360048Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5360166Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5360295Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5360508Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5360729Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5360935Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5361141Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5361273Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5361281Z 2023-03-31T06:22:41.5361288Z 2023-03-31T06:22:41.5361365Z if __name__ == "__main__": 2023-03-31T06:22:41.5361442Z import argparse 2023-03-31T06:22:41.5361570Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5361576Z 2023-03-31T06:22:41.5361686Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5361983Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5362351Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5362629Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5362725Z args = parser.parse_args() 2023-03-31T06:22:41.5362730Z 2023-03-31T06:22:41.5362821Z if args.benchmark_kernels: 2023-03-31T06:22:41.5363141Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5363211Z else: 2023-03-31T06:22:41.5363326Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5363409Z benchmark_compiled_module() 2023-03-31T06:22:41.5363428Z 2023-03-31T06:22:41.5363484Z if p: 2023-03-31T06:22:41.5363633Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5363734Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5363872Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5363878Z 2023-03-31T06:22:41.5364165Z [2023-03-31 06:14:53,167] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 122 2023-03-31T06:22:41.5364436Z [2023-03-31 06:14:53,206] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 123 2023-03-31T06:22:41.5364888Z [2023-03-31 06:15:14,160] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zv/czvud6p2y4z5jwllozasq66hm7gkcjargztkdrezuqz7uqzudfjs.py 2023-03-31T06:22:41.5365159Z [2023-03-31 06:15:14,160] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5365165Z 2023-03-31T06:22:41.5365251Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5365323Z import torch 2023-03-31T06:22:41.5365395Z import math 2023-03-31T06:22:41.5365468Z import random 2023-03-31T06:22:41.5365541Z import os 2023-03-31T06:22:41.5365615Z import tempfile 2023-03-31T06:22:41.5365737Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5365743Z 2023-03-31T06:22:41.5365852Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5365979Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5366112Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5366118Z 2023-03-31T06:22:41.5366199Z aten = torch.ops.aten 2023-03-31T06:22:41.5366338Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5366440Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5366447Z 2023-03-31T06:22:41.5366540Z async_compile.wait(globals()) 2023-03-31T06:22:41.5366615Z del async_compile 2023-03-31T06:22:41.5366723Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5366795Z wrapper = ( 2023-03-31T06:22:41.5366879Z ''' 2023-03-31T06:22:41.5366957Z #include 2023-03-31T06:22:41.5367035Z #include 2023-03-31T06:22:41.5367042Z 2023-03-31T06:22:41.5367136Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5367142Z 2023-03-31T06:22:41.5367236Z template 2023-03-31T06:22:41.5367381Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5367471Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5367599Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5367698Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5367825Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5367908Z return kernel_cpp; 2023-03-31T06:22:41.5367975Z } 2023-03-31T06:22:41.5368093Z std::vector call_147(std::vector args) { 2023-03-31T06:22:41.5368189Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5368265Z arg0_1 = args[0]; 2023-03-31T06:22:41.5368341Z arg1_1 = args[1]; 2023-03-31T06:22:41.5368414Z arg2_1 = args[2]; 2023-03-31T06:22:41.5368478Z 2023-03-31T06:22:41.5368614Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5368695Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5368784Z .findSchemaOrThrow( 2023-03-31T06:22:41.5368883Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5368952Z "") 2023-03-31T06:22:41.5369018Z .typed< 2023-03-31T06:22:41.5369092Z at::Tensor( 2023-03-31T06:22:41.5369189Z const at::Tensor& input_t, 2023-03-31T06:22:41.5369277Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5369397Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5369491Z c10::string_view attr, 2023-03-31T06:22:41.5369611Z torch::List> scalars, 2023-03-31T06:22:41.5369731Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5369795Z 2023-03-31T06:22:41.5370032Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "relu", {-1}, ""); 2023-03-31T06:22:41.5370106Z arg1_1.reset(); 2023-03-31T06:22:41.5370163Z arg2_1.reset(); 2023-03-31T06:22:41.5370267Z return std::vector({buf0}); 2023-03-31T06:22:41.5370330Z } 2023-03-31T06:22:41.5370410Z ''' 2023-03-31T06:22:41.5370471Z ) 2023-03-31T06:22:41.5370477Z 2023-03-31T06:22:41.5370552Z module = load_inline( 2023-03-31T06:22:41.5370823Z name='inline_extension_cnjwfwcz5gcvdxw5gylsnjtfo7bmeradrx4lg7g2jcwvqd26adqu', 2023-03-31T06:22:41.5370896Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5371013Z functions=['call_147'], 2023-03-31T06:22:41.5371425Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5371605Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5372312Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5372321Z 2023-03-31T06:22:41.5372397Z def _wrap_func(f): 2023-03-31T06:22:41.5372468Z def g(args): 2023-03-31T06:22:41.5372541Z return f(args) 2023-03-31T06:22:41.5372596Z return g 2023-03-31T06:22:41.5372690Z call = _wrap_func(module.call_147) 2023-03-31T06:22:41.5372696Z 2023-03-31T06:22:41.5372700Z 2023-03-31T06:22:41.5372795Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5372918Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5373044Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5373256Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5373465Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5373674Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5373787Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5373807Z 2023-03-31T06:22:41.5373838Z 2023-03-31T06:22:41.5373905Z if __name__ == "__main__": 2023-03-31T06:22:41.5373980Z import argparse 2023-03-31T06:22:41.5374109Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5374114Z 2023-03-31T06:22:41.5374222Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5374533Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5374871Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5375152Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5375272Z args = parser.parse_args() 2023-03-31T06:22:41.5375277Z 2023-03-31T06:22:41.5375354Z if args.benchmark_kernels: 2023-03-31T06:22:41.5375546Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5375613Z else: 2023-03-31T06:22:41.5375724Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5375817Z benchmark_compiled_module() 2023-03-31T06:22:41.5375823Z 2023-03-31T06:22:41.5375891Z if p: 2023-03-31T06:22:41.5376041Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5376141Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5376267Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5376272Z 2023-03-31T06:22:41.5376551Z [2023-03-31 06:15:14,161] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 123 2023-03-31T06:22:41.5376820Z [2023-03-31 06:15:14,216] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 124 2023-03-31T06:22:41.5377229Z [2023-03-31 06:15:35,503] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/vb/cvbyibd7vvzunlt54zzqhilgy27ivwstnbijjjrnpfuap5tdzsol.py 2023-03-31T06:22:41.5377447Z [2023-03-31 06:15:35,503] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5377453Z 2023-03-31T06:22:41.5377549Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5377621Z import torch 2023-03-31T06:22:41.5377691Z import math 2023-03-31T06:22:41.5377749Z import random 2023-03-31T06:22:41.5377843Z import os 2023-03-31T06:22:41.5377943Z import tempfile 2023-03-31T06:22:41.5378058Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5378064Z 2023-03-31T06:22:41.5378181Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5378305Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5378437Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5378445Z 2023-03-31T06:22:41.5378527Z aten = torch.ops.aten 2023-03-31T06:22:41.5378653Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5378748Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5378754Z 2023-03-31T06:22:41.5378844Z async_compile.wait(globals()) 2023-03-31T06:22:41.5378917Z del async_compile 2023-03-31T06:22:41.5379038Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5379109Z wrapper = ( 2023-03-31T06:22:41.5379191Z ''' 2023-03-31T06:22:41.5379254Z #include 2023-03-31T06:22:41.5379331Z #include 2023-03-31T06:22:41.5379339Z 2023-03-31T06:22:41.5379429Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5379435Z 2023-03-31T06:22:41.5379527Z template 2023-03-31T06:22:41.5379652Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5379739Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5379863Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5379950Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5380073Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5380152Z return kernel_cpp; 2023-03-31T06:22:41.5380243Z } 2023-03-31T06:22:41.5380373Z std::vector call_148(std::vector args) { 2023-03-31T06:22:41.5380477Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5380554Z arg0_1 = args[0]; 2023-03-31T06:22:41.5380614Z arg1_1 = args[1]; 2023-03-31T06:22:41.5380684Z arg2_1 = args[2]; 2023-03-31T06:22:41.5380754Z arg3_1 = args[3]; 2023-03-31T06:22:41.5380818Z 2023-03-31T06:22:41.5380918Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5381011Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5381098Z .findSchemaOrThrow( 2023-03-31T06:22:41.5381185Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5381252Z "") 2023-03-31T06:22:41.5381318Z .typed< 2023-03-31T06:22:41.5381393Z at::Tensor( 2023-03-31T06:22:41.5381517Z const at::Tensor& input_t, 2023-03-31T06:22:41.5381617Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5381735Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5381811Z c10::string_view attr, 2023-03-31T06:22:41.5381934Z torch::List> scalars, 2023-03-31T06:22:41.5382056Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5382122Z 2023-03-31T06:22:41.5382349Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5382507Z arg1_1.reset(); 2023-03-31T06:22:41.5382579Z arg2_1.reset(); 2023-03-31T06:22:41.5382650Z arg3_1.reset(); 2023-03-31T06:22:41.5382743Z return std::vector({buf0}); 2023-03-31T06:22:41.5382806Z } 2023-03-31T06:22:41.5382890Z ''' 2023-03-31T06:22:41.5382952Z ) 2023-03-31T06:22:41.5382959Z 2023-03-31T06:22:41.5383042Z module = load_inline( 2023-03-31T06:22:41.5383314Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.5383400Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5383506Z functions=['call_148'], 2023-03-31T06:22:41.5383888Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5384044Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5384806Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5384844Z 2023-03-31T06:22:41.5384920Z def _wrap_func(f): 2023-03-31T06:22:41.5384993Z def g(args): 2023-03-31T06:22:41.5385068Z return f(args) 2023-03-31T06:22:41.5385135Z return g 2023-03-31T06:22:41.5385218Z call = _wrap_func(module.call_148) 2023-03-31T06:22:41.5385224Z 2023-03-31T06:22:41.5385243Z 2023-03-31T06:22:41.5385326Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5385444Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5385573Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5385788Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5386000Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5386210Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5386426Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5386549Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5386570Z 2023-03-31T06:22:41.5386574Z 2023-03-31T06:22:41.5386640Z if __name__ == "__main__": 2023-03-31T06:22:41.5386717Z import argparse 2023-03-31T06:22:41.5386877Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5386883Z 2023-03-31T06:22:41.5386995Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5387306Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5387643Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5387922Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5388015Z args = parser.parse_args() 2023-03-31T06:22:41.5388021Z 2023-03-31T06:22:41.5388097Z if args.benchmark_kernels: 2023-03-31T06:22:41.5388284Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5388391Z else: 2023-03-31T06:22:41.5388504Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5388598Z benchmark_compiled_module() 2023-03-31T06:22:41.5388606Z 2023-03-31T06:22:41.5388674Z if p: 2023-03-31T06:22:41.5388825Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5388920Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5389045Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5389050Z 2023-03-31T06:22:41.5389337Z [2023-03-31 06:15:35,504] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 124 2023-03-31T06:22:41.5389607Z [2023-03-31 06:15:35,546] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 125 2023-03-31T06:22:41.5390009Z [2023-03-31 06:15:56,906] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zj/czjsibicpri23enklj6akk33swn6rqgxnx3xu2g5k2hiycfwmy3d.py 2023-03-31T06:22:41.5390231Z [2023-03-31 06:15:56,906] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5390238Z 2023-03-31T06:22:41.5390336Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5390407Z import torch 2023-03-31T06:22:41.5390478Z import math 2023-03-31T06:22:41.5390537Z import random 2023-03-31T06:22:41.5390604Z import os 2023-03-31T06:22:41.5390679Z import tempfile 2023-03-31T06:22:41.5390796Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5390802Z 2023-03-31T06:22:41.5390951Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5391104Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5391235Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5391241Z 2023-03-31T06:22:41.5391320Z aten = torch.ops.aten 2023-03-31T06:22:41.5391446Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5391540Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5391548Z 2023-03-31T06:22:41.5391639Z async_compile.wait(globals()) 2023-03-31T06:22:41.5391713Z del async_compile 2023-03-31T06:22:41.5391835Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5391907Z wrapper = ( 2023-03-31T06:22:41.5391987Z ''' 2023-03-31T06:22:41.5392050Z #include 2023-03-31T06:22:41.5392124Z #include 2023-03-31T06:22:41.5392129Z 2023-03-31T06:22:41.5392220Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5392225Z 2023-03-31T06:22:41.5392316Z template 2023-03-31T06:22:41.5392441Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5392529Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5392655Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5392739Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5392863Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5392942Z return kernel_cpp; 2023-03-31T06:22:41.5393006Z } 2023-03-31T06:22:41.5393138Z std::vector call_149(std::vector args) { 2023-03-31T06:22:41.5393231Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5393333Z arg0_1 = args[0]; 2023-03-31T06:22:41.5393395Z arg1_1 = args[1]; 2023-03-31T06:22:41.5393465Z arg2_1 = args[2]; 2023-03-31T06:22:41.5393527Z 2023-03-31T06:22:41.5393624Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5393719Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5393807Z .findSchemaOrThrow( 2023-03-31T06:22:41.5393907Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5393963Z "") 2023-03-31T06:22:41.5394030Z .typed< 2023-03-31T06:22:41.5394104Z at::Tensor( 2023-03-31T06:22:41.5394201Z const at::Tensor& input_t, 2023-03-31T06:22:41.5394300Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5394416Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5394531Z c10::string_view attr, 2023-03-31T06:22:41.5394640Z torch::List> scalars, 2023-03-31T06:22:41.5394764Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5394829Z 2023-03-31T06:22:41.5395069Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5395140Z arg1_1.reset(); 2023-03-31T06:22:41.5395210Z arg2_1.reset(); 2023-03-31T06:22:41.5395315Z return std::vector({buf0}); 2023-03-31T06:22:41.5395378Z } 2023-03-31T06:22:41.5395447Z ''' 2023-03-31T06:22:41.5395508Z ) 2023-03-31T06:22:41.5395513Z 2023-03-31T06:22:41.5395591Z module = load_inline( 2023-03-31T06:22:41.5395854Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.5395940Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5396057Z functions=['call_149'], 2023-03-31T06:22:41.5396432Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5396580Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5397323Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5397366Z 2023-03-31T06:22:41.5397430Z def _wrap_func(f): 2023-03-31T06:22:41.5397502Z def g(args): 2023-03-31T06:22:41.5397576Z return f(args) 2023-03-31T06:22:41.5397643Z return g 2023-03-31T06:22:41.5397739Z call = _wrap_func(module.call_149) 2023-03-31T06:22:41.5397744Z 2023-03-31T06:22:41.5397749Z 2023-03-31T06:22:41.5397843Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5397963Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5398078Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5398293Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5398505Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5398720Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5398846Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5398855Z 2023-03-31T06:22:41.5398862Z 2023-03-31T06:22:41.5398940Z if __name__ == "__main__": 2023-03-31T06:22:41.5399017Z import argparse 2023-03-31T06:22:41.5399147Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5399152Z 2023-03-31T06:22:41.5399250Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5399557Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5399954Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5400228Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5400321Z args = parser.parse_args() 2023-03-31T06:22:41.5400326Z 2023-03-31T06:22:41.5400416Z if args.benchmark_kernels: 2023-03-31T06:22:41.5400610Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5400681Z else: 2023-03-31T06:22:41.5400795Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5400877Z benchmark_compiled_module() 2023-03-31T06:22:41.5400882Z 2023-03-31T06:22:41.5400951Z if p: 2023-03-31T06:22:41.5401106Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5401204Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5401375Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5401380Z 2023-03-31T06:22:41.5401665Z [2023-03-31 06:15:56,906] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 125 2023-03-31T06:22:41.5401936Z [2023-03-31 06:15:56,959] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 126 2023-03-31T06:22:41.5402352Z [2023-03-31 06:16:18,118] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/bh/cbh4qssxnmiqcyo2hc3syucibbkvpyrtfuochlz22rsh66sgoytu.py 2023-03-31T06:22:41.5402561Z [2023-03-31 06:16:18,118] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5402580Z 2023-03-31T06:22:41.5402665Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5402739Z import torch 2023-03-31T06:22:41.5402813Z import math 2023-03-31T06:22:41.5402888Z import random 2023-03-31T06:22:41.5402958Z import os 2023-03-31T06:22:41.5403161Z import tempfile 2023-03-31T06:22:41.5403267Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5403286Z 2023-03-31T06:22:41.5403395Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5403521Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5403655Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5403660Z 2023-03-31T06:22:41.5403748Z aten = torch.ops.aten 2023-03-31T06:22:41.5403888Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5404032Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5404069Z 2023-03-31T06:22:41.5404162Z async_compile.wait(globals()) 2023-03-31T06:22:41.5404223Z del async_compile 2023-03-31T06:22:41.5404346Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5404420Z wrapper = ( 2023-03-31T06:22:41.5404508Z ''' 2023-03-31T06:22:41.5404586Z #include 2023-03-31T06:22:41.5404665Z #include 2023-03-31T06:22:41.5404673Z 2023-03-31T06:22:41.5404767Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5404772Z 2023-03-31T06:22:41.5404850Z template 2023-03-31T06:22:41.5404977Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5405064Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5405192Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5405289Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5405416Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5405499Z return kernel_cpp; 2023-03-31T06:22:41.5405551Z } 2023-03-31T06:22:41.5405683Z std::vector call_150(std::vector args) { 2023-03-31T06:22:41.5405789Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5405865Z arg0_1 = args[0]; 2023-03-31T06:22:41.5405940Z arg1_1 = args[1]; 2023-03-31T06:22:41.5406012Z arg2_1 = args[2]; 2023-03-31T06:22:41.5406084Z arg3_1 = args[3]; 2023-03-31T06:22:41.5406136Z 2023-03-31T06:22:41.5406234Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5406329Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5406456Z .findSchemaOrThrow( 2023-03-31T06:22:41.5406555Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5406623Z "") 2023-03-31T06:22:41.5406693Z .typed< 2023-03-31T06:22:41.5406755Z at::Tensor( 2023-03-31T06:22:41.5406855Z const at::Tensor& input_t, 2023-03-31T06:22:41.5406955Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5407073Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5407165Z c10::string_view attr, 2023-03-31T06:22:41.5407285Z torch::List> scalars, 2023-03-31T06:22:41.5407406Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5407470Z 2023-03-31T06:22:41.5407688Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5407808Z arg1_1.reset(); 2023-03-31T06:22:41.5407881Z arg2_1.reset(); 2023-03-31T06:22:41.5407952Z arg3_1.reset(); 2023-03-31T06:22:41.5408059Z return std::vector({buf0}); 2023-03-31T06:22:41.5408121Z } 2023-03-31T06:22:41.5408202Z ''' 2023-03-31T06:22:41.5408251Z ) 2023-03-31T06:22:41.5408257Z 2023-03-31T06:22:41.5408333Z module = load_inline( 2023-03-31T06:22:41.5408604Z name='inline_extension_cn4dsh6bhfjzkax5vppyux7rbj2fhpzjkt4xlzu47xndtscy46pz', 2023-03-31T06:22:41.5408692Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5408812Z functions=['call_150'], 2023-03-31T06:22:41.5409191Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5409348Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5410059Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5410069Z 2023-03-31T06:22:41.5410145Z def _wrap_func(f): 2023-03-31T06:22:41.5410204Z def g(args): 2023-03-31T06:22:41.5410280Z return f(args) 2023-03-31T06:22:41.5410348Z return g 2023-03-31T06:22:41.5410443Z call = _wrap_func(module.call_150) 2023-03-31T06:22:41.5410475Z 2023-03-31T06:22:41.5410515Z 2023-03-31T06:22:41.5410611Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5410731Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5410860Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5411064Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5411270Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5411479Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5411684Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5411818Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5411823Z 2023-03-31T06:22:41.5411828Z 2023-03-31T06:22:41.5411906Z if __name__ == "__main__": 2023-03-31T06:22:41.5411982Z import argparse 2023-03-31T06:22:41.5412119Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5412127Z 2023-03-31T06:22:41.5412224Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5412536Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5412873Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5413156Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5413277Z args = parser.parse_args() 2023-03-31T06:22:41.5413283Z 2023-03-31T06:22:41.5413375Z if args.benchmark_kernels: 2023-03-31T06:22:41.5413564Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5413631Z else: 2023-03-31T06:22:41.5413743Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5413828Z benchmark_compiled_module() 2023-03-31T06:22:41.5413835Z 2023-03-31T06:22:41.5413901Z if p: 2023-03-31T06:22:41.5414052Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5414148Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5414284Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5414290Z 2023-03-31T06:22:41.5414570Z [2023-03-31 06:16:18,119] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 126 2023-03-31T06:22:41.5414878Z [2023-03-31 06:16:18,159] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 127 2023-03-31T06:22:41.5415283Z [2023-03-31 06:16:38,569] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/z3/cz35zudikwyk7uqcff46dyurrgtoc63zaacfbnqng4ehytokvyrk.py 2023-03-31T06:22:41.5415489Z [2023-03-31 06:16:38,569] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5415506Z 2023-03-31T06:22:41.5415594Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5415668Z import torch 2023-03-31T06:22:41.5415739Z import math 2023-03-31T06:22:41.5415811Z import random 2023-03-31T06:22:41.5415878Z import os 2023-03-31T06:22:41.5415951Z import tempfile 2023-03-31T06:22:41.5416056Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5416073Z 2023-03-31T06:22:41.5416180Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5416307Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5416441Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5416449Z 2023-03-31T06:22:41.5416529Z aten = torch.ops.aten 2023-03-31T06:22:41.5416668Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5416761Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5416766Z 2023-03-31T06:22:41.5416857Z async_compile.wait(globals()) 2023-03-31T06:22:41.5416918Z del async_compile 2023-03-31T06:22:41.5417067Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5417166Z wrapper = ( 2023-03-31T06:22:41.5417249Z ''' 2023-03-31T06:22:41.5417323Z #include 2023-03-31T06:22:41.5417398Z #include 2023-03-31T06:22:41.5417404Z 2023-03-31T06:22:41.5417494Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5417499Z 2023-03-31T06:22:41.5417578Z template 2023-03-31T06:22:41.5417703Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5417788Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5417911Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5418011Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5418136Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5418214Z return kernel_cpp; 2023-03-31T06:22:41.5418264Z } 2023-03-31T06:22:41.5418392Z std::vector call_151(std::vector args) { 2023-03-31T06:22:41.5418484Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5418560Z arg0_1 = args[0]; 2023-03-31T06:22:41.5418633Z arg1_1 = args[1]; 2023-03-31T06:22:41.5418704Z arg2_1 = args[2]; 2023-03-31T06:22:41.5418765Z 2023-03-31T06:22:41.5418850Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5418942Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5419027Z .findSchemaOrThrow( 2023-03-31T06:22:41.5419129Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5419196Z "") 2023-03-31T06:22:41.5419263Z .typed< 2023-03-31T06:22:41.5419337Z at::Tensor( 2023-03-31T06:22:41.5419454Z const at::Tensor& input_t, 2023-03-31T06:22:41.5419555Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5419672Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5419760Z c10::string_view attr, 2023-03-31T06:22:41.5419881Z torch::List> scalars, 2023-03-31T06:22:41.5420003Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5420070Z 2023-03-31T06:22:41.5420312Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "sigmoid", {-1}, ""); 2023-03-31T06:22:41.5420373Z arg1_1.reset(); 2023-03-31T06:22:41.5420443Z arg2_1.reset(); 2023-03-31T06:22:41.5420547Z return std::vector({buf0}); 2023-03-31T06:22:41.5420611Z } 2023-03-31T06:22:41.5420716Z ''' 2023-03-31T06:22:41.5420780Z ) 2023-03-31T06:22:41.5420785Z 2023-03-31T06:22:41.5420862Z module = load_inline( 2023-03-31T06:22:41.5421116Z name='inline_extension_c3d4rk64txv2cnq2ea4t72byofnddnbhotqdbt4ypob4sohdxlnz', 2023-03-31T06:22:41.5421203Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5421320Z functions=['call_151'], 2023-03-31T06:22:41.5421705Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5421864Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5422657Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5422669Z 2023-03-31T06:22:41.5422747Z def _wrap_func(f): 2023-03-31T06:22:41.5422821Z def g(args): 2023-03-31T06:22:41.5422897Z return f(args) 2023-03-31T06:22:41.5422954Z return g 2023-03-31T06:22:41.5423051Z call = _wrap_func(module.call_151) 2023-03-31T06:22:41.5423057Z 2023-03-31T06:22:41.5423061Z 2023-03-31T06:22:41.5423157Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5423277Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5423404Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5423655Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5423891Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5424096Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5424210Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5424216Z 2023-03-31T06:22:41.5424224Z 2023-03-31T06:22:41.5424300Z if __name__ == "__main__": 2023-03-31T06:22:41.5424376Z import argparse 2023-03-31T06:22:41.5424507Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5424512Z 2023-03-31T06:22:41.5424621Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5424928Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5425271Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5425548Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5425628Z args = parser.parse_args() 2023-03-31T06:22:41.5425645Z 2023-03-31T06:22:41.5425722Z if args.benchmark_kernels: 2023-03-31T06:22:41.5425908Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5425977Z else: 2023-03-31T06:22:41.5426091Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5426185Z benchmark_compiled_module() 2023-03-31T06:22:41.5426218Z 2023-03-31T06:22:41.5426287Z if p: 2023-03-31T06:22:41.5426436Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5426519Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5426657Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5426663Z 2023-03-31T06:22:41.5426948Z [2023-03-31 06:16:38,570] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 127 2023-03-31T06:22:41.5427219Z [2023-03-31 06:16:38,622] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 128 2023-03-31T06:22:41.5427615Z [2023-03-31 06:16:59,320] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/xb/cxbrbk4xrgg22agt4xfn553rsu67263dwdoar6f3arg37xddtxv5.py 2023-03-31T06:22:41.5427863Z [2023-03-31 06:16:59,320] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5427869Z 2023-03-31T06:22:41.5427968Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5428040Z import torch 2023-03-31T06:22:41.5428111Z import math 2023-03-31T06:22:41.5428170Z import random 2023-03-31T06:22:41.5428236Z import os 2023-03-31T06:22:41.5428309Z import tempfile 2023-03-31T06:22:41.5428426Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5428432Z 2023-03-31T06:22:41.5428553Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5428682Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5428816Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5428822Z 2023-03-31T06:22:41.5428890Z aten = torch.ops.aten 2023-03-31T06:22:41.5429027Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5429121Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5429129Z 2023-03-31T06:22:41.5429221Z async_compile.wait(globals()) 2023-03-31T06:22:41.5429295Z del async_compile 2023-03-31T06:22:41.5429418Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5429491Z wrapper = ( 2023-03-31T06:22:41.5429559Z ''' 2023-03-31T06:22:41.5429633Z #include 2023-03-31T06:22:41.5429708Z #include 2023-03-31T06:22:41.5429713Z 2023-03-31T06:22:41.5429807Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5429813Z 2023-03-31T06:22:41.5429904Z template 2023-03-31T06:22:41.5430054Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5430165Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5430289Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5430375Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5430501Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5430579Z return kernel_cpp; 2023-03-31T06:22:41.5430645Z } 2023-03-31T06:22:41.5430775Z std::vector call_152(std::vector args) { 2023-03-31T06:22:41.5430880Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5430956Z arg0_1 = args[0]; 2023-03-31T06:22:41.5431016Z arg1_1 = args[1]; 2023-03-31T06:22:41.5431086Z arg2_1 = args[2]; 2023-03-31T06:22:41.5431155Z arg3_1 = args[3]; 2023-03-31T06:22:41.5431217Z 2023-03-31T06:22:41.5431315Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5431407Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5431484Z .findSchemaOrThrow( 2023-03-31T06:22:41.5431584Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5431651Z "") 2023-03-31T06:22:41.5431716Z .typed< 2023-03-31T06:22:41.5431791Z at::Tensor( 2023-03-31T06:22:41.5431887Z const at::Tensor& input_t, 2023-03-31T06:22:41.5431987Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5432104Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5432181Z c10::string_view attr, 2023-03-31T06:22:41.5432301Z torch::List> scalars, 2023-03-31T06:22:41.5432453Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5432520Z 2023-03-31T06:22:41.5432739Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.5432811Z arg1_1.reset(); 2023-03-31T06:22:41.5432882Z arg2_1.reset(); 2023-03-31T06:22:41.5432939Z arg3_1.reset(); 2023-03-31T06:22:41.5433046Z return std::vector({buf0}); 2023-03-31T06:22:41.5433111Z } 2023-03-31T06:22:41.5433189Z ''' 2023-03-31T06:22:41.5433250Z ) 2023-03-31T06:22:41.5433256Z 2023-03-31T06:22:41.5433333Z module = load_inline( 2023-03-31T06:22:41.5433597Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.5433669Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5433828Z functions=['call_152'], 2023-03-31T06:22:41.5434210Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5434364Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5435083Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5435092Z 2023-03-31T06:22:41.5435169Z def _wrap_func(f): 2023-03-31T06:22:41.5435240Z def g(args): 2023-03-31T06:22:41.5435315Z return f(args) 2023-03-31T06:22:41.5435369Z return g 2023-03-31T06:22:41.5435464Z call = _wrap_func(module.call_152) 2023-03-31T06:22:41.5435470Z 2023-03-31T06:22:41.5435477Z 2023-03-31T06:22:41.5435571Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5435690Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5435823Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5436038Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5436247Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5436453Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5436698Z arg3_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5436857Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5436863Z 2023-03-31T06:22:41.5436868Z 2023-03-31T06:22:41.5436947Z if __name__ == "__main__": 2023-03-31T06:22:41.5437023Z import argparse 2023-03-31T06:22:41.5437152Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5437161Z 2023-03-31T06:22:41.5437268Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5437576Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5437914Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5438188Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5438271Z args = parser.parse_args() 2023-03-31T06:22:41.5438290Z 2023-03-31T06:22:41.5438366Z if args.benchmark_kernels: 2023-03-31T06:22:41.5438556Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5438625Z else: 2023-03-31T06:22:41.5438739Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5438836Z benchmark_compiled_module() 2023-03-31T06:22:41.5438844Z 2023-03-31T06:22:41.5438914Z if p: 2023-03-31T06:22:41.5439066Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5439178Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5439319Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5439325Z 2023-03-31T06:22:41.5439608Z [2023-03-31 06:16:59,320] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 128 2023-03-31T06:22:41.5439881Z [2023-03-31 06:16:59,360] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 129 2023-03-31T06:22:41.5440292Z [2023-03-31 06:17:21,083] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/js/cjswudxebzrf4kmg42mjjfvj37zxxdkxpst4ntzrxmv4jegipv6h.py 2023-03-31T06:22:41.5440515Z [2023-03-31 06:17:21,083] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5440521Z 2023-03-31T06:22:41.5440650Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5440724Z import torch 2023-03-31T06:22:41.5440797Z import math 2023-03-31T06:22:41.5440856Z import random 2023-03-31T06:22:41.5440926Z import os 2023-03-31T06:22:41.5441004Z import tempfile 2023-03-31T06:22:41.5441123Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5441128Z 2023-03-31T06:22:41.5441250Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5441378Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5441513Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5441521Z 2023-03-31T06:22:41.5441591Z aten = torch.ops.aten 2023-03-31T06:22:41.5441730Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5441825Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5441830Z 2023-03-31T06:22:41.5441923Z async_compile.wait(globals()) 2023-03-31T06:22:41.5442000Z del async_compile 2023-03-31T06:22:41.5442122Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5442199Z wrapper = ( 2023-03-31T06:22:41.5442269Z ''' 2023-03-31T06:22:41.5442347Z #include 2023-03-31T06:22:41.5442424Z #include 2023-03-31T06:22:41.5442432Z 2023-03-31T06:22:41.5442525Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5442531Z 2023-03-31T06:22:41.5442625Z template 2023-03-31T06:22:41.5442748Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5442837Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5442964Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5443233Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5443395Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5443476Z return kernel_cpp; 2023-03-31T06:22:41.5443541Z } 2023-03-31T06:22:41.5443673Z std::vector call_153(std::vector args) { 2023-03-31T06:22:41.5443768Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5443846Z arg0_1 = args[0]; 2023-03-31T06:22:41.5443907Z arg1_1 = args[1]; 2023-03-31T06:22:41.5443979Z arg2_1 = args[2]; 2023-03-31T06:22:41.5444042Z 2023-03-31T06:22:41.5444145Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5444238Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5444328Z .findSchemaOrThrow( 2023-03-31T06:22:41.5444414Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5444483Z "") 2023-03-31T06:22:41.5444550Z .typed< 2023-03-31T06:22:41.5444627Z at::Tensor( 2023-03-31T06:22:41.5444727Z const at::Tensor& input_t, 2023-03-31T06:22:41.5444831Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5444948Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5445035Z c10::string_view attr, 2023-03-31T06:22:41.5445144Z torch::List> scalars, 2023-03-31T06:22:41.5445266Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5445334Z 2023-03-31T06:22:41.5445573Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.5445685Z arg1_1.reset(); 2023-03-31T06:22:41.5445756Z arg2_1.reset(); 2023-03-31T06:22:41.5445861Z return std::vector({buf0}); 2023-03-31T06:22:41.5445911Z } 2023-03-31T06:22:41.5445992Z ''' 2023-03-31T06:22:41.5446053Z ) 2023-03-31T06:22:41.5446061Z 2023-03-31T06:22:41.5446138Z module = load_inline( 2023-03-31T06:22:41.5446409Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.5446497Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5446615Z functions=['call_153'], 2023-03-31T06:22:41.5446986Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5447141Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5447887Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5447894Z 2023-03-31T06:22:41.5447969Z def _wrap_func(f): 2023-03-31T06:22:41.5448039Z def g(args): 2023-03-31T06:22:41.5448112Z return f(args) 2023-03-31T06:22:41.5448182Z return g 2023-03-31T06:22:41.5448279Z call = _wrap_func(module.call_153) 2023-03-31T06:22:41.5448285Z 2023-03-31T06:22:41.5448290Z 2023-03-31T06:22:41.5448386Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5448492Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5448618Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5448831Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5449042Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5449260Z arg2_1 = rand_strided({2, 3, 10}, {30, 10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5449385Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5449392Z 2023-03-31T06:22:41.5449396Z 2023-03-31T06:22:41.5449473Z if __name__ == "__main__": 2023-03-31T06:22:41.5449549Z import argparse 2023-03-31T06:22:41.5449693Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5449760Z 2023-03-31T06:22:41.5449859Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5450166Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5450503Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5450784Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5450878Z args = parser.parse_args() 2023-03-31T06:22:41.5450883Z 2023-03-31T06:22:41.5450975Z if args.benchmark_kernels: 2023-03-31T06:22:41.5451164Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5451230Z else: 2023-03-31T06:22:41.5451330Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5451423Z benchmark_compiled_module() 2023-03-31T06:22:41.5451431Z 2023-03-31T06:22:41.5451502Z if p: 2023-03-31T06:22:41.5451652Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5451748Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5451886Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5451892Z 2023-03-31T06:22:41.5452171Z [2023-03-31 06:17:21,083] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 129 2023-03-31T06:22:41.5452464Z [2023-03-31 06:17:21,136] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 130 2023-03-31T06:22:41.5452845Z [2023-03-31 06:17:42,145] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/zj/czj5blyc672jfkaegeq3gz265rmk53y4xe3wqoixbehm2qzqztv2.py 2023-03-31T06:22:41.5453063Z [2023-03-31 06:17:42,145] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5453069Z 2023-03-31T06:22:41.5453168Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5453242Z import torch 2023-03-31T06:22:41.5453313Z import math 2023-03-31T06:22:41.5453385Z import random 2023-03-31T06:22:41.5453453Z import os 2023-03-31T06:22:41.5453527Z import tempfile 2023-03-31T06:22:41.5453630Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5453636Z 2023-03-31T06:22:41.5453753Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5453906Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5454043Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5454051Z 2023-03-31T06:22:41.5454130Z aten = torch.ops.aten 2023-03-31T06:22:41.5454270Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5454364Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5454369Z 2023-03-31T06:22:41.5454459Z async_compile.wait(globals()) 2023-03-31T06:22:41.5454521Z del async_compile 2023-03-31T06:22:41.5454644Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5454720Z wrapper = ( 2023-03-31T06:22:41.5454802Z ''' 2023-03-31T06:22:41.5454877Z #include 2023-03-31T06:22:41.5454952Z #include 2023-03-31T06:22:41.5454957Z 2023-03-31T06:22:41.5455036Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5455053Z 2023-03-31T06:22:41.5455132Z template 2023-03-31T06:22:41.5455254Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5455342Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5455466Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5455564Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5455690Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5455768Z return kernel_cpp; 2023-03-31T06:22:41.5455817Z } 2023-03-31T06:22:41.5455945Z std::vector call_154(std::vector args) { 2023-03-31T06:22:41.5456048Z at::Tensor arg0_1, arg1_1, arg2_1, arg3_1; 2023-03-31T06:22:41.5456147Z arg0_1 = args[0]; 2023-03-31T06:22:41.5456258Z arg1_1 = args[1]; 2023-03-31T06:22:41.5456328Z arg2_1 = args[2]; 2023-03-31T06:22:41.5456399Z arg3_1 = args[3]; 2023-03-31T06:22:41.5456448Z 2023-03-31T06:22:41.5456549Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5456641Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5456727Z .findSchemaOrThrow( 2023-03-31T06:22:41.5456827Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5456894Z "") 2023-03-31T06:22:41.5456950Z .typed< 2023-03-31T06:22:41.5457025Z at::Tensor( 2023-03-31T06:22:41.5457122Z const at::Tensor& input_t, 2023-03-31T06:22:41.5457221Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5457337Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5457425Z c10::string_view attr, 2023-03-31T06:22:41.5457549Z torch::List> scalars, 2023-03-31T06:22:41.5457673Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5457727Z 2023-03-31T06:22:41.5457948Z auto buf0 = op_linear_pointwise.call(arg3_1, arg2_1, arg1_1, "tanh", {-1}, ""); 2023-03-31T06:22:41.5458022Z arg1_1.reset(); 2023-03-31T06:22:41.5458094Z arg2_1.reset(); 2023-03-31T06:22:41.5458163Z arg3_1.reset(); 2023-03-31T06:22:41.5458272Z return std::vector({buf0}); 2023-03-31T06:22:41.5458333Z } 2023-03-31T06:22:41.5458400Z ''' 2023-03-31T06:22:41.5458459Z ) 2023-03-31T06:22:41.5458465Z 2023-03-31T06:22:41.5458572Z module = load_inline( 2023-03-31T06:22:41.5458838Z name='inline_extension_cehuv5cu6biht5ds4tgr4b37pxnyj33sorso6odsfyzbk5cwujvx', 2023-03-31T06:22:41.5458924Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5459040Z functions=['call_154'], 2023-03-31T06:22:41.5459418Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5459562Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5460269Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5460311Z 2023-03-31T06:22:41.5460375Z def _wrap_func(f): 2023-03-31T06:22:41.5460449Z def g(args): 2023-03-31T06:22:41.5460524Z return f(args) 2023-03-31T06:22:41.5460591Z return g 2023-03-31T06:22:41.5460685Z call = _wrap_func(module.call_154) 2023-03-31T06:22:41.5460690Z 2023-03-31T06:22:41.5460695Z 2023-03-31T06:22:41.5460789Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5460905Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5461022Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5461238Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5461445Z arg1_1 = rand_strided({30, }, {1, }, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5461651Z arg2_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5461854Z arg3_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5461991Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1, arg3_1])) 2023-03-31T06:22:41.5461997Z 2023-03-31T06:22:41.5462004Z 2023-03-31T06:22:41.5462082Z if __name__ == "__main__": 2023-03-31T06:22:41.5462157Z import argparse 2023-03-31T06:22:41.5462275Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5462293Z 2023-03-31T06:22:41.5462474Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5462827Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5463190Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5463470Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5463562Z args = parser.parse_args() 2023-03-31T06:22:41.5463570Z 2023-03-31T06:22:41.5463660Z if args.benchmark_kernels: 2023-03-31T06:22:41.5463854Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5463921Z else: 2023-03-31T06:22:41.5464020Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5464115Z benchmark_compiled_module() 2023-03-31T06:22:41.5464121Z 2023-03-31T06:22:41.5464188Z if p: 2023-03-31T06:22:41.5464336Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5464436Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5464576Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5464582Z 2023-03-31T06:22:41.5464867Z [2023-03-31 06:17:42,145] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 130 2023-03-31T06:22:41.5465136Z [2023-03-31 06:17:42,183] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 131 2023-03-31T06:22:41.5465559Z [2023-03-31 06:18:02,913] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ee/ceeetpggqypcou25ovtpz34eo2kfoqmoxccgbt25bb5sp2ghaikg.py 2023-03-31T06:22:41.5465778Z [2023-03-31 06:18:02,913] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5465784Z 2023-03-31T06:22:41.5465880Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5465952Z import torch 2023-03-31T06:22:41.5466024Z import math 2023-03-31T06:22:41.5466096Z import random 2023-03-31T06:22:41.5466163Z import os 2023-03-31T06:22:41.5466242Z import tempfile 2023-03-31T06:22:41.5466351Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5466357Z 2023-03-31T06:22:41.5466474Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5466599Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5466732Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5466765Z 2023-03-31T06:22:41.5466847Z aten = torch.ops.aten 2023-03-31T06:22:41.5466986Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5467083Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5467088Z 2023-03-31T06:22:41.5467180Z async_compile.wait(globals()) 2023-03-31T06:22:41.5467242Z del async_compile 2023-03-31T06:22:41.5467363Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5467435Z wrapper = ( 2023-03-31T06:22:41.5467519Z ''' 2023-03-31T06:22:41.5467597Z #include 2023-03-31T06:22:41.5467671Z #include 2023-03-31T06:22:41.5467679Z 2023-03-31T06:22:41.5467760Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5467777Z 2023-03-31T06:22:41.5467857Z template 2023-03-31T06:22:41.5467980Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5468070Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5468194Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5468293Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5468416Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5468496Z return kernel_cpp; 2023-03-31T06:22:41.5468548Z } 2023-03-31T06:22:41.5468678Z std::vector call_155(std::vector args) { 2023-03-31T06:22:41.5468770Z at::Tensor arg0_1, arg1_1, arg2_1; 2023-03-31T06:22:41.5468844Z arg0_1 = args[0]; 2023-03-31T06:22:41.5468915Z arg1_1 = args[1]; 2023-03-31T06:22:41.5468989Z arg2_1 = args[2]; 2023-03-31T06:22:41.5469050Z 2023-03-31T06:22:41.5469162Z static auto op_linear_pointwise = 2023-03-31T06:22:41.5469280Z c10::Dispatcher::singleton() 2023-03-31T06:22:41.5469368Z .findSchemaOrThrow( 2023-03-31T06:22:41.5469472Z "mkldnn::_linear_pointwise", 2023-03-31T06:22:41.5469539Z "") 2023-03-31T06:22:41.5469607Z .typed< 2023-03-31T06:22:41.5469670Z at::Tensor( 2023-03-31T06:22:41.5469772Z const at::Tensor& input_t, 2023-03-31T06:22:41.5469877Z const at::Tensor& weight_t, 2023-03-31T06:22:41.5469994Z const c10::optional& bias_opt, 2023-03-31T06:22:41.5470086Z c10::string_view attr, 2023-03-31T06:22:41.5470207Z torch::List> scalars, 2023-03-31T06:22:41.5470327Z c10::optional algorithm)>(); 2023-03-31T06:22:41.5470392Z 2023-03-31T06:22:41.5470613Z auto buf0 = op_linear_pointwise.call(arg2_1, arg1_1, at::Tensor(), "tanh", {-1}, ""); 2023-03-31T06:22:41.5470687Z arg1_1.reset(); 2023-03-31T06:22:41.5470760Z arg2_1.reset(); 2023-03-31T06:22:41.5470864Z return std::vector({buf0}); 2023-03-31T06:22:41.5470927Z } 2023-03-31T06:22:41.5471008Z ''' 2023-03-31T06:22:41.5471070Z ) 2023-03-31T06:22:41.5471075Z 2023-03-31T06:22:41.5471140Z module = load_inline( 2023-03-31T06:22:41.5471407Z name='inline_extension_c4jjw6ebp3ztkhubzmygathkb6bd25it53fmyxh65ibpeyavf6wu', 2023-03-31T06:22:41.5471496Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5471614Z functions=['call_155'], 2023-03-31T06:22:41.5472026Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5472184Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5472895Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5472904Z 2023-03-31T06:22:41.5472979Z def _wrap_func(f): 2023-03-31T06:22:41.5473049Z def g(args): 2023-03-31T06:22:41.5473111Z return f(args) 2023-03-31T06:22:41.5473178Z return g 2023-03-31T06:22:41.5473307Z call = _wrap_func(module.call_155) 2023-03-31T06:22:41.5473312Z 2023-03-31T06:22:41.5473317Z 2023-03-31T06:22:41.5473411Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5473531Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5473657Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5473875Z arg0_1 = rand_strided({30, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5474086Z arg1_1 = rand_strided({30, 10}, {1, 0}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5474283Z arg2_1 = rand_strided({2, 10}, {10, 1}, device='cpu', dtype=torch.bfloat16) 2023-03-31T06:22:41.5474415Z print_performance(lambda: call([arg0_1, arg1_1, arg2_1])) 2023-03-31T06:22:41.5474420Z 2023-03-31T06:22:41.5474425Z 2023-03-31T06:22:41.5474500Z if __name__ == "__main__": 2023-03-31T06:22:41.5474576Z import argparse 2023-03-31T06:22:41.5474705Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5474713Z 2023-03-31T06:22:41.5474824Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5475134Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5475474Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5475751Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5475857Z args = parser.parse_args() 2023-03-31T06:22:41.5475884Z 2023-03-31T06:22:41.5475974Z if args.benchmark_kernels: 2023-03-31T06:22:41.5476164Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5476231Z else: 2023-03-31T06:22:41.5476344Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5476443Z benchmark_compiled_module() 2023-03-31T06:22:41.5476448Z 2023-03-31T06:22:41.5476519Z if p: 2023-03-31T06:22:41.5476656Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5476757Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5476894Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5476899Z 2023-03-31T06:22:41.5477184Z [2023-03-31 06:18:02,914] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 131 2023-03-31T06:22:41.5477261Z PASSED [ 50%] 2023-03-31T06:22:41.5477670Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_lowmem_dropout1_cpu [2023-03-31 06:18:02,953] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 132 2023-03-31T06:22:41.5478078Z [2023-03-31 06:18:24,389] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/rl/crlwi6bcswlydcrod5y4otwenld44cuy7cl4wf6utzcx4kgjuwso.py 2023-03-31T06:22:41.5478300Z [2023-03-31 06:18:24,389] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5478308Z 2023-03-31T06:22:41.5478407Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5478467Z import torch 2023-03-31T06:22:41.5478570Z import math 2023-03-31T06:22:41.5478646Z import random 2023-03-31T06:22:41.5478715Z import os 2023-03-31T06:22:41.5478792Z import tempfile 2023-03-31T06:22:41.5478911Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5478917Z 2023-03-31T06:22:41.5479038Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5479150Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5479287Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5479295Z 2023-03-31T06:22:41.5479376Z aten = torch.ops.aten 2023-03-31T06:22:41.5479517Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5479615Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5479620Z 2023-03-31T06:22:41.5479624Z 2023-03-31T06:22:41.5479767Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5480025Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5480139Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5480223Z const float* in_ptr1, 2023-03-31T06:22:41.5480311Z float* out_ptr0) 2023-03-31T06:22:41.5480375Z { 2023-03-31T06:22:41.5480477Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5480542Z { 2023-03-31T06:22:41.5480611Z { 2023-03-31T06:22:41.5480698Z #pragma omp for 2023-03-31T06:22:41.5480843Z for(long i0=static_cast(0); i0(12500); i0+=static_cast(1)) 2023-03-31T06:22:41.5480914Z { 2023-03-31T06:22:41.5481077Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5481235Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5481332Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5481453Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5481524Z } 2023-03-31T06:22:41.5481627Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5481771Z for(long i0=static_cast(100000); i0(100000); i0+=static_cast(1)) 2023-03-31T06:22:41.5481839Z { 2023-03-31T06:22:41.5481955Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5482099Z auto tmp1 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5482218Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5482329Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.5482396Z } 2023-03-31T06:22:41.5482450Z } 2023-03-31T06:22:41.5482514Z } 2023-03-31T06:22:41.5482575Z } 2023-03-31T06:22:41.5482662Z ''') 2023-03-31T06:22:41.5482668Z 2023-03-31T06:22:41.5482809Z async_compile.wait(globals()) 2023-03-31T06:22:41.5482917Z del async_compile 2023-03-31T06:22:41.5483203Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5483265Z wrapper = ( 2023-03-31T06:22:41.5483503Z ''' 2023-03-31T06:22:41.5483606Z #include 2023-03-31T06:22:41.5483749Z #include 2023-03-31T06:22:41.5483755Z 2023-03-31T06:22:41.5483878Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5483883Z 2023-03-31T06:22:41.5483964Z template 2023-03-31T06:22:41.5488611Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5488719Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5488856Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5488958Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5489086Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5489168Z return kernel_cpp; 2023-03-31T06:22:41.5489232Z } 2023-03-31T06:22:41.5489373Z std::vector call_156(std::vector args) { 2023-03-31T06:22:41.5489464Z at::Tensor primals_1, primals_2; 2023-03-31T06:22:41.5489546Z primals_1 = args[0]; 2023-03-31T06:22:41.5489718Z primals_2 = args[1]; 2023-03-31T06:22:41.5489860Z auto buf0 = at::empty_strided({100000, }, {1, }, at::ScalarType::Float); 2023-03-31T06:22:41.5490174Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/pc/cpcmjer6uoj4fiv5tu3frpax6dqw2mu54ynzoyia7i57hix72so6.so"); 2023-03-31T06:22:41.5490358Z kernel_cpp_0((float*)(primals_1.data_ptr()), (float*)(primals_2.data_ptr()), (float*)(buf0.data_ptr())); 2023-03-31T06:22:41.5490441Z primals_2.reset(); 2023-03-31T06:22:41.5490567Z return std::vector({buf0, primals_1}); 2023-03-31T06:22:41.5490618Z } 2023-03-31T06:22:41.5490725Z ''' 2023-03-31T06:22:41.5490788Z ) 2023-03-31T06:22:41.5490795Z 2023-03-31T06:22:41.5490874Z module = load_inline( 2023-03-31T06:22:41.5491196Z name='inline_extension_cuhcantr3d2h3vad7elwb7u3kgwfisu3juy3t7d7ttpo5jmboh6v', 2023-03-31T06:22:41.5491285Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5491407Z functions=['call_156'], 2023-03-31T06:22:41.5491788Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5491948Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5492657Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5492666Z 2023-03-31T06:22:41.5492745Z def _wrap_func(f): 2023-03-31T06:22:41.5492817Z def g(args): 2023-03-31T06:22:41.5492894Z return f(args) 2023-03-31T06:22:41.5492967Z return g 2023-03-31T06:22:41.5493065Z call = _wrap_func(module.call_156) 2023-03-31T06:22:41.5493071Z 2023-03-31T06:22:41.5493076Z 2023-03-31T06:22:41.5493174Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5493282Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5493411Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5493631Z primals_1 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5493893Z primals_2 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5494057Z print_performance(lambda: call([primals_1, primals_2])) 2023-03-31T06:22:41.5494063Z 2023-03-31T06:22:41.5494068Z 2023-03-31T06:22:41.5494147Z if __name__ == "__main__": 2023-03-31T06:22:41.5494226Z import argparse 2023-03-31T06:22:41.5494357Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5494363Z 2023-03-31T06:22:41.5494464Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5494781Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5495124Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5495408Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5495502Z args = parser.parse_args() 2023-03-31T06:22:41.5495507Z 2023-03-31T06:22:41.5495597Z if args.benchmark_kernels: 2023-03-31T06:22:41.5495791Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5495861Z else: 2023-03-31T06:22:41.5495975Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5496057Z benchmark_compiled_module() 2023-03-31T06:22:41.5496063Z 2023-03-31T06:22:41.5496130Z if p: 2023-03-31T06:22:41.5496284Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5496382Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5496548Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5496555Z 2023-03-31T06:22:41.5496841Z [2023-03-31 06:18:24,389] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 132 2023-03-31T06:22:41.5497112Z [2023-03-31 06:18:24,394] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling BACKWARDS graph 132 2023-03-31T06:22:41.5497521Z [2023-03-31 06:18:44,634] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/jk/cjkggqt3kogac6psrgnx2ramshgcqcmrg57fj57esogkivvflpqj.py 2023-03-31T06:22:41.5497729Z [2023-03-31 06:18:44,635] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5497746Z 2023-03-31T06:22:41.5497832Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5497906Z import torch 2023-03-31T06:22:41.5498005Z import math 2023-03-31T06:22:41.5498078Z import random 2023-03-31T06:22:41.5498145Z import os 2023-03-31T06:22:41.5498218Z import tempfile 2023-03-31T06:22:41.5498325Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5498342Z 2023-03-31T06:22:41.5498449Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5498572Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5498705Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5498711Z 2023-03-31T06:22:41.5498794Z aten = torch.ops.aten 2023-03-31T06:22:41.5498937Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5499031Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5499036Z 2023-03-31T06:22:41.5499041Z 2023-03-31T06:22:41.5499179Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5499401Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5499501Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5499598Z const float* in_ptr1, 2023-03-31T06:22:41.5499685Z float* out_ptr0) 2023-03-31T06:22:41.5499748Z { 2023-03-31T06:22:41.5499847Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5499910Z { 2023-03-31T06:22:41.5499974Z { 2023-03-31T06:22:41.5500043Z #pragma omp for 2023-03-31T06:22:41.5500199Z for(long i0=static_cast(0); i0(12500); i0+=static_cast(1)) 2023-03-31T06:22:41.5500296Z { 2023-03-31T06:22:41.5500485Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5500643Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5500734Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5500854Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5500910Z } 2023-03-31T06:22:41.5501011Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5501171Z for(long i0=static_cast(100000); i0(100000); i0+=static_cast(1)) 2023-03-31T06:22:41.5501238Z { 2023-03-31T06:22:41.5501352Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5501465Z auto tmp1 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5501555Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5501666Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.5501721Z } 2023-03-31T06:22:41.5501785Z } 2023-03-31T06:22:41.5501847Z } 2023-03-31T06:22:41.5501910Z } 2023-03-31T06:22:41.5501993Z ''') 2023-03-31T06:22:41.5501999Z 2023-03-31T06:22:41.5502091Z async_compile.wait(globals()) 2023-03-31T06:22:41.5502153Z del async_compile 2023-03-31T06:22:41.5502274Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5502348Z wrapper = ( 2023-03-31T06:22:41.5502549Z ''' 2023-03-31T06:22:41.5502630Z #include 2023-03-31T06:22:41.5502706Z #include 2023-03-31T06:22:41.5502746Z 2023-03-31T06:22:41.5502841Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5502847Z 2023-03-31T06:22:41.5502939Z template 2023-03-31T06:22:41.5503052Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5503138Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5503264Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5503365Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5503493Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5503573Z return kernel_cpp; 2023-03-31T06:22:41.5503634Z } 2023-03-31T06:22:41.5503751Z std::vector call_157(std::vector args) { 2023-03-31T06:22:41.5503851Z at::Tensor primals_1, tangents_1; 2023-03-31T06:22:41.5503972Z primals_1 = args[0]; 2023-03-31T06:22:41.5504051Z tangents_1 = args[1]; 2023-03-31T06:22:41.5504189Z auto buf0 = at::empty_strided({100000, }, {1, }, at::ScalarType::Float); 2023-03-31T06:22:41.5504499Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/pc/cpcmjer6uoj4fiv5tu3frpax6dqw2mu54ynzoyia7i57hix72so6.so"); 2023-03-31T06:22:41.5504681Z kernel_cpp_0((float*)(tangents_1.data_ptr()), (float*)(primals_1.data_ptr()), (float*)(buf0.data_ptr())); 2023-03-31T06:22:41.5504748Z primals_1.reset(); 2023-03-31T06:22:41.5504827Z tangents_1.reset(); 2023-03-31T06:22:41.5504952Z return std::vector({at::Tensor(), buf0}); 2023-03-31T06:22:41.5505014Z } 2023-03-31T06:22:41.5505097Z ''' 2023-03-31T06:22:41.5505157Z ) 2023-03-31T06:22:41.5505162Z 2023-03-31T06:22:41.5505239Z module = load_inline( 2023-03-31T06:22:41.5505506Z name='inline_extension_cm4gr3qn532sb3lzqodbt3awo5qpfdibupnrapxlescr2mdkms5x', 2023-03-31T06:22:41.5505582Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5505699Z functions=['call_157'], 2023-03-31T06:22:41.5506084Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5506240Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5506985Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5507018Z 2023-03-31T06:22:41.5507093Z def _wrap_func(f): 2023-03-31T06:22:41.5507165Z def g(args): 2023-03-31T06:22:41.5507240Z return f(args) 2023-03-31T06:22:41.5507294Z return g 2023-03-31T06:22:41.5507390Z call = _wrap_func(module.call_157) 2023-03-31T06:22:41.5507398Z 2023-03-31T06:22:41.5507403Z 2023-03-31T06:22:41.5507497Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5507619Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5507747Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5507965Z primals_1 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5508182Z tangents_1 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5508316Z print_performance(lambda: call([primals_1, tangents_1])) 2023-03-31T06:22:41.5508324Z 2023-03-31T06:22:41.5508328Z 2023-03-31T06:22:41.5508406Z if __name__ == "__main__": 2023-03-31T06:22:41.5508470Z import argparse 2023-03-31T06:22:41.5508601Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5508606Z 2023-03-31T06:22:41.5508716Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5509029Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5509398Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5509677Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5509771Z args = parser.parse_args() 2023-03-31T06:22:41.5509776Z 2023-03-31T06:22:41.5509865Z if args.benchmark_kernels: 2023-03-31T06:22:41.5510045Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5510113Z else: 2023-03-31T06:22:41.5510226Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5510321Z benchmark_compiled_module() 2023-03-31T06:22:41.5510327Z 2023-03-31T06:22:41.5510396Z if p: 2023-03-31T06:22:41.5510544Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5510671Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5510796Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5510813Z 2023-03-31T06:22:41.5511086Z [2023-03-31 06:18:44,635] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling BACKWARDS graph 132 2023-03-31T06:22:41.5511358Z [2023-03-31 06:18:44,676] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 133 2023-03-31T06:22:41.5511622Z [2023-03-31 06:18:44,684] torch._inductor.graph: [DEBUG] Set _can_use_cpp_wrapper to False due to ExternKernel 2023-03-31T06:22:41.5512024Z [2023-03-31 06:18:45,863] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/5m/c5mpxggc2afg7eu5zzoh7w3ct2zceuaxsvxn5w3a7tjr53oa2oyd.py 2023-03-31T06:22:41.5512242Z [2023-03-31 06:18:45,863] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5512248Z 2023-03-31T06:22:41.5512344Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5512417Z import torch 2023-03-31T06:22:41.5512487Z import math 2023-03-31T06:22:41.5512546Z import random 2023-03-31T06:22:41.5512613Z import os 2023-03-31T06:22:41.5512688Z import tempfile 2023-03-31T06:22:41.5512807Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5512812Z 2023-03-31T06:22:41.5512930Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5513055Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5513187Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5513193Z 2023-03-31T06:22:41.5513304Z aten = torch.ops.aten 2023-03-31T06:22:41.5513456Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5513550Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5513556Z 2023-03-31T06:22:41.5513561Z 2023-03-31T06:22:41.5513699Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5513921Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5514030Z extern "C" void kernel(float* in_out_ptr0) 2023-03-31T06:22:41.5514093Z { 2023-03-31T06:22:41.5514195Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5514247Z { 2023-03-31T06:22:41.5514310Z { 2023-03-31T06:22:41.5514391Z #pragma omp for 2023-03-31T06:22:41.5514544Z for(long i0=static_cast(0); i0(12500); i0+=static_cast(1)) 2023-03-31T06:22:41.5514610Z { 2023-03-31T06:22:41.5514778Z auto tmp0 = at::vec::Vectorized::loadu(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5514904Z tmp0.store(in_out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5514969Z } 2023-03-31T06:22:41.5515059Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5515217Z for(long i0=static_cast(100000); i0(100000); i0+=static_cast(1)) 2023-03-31T06:22:41.5515286Z { 2023-03-31T06:22:41.5515404Z auto tmp0 = in_out_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5515518Z in_out_ptr0[static_cast(i0)] = tmp0; 2023-03-31T06:22:41.5515615Z } 2023-03-31T06:22:41.5515682Z } 2023-03-31T06:22:41.5515733Z } 2023-03-31T06:22:41.5515794Z } 2023-03-31T06:22:41.5515875Z ''') 2023-03-31T06:22:41.5515881Z 2023-03-31T06:22:41.5515885Z 2023-03-31T06:22:41.5516026Z kernel_cpp_1 = async_compile.cpp(''' 2023-03-31T06:22:41.5516247Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5516357Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5516451Z const float* in_ptr1, 2023-03-31T06:22:41.5516530Z const float* in_ptr2, 2023-03-31T06:22:41.5516615Z float* out_ptr0) 2023-03-31T06:22:41.5516676Z { 2023-03-31T06:22:41.5516778Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5516872Z { 2023-03-31T06:22:41.5516937Z { 2023-03-31T06:22:41.5517018Z #pragma omp for 2023-03-31T06:22:41.5517164Z for(long i0=static_cast(0); i0(12500); i0+=static_cast(1)) 2023-03-31T06:22:41.5517229Z { 2023-03-31T06:22:41.5517388Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5517546Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5517711Z auto tmp3 = at::vec::Vectorized::loadu(in_ptr2 + static_cast(8*i0)); 2023-03-31T06:22:41.5517803Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5517951Z auto tmp4 = at::vec::Vectorized(static_cast(0.6699999999999999)); 2023-03-31T06:22:41.5518043Z auto tmp5 = tmp3 / tmp4; 2023-03-31T06:22:41.5518120Z auto tmp6 = tmp2 * tmp5; 2023-03-31T06:22:41.5518242Z tmp6.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5518307Z } 2023-03-31T06:22:41.5518410Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5518567Z for(long i0=static_cast(100000); i0(100000); i0+=static_cast(1)) 2023-03-31T06:22:41.5518632Z { 2023-03-31T06:22:41.5518745Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5518858Z auto tmp1 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5518984Z auto tmp3 = in_ptr2[static_cast(i0)]; 2023-03-31T06:22:41.5519107Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5519224Z auto tmp4 = static_cast(0.6699999999999999); 2023-03-31T06:22:41.5519312Z auto tmp5 = tmp3 / tmp4; 2023-03-31T06:22:41.5519399Z auto tmp6 = tmp2 * tmp5; 2023-03-31T06:22:41.5519507Z out_ptr0[static_cast(i0)] = tmp6; 2023-03-31T06:22:41.5519578Z } 2023-03-31T06:22:41.5519631Z } 2023-03-31T06:22:41.5519694Z } 2023-03-31T06:22:41.5519755Z } 2023-03-31T06:22:41.5519837Z ''') 2023-03-31T06:22:41.5519845Z 2023-03-31T06:22:41.5519851Z 2023-03-31T06:22:41.5519943Z async_compile.wait(globals()) 2023-03-31T06:22:41.5520018Z del async_compile 2023-03-31T06:22:41.5520023Z 2023-03-31T06:22:41.5520095Z def call(args): 2023-03-31T06:22:41.5520172Z primals_1, primals_2 = args 2023-03-31T06:22:41.5520245Z args.clear() 2023-03-31T06:22:41.5520459Z buf0 = empty_strided((100000, ), (1, ), device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5520550Z buf1 = buf0; del buf0 # reuse 2023-03-31T06:22:41.5520657Z kernel_cpp_0(c_void_p(buf1.data_ptr())) 2023-03-31T06:22:41.5520759Z aten.bernoulli_(buf1, 0.6699999999999999) 2023-03-31T06:22:41.5520968Z buf3 = empty_strided((100000, ), (1, ), device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5521163Z kernel_cpp_1(c_void_p(primals_1.data_ptr()), c_void_p(primals_2.data_ptr()), c_void_p(buf1.data_ptr()), c_void_p(buf3.data_ptr())) 2023-03-31T06:22:41.5521241Z del primals_2 2023-03-31T06:22:41.5521363Z return (buf3, primals_1, buf1, ) 2023-03-31T06:22:41.5521369Z 2023-03-31T06:22:41.5521374Z 2023-03-31T06:22:41.5521469Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5521585Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5521712Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5521930Z primals_1 = rand_strided((100000, ), (1, ), device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5522142Z primals_2 = rand_strided((100000, ), (1, ), device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5522273Z print_performance(lambda: call([primals_1, primals_2])) 2023-03-31T06:22:41.5522279Z 2023-03-31T06:22:41.5522284Z 2023-03-31T06:22:41.5522350Z if __name__ == "__main__": 2023-03-31T06:22:41.5522424Z import argparse 2023-03-31T06:22:41.5522552Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5522585Z 2023-03-31T06:22:41.5522697Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5523143Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5523484Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5523764Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5523860Z args = parser.parse_args() 2023-03-31T06:22:41.5523865Z 2023-03-31T06:22:41.5523958Z if args.benchmark_kernels: 2023-03-31T06:22:41.5524135Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5524203Z else: 2023-03-31T06:22:41.5524316Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5524411Z benchmark_compiled_module() 2023-03-31T06:22:41.5524420Z 2023-03-31T06:22:41.5524489Z if p: 2023-03-31T06:22:41.5524639Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5524737Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5524863Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5524869Z 2023-03-31T06:22:41.5525152Z [2023-03-31 06:18:45,863] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 133 2023-03-31T06:22:41.5525473Z [2023-03-31 06:18:45,869] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling BACKWARDS graph 133 2023-03-31T06:22:41.5525918Z [2023-03-31 06:19:07,131] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/hv/chv6eqvywgi7uhi56dtbj6h2wxasm5lo6e6c7rtpaqk6j6ihpiyt.py 2023-03-31T06:22:41.5526138Z [2023-03-31 06:19:07,131] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5526147Z 2023-03-31T06:22:41.5526245Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5526318Z import torch 2023-03-31T06:22:41.5526389Z import math 2023-03-31T06:22:41.5526448Z import random 2023-03-31T06:22:41.5526521Z import os 2023-03-31T06:22:41.5526597Z import tempfile 2023-03-31T06:22:41.5526715Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5526721Z 2023-03-31T06:22:41.5526841Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5526967Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5527106Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5527113Z 2023-03-31T06:22:41.5527197Z aten = torch.ops.aten 2023-03-31T06:22:41.5527324Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5527419Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5527424Z 2023-03-31T06:22:41.5527429Z 2023-03-31T06:22:41.5527569Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5527793Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5527904Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5528035Z const float* in_ptr1, 2023-03-31T06:22:41.5528130Z const float* in_ptr2, 2023-03-31T06:22:41.5528214Z float* out_ptr0) 2023-03-31T06:22:41.5528266Z { 2023-03-31T06:22:41.5528367Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5528433Z { 2023-03-31T06:22:41.5528499Z { 2023-03-31T06:22:41.5528585Z #pragma omp for 2023-03-31T06:22:41.5528744Z for(long i0=static_cast(0); i0(12500); i0+=static_cast(1)) 2023-03-31T06:22:41.5528813Z { 2023-03-31T06:22:41.5528959Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5529122Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5529346Z auto tmp5 = at::vec::Vectorized::loadu(in_ptr2 + static_cast(8*i0)); 2023-03-31T06:22:41.5529498Z auto tmp2 = at::vec::Vectorized(static_cast(0.6699999999999999)); 2023-03-31T06:22:41.5529592Z auto tmp3 = tmp1 / tmp2; 2023-03-31T06:22:41.5529684Z auto tmp4 = tmp0 * tmp3; 2023-03-31T06:22:41.5529774Z auto tmp6 = tmp4 * tmp5; 2023-03-31T06:22:41.5529896Z tmp6.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5529952Z } 2023-03-31T06:22:41.5530058Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5530218Z for(long i0=static_cast(100000); i0(100000); i0+=static_cast(1)) 2023-03-31T06:22:41.5530287Z { 2023-03-31T06:22:41.5530403Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5530517Z auto tmp1 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5530632Z auto tmp5 = in_ptr2[static_cast(i0)]; 2023-03-31T06:22:41.5530737Z auto tmp2 = static_cast(0.6699999999999999); 2023-03-31T06:22:41.5530829Z auto tmp3 = tmp1 / tmp2; 2023-03-31T06:22:41.5530919Z auto tmp4 = tmp0 * tmp3; 2023-03-31T06:22:41.5531008Z auto tmp6 = tmp4 * tmp5; 2023-03-31T06:22:41.5531118Z out_ptr0[static_cast(i0)] = tmp6; 2023-03-31T06:22:41.5531185Z } 2023-03-31T06:22:41.5531252Z } 2023-03-31T06:22:41.5531330Z } 2023-03-31T06:22:41.5531417Z } 2023-03-31T06:22:41.5531501Z ''') 2023-03-31T06:22:41.5531506Z 2023-03-31T06:22:41.5531602Z async_compile.wait(globals()) 2023-03-31T06:22:41.5531678Z del async_compile 2023-03-31T06:22:41.5531804Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5531878Z wrapper = ( 2023-03-31T06:22:41.5531946Z ''' 2023-03-31T06:22:41.5532024Z #include 2023-03-31T06:22:41.5532105Z #include 2023-03-31T06:22:41.5532110Z 2023-03-31T06:22:41.5532203Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5532208Z 2023-03-31T06:22:41.5532305Z template 2023-03-31T06:22:41.5532433Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5532519Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5532631Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5532728Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5532857Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5532938Z return kernel_cpp; 2023-03-31T06:22:41.5532999Z } 2023-03-31T06:22:41.5533129Z std::vector call_158(std::vector args) { 2023-03-31T06:22:41.5533244Z at::Tensor primals_1, bernoulli, tangents_1; 2023-03-31T06:22:41.5533311Z primals_1 = args[0]; 2023-03-31T06:22:41.5533393Z bernoulli = args[1]; 2023-03-31T06:22:41.5533471Z tangents_1 = args[2]; 2023-03-31T06:22:41.5533609Z auto buf0 = at::empty_strided({100000, }, {1, }, at::ScalarType::Float); 2023-03-31T06:22:41.5533958Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/2f/c2fbmr4vc3f5coivzrow2k2qteslxhc3dctgf5z4spjogvgaluue.so"); 2023-03-31T06:22:41.5534172Z kernel_cpp_0((float*)(tangents_1.data_ptr()), (float*)(bernoulli.data_ptr()), (float*)(primals_1.data_ptr()), (float*)(buf0.data_ptr())); 2023-03-31T06:22:41.5534255Z bernoulli.reset(); 2023-03-31T06:22:41.5534333Z primals_1.reset(); 2023-03-31T06:22:41.5534400Z tangents_1.reset(); 2023-03-31T06:22:41.5534523Z return std::vector({at::Tensor(), buf0}); 2023-03-31T06:22:41.5534585Z } 2023-03-31T06:22:41.5534666Z ''' 2023-03-31T06:22:41.5534727Z ) 2023-03-31T06:22:41.5534733Z 2023-03-31T06:22:41.5534809Z module = load_inline( 2023-03-31T06:22:41.5535081Z name='inline_extension_chpaal7hutjf2ltbmgqrm6kwoyegozxln3ltx4sv5ush7mfvlqls', 2023-03-31T06:22:41.5535188Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5535308Z functions=['call_158'], 2023-03-31T06:22:41.5535692Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5535847Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5536557Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5536567Z 2023-03-31T06:22:41.5536642Z def _wrap_func(f): 2023-03-31T06:22:41.5536712Z def g(args): 2023-03-31T06:22:41.5536787Z return f(args) 2023-03-31T06:22:41.5536853Z return g 2023-03-31T06:22:41.5536938Z call = _wrap_func(module.call_158) 2023-03-31T06:22:41.5536944Z 2023-03-31T06:22:41.5536961Z 2023-03-31T06:22:41.5537043Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5537163Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5537291Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5537509Z primals_1 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5537722Z bernoulli = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5537970Z tangents_1 = rand_strided({100000, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5538142Z print_performance(lambda: call([primals_1, bernoulli, tangents_1])) 2023-03-31T06:22:41.5538148Z 2023-03-31T06:22:41.5538153Z 2023-03-31T06:22:41.5538230Z if __name__ == "__main__": 2023-03-31T06:22:41.5538294Z import argparse 2023-03-31T06:22:41.5538427Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5538436Z 2023-03-31T06:22:41.5538544Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5538854Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5539192Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5539467Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5539563Z args = parser.parse_args() 2023-03-31T06:22:41.5539571Z 2023-03-31T06:22:41.5539659Z if args.benchmark_kernels: 2023-03-31T06:22:41.5539837Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5539903Z else: 2023-03-31T06:22:41.5540015Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5540109Z benchmark_compiled_module() 2023-03-31T06:22:41.5540117Z 2023-03-31T06:22:41.5540184Z if p: 2023-03-31T06:22:41.5540333Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5540457Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5540596Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5540602Z 2023-03-31T06:22:41.5540873Z [2023-03-31 06:19:07,131] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling BACKWARDS graph 133 2023-03-31T06:22:41.5540948Z PASSED [ 55%] 2023-03-31T06:22:41.5541340Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_mm_views_cpu [2023-03-31 06:19:07,165] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 134 2023-03-31T06:22:41.5541750Z [2023-03-31 06:19:27,697] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/hd/chd3bijlis3prmzwvr5jfirisskhl4vszlfxstt5kh2c3agmlbj2.py 2023-03-31T06:22:41.5541969Z [2023-03-31 06:19:27,697] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5542003Z 2023-03-31T06:22:41.5542103Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5542174Z import torch 2023-03-31T06:22:41.5542247Z import math 2023-03-31T06:22:41.5542308Z import random 2023-03-31T06:22:41.5542461Z import os 2023-03-31T06:22:41.5542545Z import tempfile 2023-03-31T06:22:41.5542663Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5542669Z 2023-03-31T06:22:41.5542791Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5542918Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5543054Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5543060Z 2023-03-31T06:22:41.5543144Z aten = torch.ops.aten 2023-03-31T06:22:41.5543270Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5543364Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5543369Z 2023-03-31T06:22:41.5543460Z async_compile.wait(globals()) 2023-03-31T06:22:41.5543539Z del async_compile 2023-03-31T06:22:41.5543658Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5543731Z wrapper = ( 2023-03-31T06:22:41.5543817Z ''' 2023-03-31T06:22:41.5543882Z #include 2023-03-31T06:22:41.5543959Z #include 2023-03-31T06:22:41.5543966Z 2023-03-31T06:22:41.5544060Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5544065Z 2023-03-31T06:22:41.5544157Z template 2023-03-31T06:22:41.5544281Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5544408Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5544560Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5544645Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5544769Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5544847Z return kernel_cpp; 2023-03-31T06:22:41.5544909Z } 2023-03-31T06:22:41.5545040Z std::vector call_159(std::vector args) { 2023-03-31T06:22:41.5545131Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.5545204Z arg0_1 = args[0]; 2023-03-31T06:22:41.5545265Z arg1_1 = args[1]; 2023-03-31T06:22:41.5545403Z auto buf0 = at::empty_strided({32, 32}, {32, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.5545529Z at::mm_out(buf0, arg0_1, at::as_strided(arg1_1, {32, 32}, {32, 1})); 2023-03-31T06:22:41.5545602Z arg0_1.reset(); 2023-03-31T06:22:41.5545673Z arg1_1.reset(); 2023-03-31T06:22:41.5545779Z return std::vector({buf0}); 2023-03-31T06:22:41.5545843Z } 2023-03-31T06:22:41.5545914Z ''' 2023-03-31T06:22:41.5545978Z ) 2023-03-31T06:22:41.5545983Z 2023-03-31T06:22:41.5546060Z module = load_inline( 2023-03-31T06:22:41.5546333Z name='inline_extension_chf5faklnv6zin4pn6c7kr7yt5sngrbycgvyvlads6xzyux6y456', 2023-03-31T06:22:41.5546420Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5546538Z functions=['call_159'], 2023-03-31T06:22:41.5546921Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5547080Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5547823Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5547833Z 2023-03-31T06:22:41.5547898Z def _wrap_func(f): 2023-03-31T06:22:41.5547969Z def g(args): 2023-03-31T06:22:41.5548043Z return f(args) 2023-03-31T06:22:41.5548111Z return g 2023-03-31T06:22:41.5548209Z call = _wrap_func(module.call_159) 2023-03-31T06:22:41.5548214Z 2023-03-31T06:22:41.5548219Z 2023-03-31T06:22:41.5548314Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5548473Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5548603Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5548804Z arg0_1 = rand_strided({32, 32}, {1, 32}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5549023Z arg1_1 = rand_strided({32, 1, 32}, {32, 1024, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5549142Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.5549148Z 2023-03-31T06:22:41.5549153Z 2023-03-31T06:22:41.5549230Z if __name__ == "__main__": 2023-03-31T06:22:41.5549308Z import argparse 2023-03-31T06:22:41.5549439Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5549444Z 2023-03-31T06:22:41.5549553Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5549864Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5550193Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5550477Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5550571Z args = parser.parse_args() 2023-03-31T06:22:41.5550576Z 2023-03-31T06:22:41.5550666Z if args.benchmark_kernels: 2023-03-31T06:22:41.5550855Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5550925Z else: 2023-03-31T06:22:41.5551038Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5551164Z benchmark_compiled_module() 2023-03-31T06:22:41.5551200Z 2023-03-31T06:22:41.5551257Z if p: 2023-03-31T06:22:41.5551407Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5551502Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5551638Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5551647Z 2023-03-31T06:22:41.5551928Z [2023-03-31 06:19:27,698] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 134 2023-03-31T06:22:41.5552006Z PASSED [ 61%] 2023-03-31T06:22:41.5552412Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_profiler_mark_wrapper_call_cpu STAGE:2023-03-31 06:19:27 695:695 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:22:41.5552682Z [2023-03-31 06:19:27,722] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 135 2023-03-31T06:22:41.5553076Z [2023-03-31 06:19:48,850] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/l4/cl4b42m4m5ctz6t3ivr6lj2vjvcny3r4csc6mqpyusjbyiszbozi.py 2023-03-31T06:22:41.5553285Z [2023-03-31 06:19:48,850] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5553302Z 2023-03-31T06:22:41.5553387Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5553458Z import torch 2023-03-31T06:22:41.5553531Z import math 2023-03-31T06:22:41.5553604Z import random 2023-03-31T06:22:41.5553670Z import os 2023-03-31T06:22:41.5553743Z import tempfile 2023-03-31T06:22:41.5553877Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5553895Z 2023-03-31T06:22:41.5554002Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5554126Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5554259Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5554264Z 2023-03-31T06:22:41.5554345Z aten = torch.ops.aten 2023-03-31T06:22:41.5554487Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5554583Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5554588Z 2023-03-31T06:22:41.5554592Z 2023-03-31T06:22:41.5554733Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5554952Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5555078Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5555173Z const float* in_ptr1, 2023-03-31T06:22:41.5555257Z float* out_ptr0) 2023-03-31T06:22:41.5555322Z { 2023-03-31T06:22:41.5555422Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5555487Z { 2023-03-31T06:22:41.5555539Z { 2023-03-31T06:22:41.5555619Z #pragma omp for 2023-03-31T06:22:41.5555773Z for(long i0=static_cast(0); i0(12); i0+=static_cast(1)) 2023-03-31T06:22:41.5555841Z { 2023-03-31T06:22:41.5556003Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5556165Z auto tmp1 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5556255Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.5556375Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5556431Z } 2023-03-31T06:22:41.5556531Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5556685Z for(long i0=static_cast(96); i0(100); i0+=static_cast(1)) 2023-03-31T06:22:41.5556751Z { 2023-03-31T06:22:41.5556865Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5556980Z auto tmp1 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5557070Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.5557199Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.5557293Z } 2023-03-31T06:22:41.5557358Z } 2023-03-31T06:22:41.5557424Z } 2023-03-31T06:22:41.5557485Z } 2023-03-31T06:22:41.5557568Z ''') 2023-03-31T06:22:41.5557573Z 2023-03-31T06:22:41.5557666Z async_compile.wait(globals()) 2023-03-31T06:22:41.5557729Z del async_compile 2023-03-31T06:22:41.5557851Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5557925Z wrapper = ( 2023-03-31T06:22:41.5558006Z ''' 2023-03-31T06:22:41.5558082Z #include 2023-03-31T06:22:41.5558159Z #include 2023-03-31T06:22:41.5558164Z 2023-03-31T06:22:41.5558258Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5558264Z 2023-03-31T06:22:41.5558343Z template 2023-03-31T06:22:41.5558466Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5558552Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5558675Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5558774Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5558902Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5558980Z return kernel_cpp; 2023-03-31T06:22:41.5559030Z } 2023-03-31T06:22:41.5559160Z std::vector call_160(std::vector args) { 2023-03-31T06:22:41.5559245Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.5559318Z arg0_1 = args[0]; 2023-03-31T06:22:41.5559394Z arg1_1 = args[1]; 2023-03-31T06:22:41.5559545Z RECORD_FUNCTION("inductor_wrapper_call", c10::ArrayRef({{}})); 2023-03-31T06:22:41.5559711Z auto buf0 = at::empty_strided({100, }, {1, }, at::ScalarType::Float); 2023-03-31T06:22:41.5560017Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/lj/cljluapn5qw47azfq2ib4gybcvjiss7ibumzyt2z57utss4lu36u.so"); 2023-03-31T06:22:41.5560169Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(arg1_1.data_ptr()), (float*)(buf0.data_ptr())); 2023-03-31T06:22:41.5560243Z arg0_1.reset(); 2023-03-31T06:22:41.5560314Z arg1_1.reset(); 2023-03-31T06:22:41.5560424Z return std::vector({buf0}); 2023-03-31T06:22:41.5560486Z } 2023-03-31T06:22:41.5560570Z ''' 2023-03-31T06:22:41.5560635Z ) 2023-03-31T06:22:41.5560640Z 2023-03-31T06:22:41.5560704Z module = load_inline( 2023-03-31T06:22:41.5560977Z name='inline_extension_chtlilxfi36hbnclt2pyiwaltongglflihl5vbqgjdytsttbbkmw', 2023-03-31T06:22:41.5561094Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5561211Z functions=['call_160'], 2023-03-31T06:22:41.5561594Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5561747Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5562470Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5562479Z 2023-03-31T06:22:41.5562554Z def _wrap_func(f): 2023-03-31T06:22:41.5562625Z def g(args): 2023-03-31T06:22:41.5562688Z return f(args) 2023-03-31T06:22:41.5562756Z return g 2023-03-31T06:22:41.5562855Z call = _wrap_func(module.call_160) 2023-03-31T06:22:41.5562860Z 2023-03-31T06:22:41.5562865Z 2023-03-31T06:22:41.5562960Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5563209Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5563337Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5563548Z arg0_1 = rand_strided({100, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5563755Z arg1_1 = rand_strided({100, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5563916Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.5563952Z 2023-03-31T06:22:41.5563971Z 2023-03-31T06:22:41.5564038Z if __name__ == "__main__": 2023-03-31T06:22:41.5564116Z import argparse 2023-03-31T06:22:41.5564245Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5564250Z 2023-03-31T06:22:41.5564361Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5564676Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5565022Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5565302Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5565396Z args = parser.parse_args() 2023-03-31T06:22:41.5565401Z 2023-03-31T06:22:41.5565477Z if args.benchmark_kernels: 2023-03-31T06:22:41.5565671Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5565742Z else: 2023-03-31T06:22:41.5565856Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5565953Z benchmark_compiled_module() 2023-03-31T06:22:41.5565959Z 2023-03-31T06:22:41.5566030Z if p: 2023-03-31T06:22:41.5566181Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5566266Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5566406Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5566451Z 2023-03-31T06:22:41.5566737Z [2023-03-31 06:19:48,851] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 135 2023-03-31T06:22:41.5566999Z STAGE:2023-03-31 06:19:48 695:695 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:22:41.5567275Z STAGE:2023-03-31 06:19:48 695:695 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:22:41.5567353Z PASSED [ 66%] 2023-03-31T06:22:41.5567752Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_reduction1_cpu [2023-03-31 06:19:48,889] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 136 2023-03-31T06:22:41.5568131Z [2023-03-31 06:19:48,909] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.float32, reduction_type argmax 2023-03-31T06:22:41.5568553Z [2023-03-31 06:19:48,909] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.float32, reduction_type argmin 2023-03-31T06:22:41.5568960Z [2023-03-31 06:20:10,270] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/wz/cwzvjqrol3l5tza4qa3kt3fskowtwyk5hazdbyioakemchwgbpjc.py 2023-03-31T06:22:41.5569168Z [2023-03-31 06:20:10,270] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5569189Z 2023-03-31T06:22:41.5569276Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5569351Z import torch 2023-03-31T06:22:41.5569425Z import math 2023-03-31T06:22:41.5569499Z import random 2023-03-31T06:22:41.5569568Z import os 2023-03-31T06:22:41.5569645Z import tempfile 2023-03-31T06:22:41.5569751Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5569756Z 2023-03-31T06:22:41.5569878Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5570007Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5570141Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5570149Z 2023-03-31T06:22:41.5570233Z aten = torch.ops.aten 2023-03-31T06:22:41.5570373Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5570471Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5570476Z 2023-03-31T06:22:41.5570481Z 2023-03-31T06:22:41.5570624Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5570877Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5571015Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5571103Z float* out_ptr0, 2023-03-31T06:22:41.5571194Z float* out_ptr1, 2023-03-31T06:22:41.5571278Z float* out_ptr2, 2023-03-31T06:22:41.5571363Z long* out_ptr3, 2023-03-31T06:22:41.5571448Z long* out_ptr4) 2023-03-31T06:22:41.5571512Z { 2023-03-31T06:22:41.5571563Z { 2023-03-31T06:22:41.5571629Z { 2023-03-31T06:22:41.5571713Z float tmp1 = 0; 2023-03-31T06:22:41.5571967Z float tmp2 = -std::numeric_limits::infinity(); 2023-03-31T06:22:41.5572097Z float tmp3 = std::numeric_limits::infinity(); 2023-03-31T06:22:41.5572221Z struct IndexValue_1 {size_t index; float value;}; 2023-03-31T06:22:41.5572454Z IndexValue_1 tmp4{0, -std::numeric_limits::infinity()}; 2023-03-31T06:22:41.5572586Z #pragma omp declare reduction(argmax : struct IndexValue_1 :\ 2023-03-31T06:22:41.5572742Z omp_out.value = omp_in.value < omp_out.value ? omp_out.value : omp_in.value,\ 2023-03-31T06:22:41.5572893Z omp_out.index = omp_in.value < omp_out.value ? omp_out.index : omp_in.index)\ 2023-03-31T06:22:41.5573133Z initializer(omp_priv = {0, -std::numeric_limits::infinity()}) 2023-03-31T06:22:41.5573259Z struct IndexValue_2 {size_t index; float value;}; 2023-03-31T06:22:41.5573429Z IndexValue_2 tmp5{0, std::numeric_limits::infinity()}; 2023-03-31T06:22:41.5573571Z #pragma omp declare reduction(argmin : struct IndexValue_2 :\ 2023-03-31T06:22:41.5573724Z omp_out.value = omp_in.value > omp_out.value ? omp_out.value : omp_in.value,\ 2023-03-31T06:22:41.5573875Z omp_out.index = omp_in.value > omp_out.value ? omp_out.index : omp_in.index)\ 2023-03-31T06:22:41.5574012Z initializer(omp_priv = {0, std::numeric_limits::infinity()}) 2023-03-31T06:22:41.5574167Z for(long i0=static_cast(0); i0(3); i0+=static_cast(1)) 2023-03-31T06:22:41.5574235Z { 2023-03-31T06:22:41.5574351Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5574429Z tmp1 += tmp0; 2023-03-31T06:22:41.5574557Z tmp2 = std::max(tmp2, tmp0); 2023-03-31T06:22:41.5574656Z tmp3 = std::min(tmp3, tmp0); 2023-03-31T06:22:41.5574739Z if (tmp4.value < tmp0) { 2023-03-31T06:22:41.5574849Z tmp4.index = i0; tmp4.value = tmp0; 2023-03-31T06:22:41.5574915Z } 2023-03-31T06:22:41.5575004Z if (tmp5.value > tmp0) { 2023-03-31T06:22:41.5575109Z tmp5.index = i0; tmp5.value = tmp0; 2023-03-31T06:22:41.5575175Z } 2023-03-31T06:22:41.5575241Z } 2023-03-31T06:22:41.5575338Z out_ptr0[static_cast(0)] = tmp1; 2023-03-31T06:22:41.5575445Z out_ptr1[static_cast(0)] = tmp2; 2023-03-31T06:22:41.5575544Z out_ptr2[static_cast(0)] = tmp3; 2023-03-31T06:22:41.5575657Z out_ptr3[static_cast(0)] = tmp4.index; 2023-03-31T06:22:41.5575768Z out_ptr4[static_cast(0)] = tmp5.index; 2023-03-31T06:22:41.5575835Z } 2023-03-31T06:22:41.5575898Z } 2023-03-31T06:22:41.5575947Z } 2023-03-31T06:22:41.5576030Z ''') 2023-03-31T06:22:41.5576035Z 2023-03-31T06:22:41.5576129Z async_compile.wait(globals()) 2023-03-31T06:22:41.5576203Z del async_compile 2023-03-31T06:22:41.5576328Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5576402Z wrapper = ( 2023-03-31T06:22:41.5576481Z ''' 2023-03-31T06:22:41.5576544Z #include 2023-03-31T06:22:41.5576619Z #include 2023-03-31T06:22:41.5576624Z 2023-03-31T06:22:41.5576723Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5576756Z 2023-03-31T06:22:41.5576874Z template 2023-03-31T06:22:41.5577000Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5577086Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5577212Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5577296Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5577424Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5577504Z return kernel_cpp; 2023-03-31T06:22:41.5577565Z } 2023-03-31T06:22:41.5577698Z std::vector call_161(std::vector args) { 2023-03-31T06:22:41.5577776Z at::Tensor arg0_1; 2023-03-31T06:22:41.5577849Z arg0_1 = args[0]; 2023-03-31T06:22:41.5577964Z auto buf0 = at::empty_strided({}, {}, at::ScalarType::Float); 2023-03-31T06:22:41.5578091Z auto buf1 = at::empty_strided({}, {}, at::ScalarType::Float); 2023-03-31T06:22:41.5578217Z auto buf2 = at::empty_strided({}, {}, at::ScalarType::Float); 2023-03-31T06:22:41.5578344Z auto buf3 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5578467Z auto buf4 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5578783Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/xt/cxtwm2dxnu4ft3a4uuygpvr5pqaw2wpmq3mwnxt3uv4b6hqg2g3v.so"); 2023-03-31T06:22:41.5579086Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf1.data_ptr()), (float*)(buf2.data_ptr()), (long*)(buf3.data_ptr()), (long*)(buf4.data_ptr())); 2023-03-31T06:22:41.5579161Z arg0_1.reset(); 2023-03-31T06:22:41.5579291Z return std::vector({buf0, buf1, buf2, buf3, buf4}); 2023-03-31T06:22:41.5579340Z } 2023-03-31T06:22:41.5579423Z ''' 2023-03-31T06:22:41.5579486Z ) 2023-03-31T06:22:41.5579491Z 2023-03-31T06:22:41.5579569Z module = load_inline( 2023-03-31T06:22:41.5579849Z name='inline_extension_cojqbzhl4ego5jpmey2toxymletygxqix5ghpe4c6im5wnekp4bh', 2023-03-31T06:22:41.5579939Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5580057Z functions=['call_161'], 2023-03-31T06:22:41.5580439Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5580581Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5581319Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5581326Z 2023-03-31T06:22:41.5581403Z def _wrap_func(f): 2023-03-31T06:22:41.5581475Z def g(args): 2023-03-31T06:22:41.5581549Z return f(args) 2023-03-31T06:22:41.5581617Z return g 2023-03-31T06:22:41.5581714Z call = _wrap_func(module.call_161) 2023-03-31T06:22:41.5581720Z 2023-03-31T06:22:41.5581725Z 2023-03-31T06:22:41.5581819Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5581939Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5582054Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5582258Z arg0_1 = rand_strided({3, }, {1, }, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5582464Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5582473Z 2023-03-31T06:22:41.5582477Z 2023-03-31T06:22:41.5582561Z if __name__ == "__main__": 2023-03-31T06:22:41.5582637Z import argparse 2023-03-31T06:22:41.5582770Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5582776Z 2023-03-31T06:22:41.5582890Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5583242Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5583596Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5583873Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5583966Z args = parser.parse_args() 2023-03-31T06:22:41.5583974Z 2023-03-31T06:22:41.5584065Z if args.benchmark_kernels: 2023-03-31T06:22:41.5584254Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5584323Z else: 2023-03-31T06:22:41.5584435Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5584530Z benchmark_compiled_module() 2023-03-31T06:22:41.5584536Z 2023-03-31T06:22:41.5584591Z if p: 2023-03-31T06:22:41.5584740Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5584835Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5584974Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5584982Z 2023-03-31T06:22:41.5585267Z [2023-03-31 06:20:10,270] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 136 2023-03-31T06:22:41.5585341Z PASSED [ 72%] 2023-03-31T06:22:41.5585728Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_relu_cpu [2023-03-31 06:20:10,305] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 137 2023-03-31T06:22:41.5586168Z [2023-03-31 06:20:32,090] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3d/c3dsaomkt4lidgcldi4367ij2zux4pcjtkpbwyqd54ahyeytaah3.py 2023-03-31T06:22:41.5586388Z [2023-03-31 06:20:32,091] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5586394Z 2023-03-31T06:22:41.5586479Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5586549Z import torch 2023-03-31T06:22:41.5586621Z import math 2023-03-31T06:22:41.5586695Z import random 2023-03-31T06:22:41.5586764Z import os 2023-03-31T06:22:41.5586838Z import tempfile 2023-03-31T06:22:41.5586957Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5586963Z 2023-03-31T06:22:41.5587070Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5587196Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5587328Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5587359Z 2023-03-31T06:22:41.5587440Z aten = torch.ops.aten 2023-03-31T06:22:41.5587581Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5587678Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5587684Z 2023-03-31T06:22:41.5587688Z 2023-03-31T06:22:41.5587828Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5588051Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5588146Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5588242Z const float* in_ptr1, 2023-03-31T06:22:41.5588328Z float* out_ptr0, 2023-03-31T06:22:41.5588411Z float* out_ptr1) 2023-03-31T06:22:41.5588473Z { 2023-03-31T06:22:41.5588572Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5588635Z { 2023-03-31T06:22:41.5588687Z { 2023-03-31T06:22:41.5588767Z #pragma omp for 2023-03-31T06:22:41.5588923Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.5588988Z { 2023-03-31T06:22:41.5589151Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5589307Z auto tmp2 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5589440Z auto tmp1 = at::vec::clamp_min(tmp0, decltype(tmp0)(0)); 2023-03-31T06:22:41.5589562Z auto tmp3 = tmp0 + tmp2; 2023-03-31T06:22:41.5589709Z auto tmp4 = at::vec::clamp_min(tmp3, decltype(tmp3)(0)); 2023-03-31T06:22:41.5589850Z auto tmp5 = at::vec::Vectorized(static_cast(10)); 2023-03-31T06:22:41.5589942Z auto tmp6 = tmp4 / tmp5; 2023-03-31T06:22:41.5590061Z tmp1.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5590178Z tmp6.store(out_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5590247Z } 2023-03-31T06:22:41.5590347Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5590488Z for(long i0=static_cast(64); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5590554Z { 2023-03-31T06:22:41.5590669Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5590781Z auto tmp2 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5590873Z auto tmp1 = tmp0 * (tmp0>0); 2023-03-31T06:22:41.5590965Z auto tmp3 = tmp0 + tmp2; 2023-03-31T06:22:41.5591061Z auto tmp4 = tmp3 * (tmp3>0); 2023-03-31T06:22:41.5591166Z auto tmp5 = static_cast(10); 2023-03-31T06:22:41.5591243Z auto tmp6 = tmp4 / tmp5; 2023-03-31T06:22:41.5591352Z out_ptr0[static_cast(i0)] = tmp1; 2023-03-31T06:22:41.5591460Z out_ptr1[static_cast(i0)] = tmp6; 2023-03-31T06:22:41.5591528Z } 2023-03-31T06:22:41.5591593Z } 2023-03-31T06:22:41.5591657Z } 2023-03-31T06:22:41.5591706Z } 2023-03-31T06:22:41.5591816Z ''') 2023-03-31T06:22:41.5591822Z 2023-03-31T06:22:41.5591917Z async_compile.wait(globals()) 2023-03-31T06:22:41.5591992Z del async_compile 2023-03-31T06:22:41.5592115Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5592188Z wrapper = ( 2023-03-31T06:22:41.5592268Z ''' 2023-03-31T06:22:41.5592331Z #include 2023-03-31T06:22:41.5592409Z #include 2023-03-31T06:22:41.5592416Z 2023-03-31T06:22:41.5592509Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5592515Z 2023-03-31T06:22:41.5592607Z template 2023-03-31T06:22:41.5592729Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5592815Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5592938Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5593081Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5593192Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5593270Z return kernel_cpp; 2023-03-31T06:22:41.5593334Z } 2023-03-31T06:22:41.5593463Z std::vector call_162(std::vector args) { 2023-03-31T06:22:41.5593549Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.5593622Z arg0_1 = args[0]; 2023-03-31T06:22:41.5593694Z arg1_1 = args[1]; 2023-03-31T06:22:41.5593817Z auto buf0 = at::empty_strided({8, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.5593951Z auto buf1 = at::empty_strided({8, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.5594250Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/ip/cipw76463ensd2j7za6fikst75hj2jyn752iygf2dp22m6vpmz5x.so"); 2023-03-31T06:22:41.5594443Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(arg1_1.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf1.data_ptr())); 2023-03-31T06:22:41.5594518Z arg0_1.reset(); 2023-03-31T06:22:41.5594590Z arg1_1.reset(); 2023-03-31T06:22:41.5594703Z return std::vector({buf0, buf1}); 2023-03-31T06:22:41.5594765Z } 2023-03-31T06:22:41.5594833Z ''' 2023-03-31T06:22:41.5594894Z ) 2023-03-31T06:22:41.5594899Z 2023-03-31T06:22:41.5594977Z module = load_inline( 2023-03-31T06:22:41.5595250Z name='inline_extension_cxdfrdgpgznitup6l2x6r3te3e3irzemrmmosf236zrqfx3tci36', 2023-03-31T06:22:41.5595336Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5595484Z functions=['call_162'], 2023-03-31T06:22:41.5595894Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5596037Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5596745Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5596767Z 2023-03-31T06:22:41.5596829Z def _wrap_func(f): 2023-03-31T06:22:41.5596899Z def g(args): 2023-03-31T06:22:41.5596974Z return f(args) 2023-03-31T06:22:41.5597041Z return g 2023-03-31T06:22:41.5597136Z call = _wrap_func(module.call_162) 2023-03-31T06:22:41.5597142Z 2023-03-31T06:22:41.5597147Z 2023-03-31T06:22:41.5597244Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5597364Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5597478Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5597684Z arg0_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5597889Z arg1_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5598010Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.5598015Z 2023-03-31T06:22:41.5598020Z 2023-03-31T06:22:41.5598125Z if __name__ == "__main__": 2023-03-31T06:22:41.5598203Z import argparse 2023-03-31T06:22:41.5598332Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5598338Z 2023-03-31T06:22:41.5598449Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5598752Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5599094Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5599373Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5599466Z args = parser.parse_args() 2023-03-31T06:22:41.5599471Z 2023-03-31T06:22:41.5599597Z if args.benchmark_kernels: 2023-03-31T06:22:41.5599787Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5599854Z else: 2023-03-31T06:22:41.5599969Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5600052Z benchmark_compiled_module() 2023-03-31T06:22:41.5600070Z 2023-03-31T06:22:41.5600126Z if p: 2023-03-31T06:22:41.5600274Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5600370Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5600509Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5600517Z 2023-03-31T06:22:41.5600798Z [2023-03-31 06:20:32,091] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 137 2023-03-31T06:22:41.5600874Z PASSED [ 77%] 2023-03-31T06:22:41.5601256Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_silu_cpu [2023-03-31 06:20:32,120] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 138 2023-03-31T06:22:41.5601659Z [2023-03-31 06:20:53,757] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/je/cjed2ynd7zonelscr57vq5bau4znfvgqp7h352doirlhok5neaxh.py 2023-03-31T06:22:41.5601866Z [2023-03-31 06:20:53,757] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5601885Z 2023-03-31T06:22:41.5601968Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5602039Z import torch 2023-03-31T06:22:41.5602109Z import math 2023-03-31T06:22:41.5602180Z import random 2023-03-31T06:22:41.5602282Z import os 2023-03-31T06:22:41.5602383Z import tempfile 2023-03-31T06:22:41.5602490Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5602507Z 2023-03-31T06:22:41.5602614Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5602737Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5602872Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5602880Z 2023-03-31T06:22:41.5602960Z aten = torch.ops.aten 2023-03-31T06:22:41.5603293Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5603394Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5603400Z 2023-03-31T06:22:41.5603404Z 2023-03-31T06:22:41.5603547Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5603769Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5603866Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5603956Z float* out_ptr0) 2023-03-31T06:22:41.5604022Z { 2023-03-31T06:22:41.5604123Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5604189Z { 2023-03-31T06:22:41.5604256Z { 2023-03-31T06:22:41.5604324Z #pragma omp for 2023-03-31T06:22:41.5604480Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.5604550Z { 2023-03-31T06:22:41.5604712Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5604907Z auto tmp1 = decltype(tmp0)(1)/(decltype(tmp0)(1) + tmp0.neg().exp()); 2023-03-31T06:22:41.5605002Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5605125Z tmp2.store(out_ptr0 + static_cast(8*i0)); 2023-03-31T06:22:41.5605193Z } 2023-03-31T06:22:41.5605283Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5605441Z for(long i0=static_cast(64); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5605510Z { 2023-03-31T06:22:41.5605627Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5605853Z auto tmp1 = decltype(tmp0)(1) / (decltype(tmp0)(1) + std::exp(-tmp0)); 2023-03-31T06:22:41.5605947Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5606094Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.5606149Z } 2023-03-31T06:22:41.5606215Z } 2023-03-31T06:22:41.5606280Z } 2023-03-31T06:22:41.5606343Z } 2023-03-31T06:22:41.5606431Z ''') 2023-03-31T06:22:41.5606437Z 2023-03-31T06:22:41.5606531Z async_compile.wait(globals()) 2023-03-31T06:22:41.5606609Z del async_compile 2023-03-31T06:22:41.5606720Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5606794Z wrapper = ( 2023-03-31T06:22:41.5606876Z ''' 2023-03-31T06:22:41.5606953Z #include 2023-03-31T06:22:41.5607033Z #include 2023-03-31T06:22:41.5607041Z 2023-03-31T06:22:41.5607138Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5607144Z 2023-03-31T06:22:41.5607239Z template 2023-03-31T06:22:41.5607351Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5607438Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5607564Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5607665Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5607792Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5607874Z return kernel_cpp; 2023-03-31T06:22:41.5607937Z } 2023-03-31T06:22:41.5608054Z std::vector call_163(std::vector args) { 2023-03-31T06:22:41.5608138Z at::Tensor arg0_1; 2023-03-31T06:22:41.5608213Z arg0_1 = args[0]; 2023-03-31T06:22:41.5608348Z auto buf0 = at::empty_strided({8, 8}, {8, 1}, at::ScalarType::Float); 2023-03-31T06:22:41.5608683Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/yc/cycgfrva3u6uwpo6rrqo65m62f4cljwozewwbzpqkqr735rm5h2b.so"); 2023-03-31T06:22:41.5608853Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(buf0.data_ptr())); 2023-03-31T06:22:41.5608928Z arg0_1.reset(); 2023-03-31T06:22:41.5609035Z return std::vector({buf0}); 2023-03-31T06:22:41.5609085Z } 2023-03-31T06:22:41.5609173Z ''' 2023-03-31T06:22:41.5609236Z ) 2023-03-31T06:22:41.5609241Z 2023-03-31T06:22:41.5609321Z module = load_inline( 2023-03-31T06:22:41.5609596Z name='inline_extension_cunrlso7hew4cur2akhzgeijs43luynhwhpt25iuz3iygbpe6ci3', 2023-03-31T06:22:41.5609684Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5609802Z functions=['call_163'], 2023-03-31T06:22:41.5610175Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5610335Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5611046Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5611055Z 2023-03-31T06:22:41.5611134Z def _wrap_func(f): 2023-03-31T06:22:41.5611207Z def g(args): 2023-03-31T06:22:41.5611282Z return f(args) 2023-03-31T06:22:41.5611350Z return g 2023-03-31T06:22:41.5611476Z call = _wrap_func(module.call_163) 2023-03-31T06:22:41.5611483Z 2023-03-31T06:22:41.5611488Z 2023-03-31T06:22:41.5611582Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5611689Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5611816Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5612027Z arg0_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5612140Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5612146Z 2023-03-31T06:22:41.5612151Z 2023-03-31T06:22:41.5612227Z if __name__ == "__main__": 2023-03-31T06:22:41.5612303Z import argparse 2023-03-31T06:22:41.5612435Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5612441Z 2023-03-31T06:22:41.5612580Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5612879Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5613216Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5613492Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5613584Z args = parser.parse_args() 2023-03-31T06:22:41.5613592Z 2023-03-31T06:22:41.5613683Z if args.benchmark_kernels: 2023-03-31T06:22:41.5613874Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5613940Z else: 2023-03-31T06:22:41.5614052Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5614133Z benchmark_compiled_module() 2023-03-31T06:22:41.5614151Z 2023-03-31T06:22:41.5614207Z if p: 2023-03-31T06:22:41.5614359Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5614455Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5614593Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5614599Z 2023-03-31T06:22:41.5614880Z [2023-03-31 06:20:53,758] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 138 2023-03-31T06:22:41.5614955Z PASSED [ 83%] 2023-03-31T06:22:41.5615371Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_sum_dtype_cpu [2023-03-31 06:20:53,792] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 139 2023-03-31T06:22:41.5615716Z [2023-03-31 06:20:53,812] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5616079Z [2023-03-31 06:20:53,812] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.float64, src_dtype torch.float64, reduction_type sum 2023-03-31T06:22:41.5616380Z [2023-03-31 06:20:53,813] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5616677Z [2023-03-31 06:20:53,815] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5617047Z [2023-03-31 06:20:53,815] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.float64, src_dtype torch.float64, reduction_type sum 2023-03-31T06:22:41.5617341Z [2023-03-31 06:20:53,816] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5617638Z [2023-03-31 06:20:53,818] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5617946Z [2023-03-31 06:20:53,818] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by load 2023-03-31T06:22:41.5618251Z [2023-03-31 06:20:53,818] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by load 2023-03-31T06:22:41.5618589Z [2023-03-31 06:20:53,818] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.float64 not supported by store 2023-03-31T06:22:41.5618887Z [2023-03-31 06:20:53,818] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.float64 2023-03-31T06:22:41.5619294Z [2023-03-31 06:21:15,387] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/no/cnolzwbnq3w2ck45pdazsgf6sjytbnmocc77p2w5lg54wqcc2fh3.py 2023-03-31T06:22:41.5619504Z [2023-03-31 06:21:15,387] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5619522Z 2023-03-31T06:22:41.5619608Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5619679Z import torch 2023-03-31T06:22:41.5619750Z import math 2023-03-31T06:22:41.5619822Z import random 2023-03-31T06:22:41.5619923Z import os 2023-03-31T06:22:41.5619998Z import tempfile 2023-03-31T06:22:41.5620106Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5620112Z 2023-03-31T06:22:41.5620233Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5620358Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5620490Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5620496Z 2023-03-31T06:22:41.5620578Z aten = torch.ops.aten 2023-03-31T06:22:41.5620718Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5620816Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5620821Z 2023-03-31T06:22:41.5620826Z 2023-03-31T06:22:41.5620967Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5621177Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5621286Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5621379Z double* out_ptr0, 2023-03-31T06:22:41.5621465Z double* out_ptr1, 2023-03-31T06:22:41.5621548Z double* out_ptr2) 2023-03-31T06:22:41.5621613Z { 2023-03-31T06:22:41.5621714Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5621764Z { 2023-03-31T06:22:41.5621828Z { 2023-03-31T06:22:41.5621908Z #pragma omp for 2023-03-31T06:22:41.5622066Z for(long i0=static_cast(0); i0(32); i0+=static_cast(1)) 2023-03-31T06:22:41.5622131Z { 2023-03-31T06:22:41.5622224Z { 2023-03-31T06:22:41.5622346Z double tmp2 = 0; 2023-03-31T06:22:41.5622580Z for(long i1=static_cast(0); i1(32); i1+=static_cast(1)) 2023-03-31T06:22:41.5622654Z { 2023-03-31T06:22:41.5622780Z auto tmp0 = in_ptr0[static_cast(i1 + (32*i0))]; 2023-03-31T06:22:41.5622900Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5622982Z tmp2 += tmp1; 2023-03-31T06:22:41.5623050Z } 2023-03-31T06:22:41.5623164Z out_ptr0[static_cast(i0)] = tmp2; 2023-03-31T06:22:41.5623231Z } 2023-03-31T06:22:41.5623284Z } 2023-03-31T06:22:41.5623347Z } 2023-03-31T06:22:41.5623410Z } 2023-03-31T06:22:41.5623473Z { 2023-03-31T06:22:41.5623536Z { 2023-03-31T06:22:41.5623617Z double tmp2 = 0; 2023-03-31T06:22:41.5623717Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5623784Z { 2023-03-31T06:22:41.5623893Z #pragma omp for reduction(+:tmp2) 2023-03-31T06:22:41.5624050Z for(long i0=static_cast(0); i0(1024); i0+=static_cast(1)) 2023-03-31T06:22:41.5624118Z { 2023-03-31T06:22:41.5624234Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5624350Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5624418Z tmp2 += tmp1; 2023-03-31T06:22:41.5624520Z } 2023-03-31T06:22:41.5624586Z } 2023-03-31T06:22:41.5624689Z out_ptr1[static_cast(0)] = tmp2; 2023-03-31T06:22:41.5624752Z } 2023-03-31T06:22:41.5624813Z } 2023-03-31T06:22:41.5624914Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5624964Z { 2023-03-31T06:22:41.5625027Z { 2023-03-31T06:22:41.5625113Z #pragma omp for 2023-03-31T06:22:41.5625267Z for(long i0=static_cast(0); i0(32); i0+=static_cast(1)) 2023-03-31T06:22:41.5625336Z { 2023-03-31T06:22:41.5625419Z #pragma GCC ivdep 2023-03-31T06:22:41.5625572Z for(long i1=static_cast(0); i1(32); i1+=static_cast(1)) 2023-03-31T06:22:41.5625625Z { 2023-03-31T06:22:41.5625781Z auto tmp0 = in_ptr0[static_cast(i1 + (32*i0))]; 2023-03-31T06:22:41.5625899Z auto tmp2 = out_ptr0[static_cast(i1)]; 2023-03-31T06:22:41.5626017Z auto tmp4 = out_ptr1[static_cast(0)]; 2023-03-31T06:22:41.5626131Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5626226Z auto tmp3 = tmp1 * tmp2; 2023-03-31T06:22:41.5626320Z auto tmp5 = tmp3 + tmp4; 2023-03-31T06:22:41.5626429Z out_ptr2[static_cast(i1 + (32*i0))] = tmp5; 2023-03-31T06:22:41.5626497Z } 2023-03-31T06:22:41.5626561Z } 2023-03-31T06:22:41.5626624Z } 2023-03-31T06:22:41.5626685Z } 2023-03-31T06:22:41.5626746Z } 2023-03-31T06:22:41.5626836Z ''') 2023-03-31T06:22:41.5626842Z 2023-03-31T06:22:41.5626923Z async_compile.wait(globals()) 2023-03-31T06:22:41.5626997Z del async_compile 2023-03-31T06:22:41.5627122Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5627195Z wrapper = ( 2023-03-31T06:22:41.5627276Z ''' 2023-03-31T06:22:41.5627352Z #include 2023-03-31T06:22:41.5627430Z #include 2023-03-31T06:22:41.5627435Z 2023-03-31T06:22:41.5627520Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5627525Z 2023-03-31T06:22:41.5627617Z template 2023-03-31T06:22:41.5627741Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5627826Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5627977Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5628142Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5628268Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5628335Z return kernel_cpp; 2023-03-31T06:22:41.5628396Z } 2023-03-31T06:22:41.5628527Z std::vector call_164(std::vector args) { 2023-03-31T06:22:41.5628606Z at::Tensor arg0_1; 2023-03-31T06:22:41.5628683Z arg0_1 = args[0]; 2023-03-31T06:22:41.5628818Z auto buf0 = at::empty_strided({32, }, {1, }, at::ScalarType::Double); 2023-03-31T06:22:41.5628950Z auto buf1 = at::empty_strided({}, {}, at::ScalarType::Double); 2023-03-31T06:22:41.5629088Z auto buf2 = at::empty_strided({32, 32}, {32, 1}, at::ScalarType::Double); 2023-03-31T06:22:41.5629397Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/gq/cgqkfhlzunqhpziishakl6nmq2id5urehmk3hxz4r7fjat4dfhki.so"); 2023-03-31T06:22:41.5629596Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (double*)(buf0.data_ptr()), (double*)(buf1.data_ptr()), (double*)(buf2.data_ptr())); 2023-03-31T06:22:41.5629672Z arg0_1.reset(); 2023-03-31T06:22:41.5629778Z return std::vector({buf2}); 2023-03-31T06:22:41.5629843Z } 2023-03-31T06:22:41.5629927Z ''' 2023-03-31T06:22:41.5629989Z ) 2023-03-31T06:22:41.5629994Z 2023-03-31T06:22:41.5630071Z module = load_inline( 2023-03-31T06:22:41.5630334Z name='inline_extension_c6g7looj2dadfpvr2avhearuy2dutpbscdvh4wvl37hmhoegqnsl', 2023-03-31T06:22:41.5630421Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5630564Z functions=['call_164'], 2023-03-31T06:22:41.5630947Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5631102Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5631821Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5631830Z 2023-03-31T06:22:41.5631904Z def _wrap_func(f): 2023-03-31T06:22:41.5631977Z def g(args): 2023-03-31T06:22:41.5632066Z return f(args) 2023-03-31T06:22:41.5632133Z return g 2023-03-31T06:22:41.5632229Z call = _wrap_func(module.call_164) 2023-03-31T06:22:41.5632235Z 2023-03-31T06:22:41.5632240Z 2023-03-31T06:22:41.5632335Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5632453Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5632580Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5632794Z arg0_1 = rand_strided({32, 32}, {32, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5632907Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5632915Z 2023-03-31T06:22:41.5632919Z 2023-03-31T06:22:41.5632997Z if __name__ == "__main__": 2023-03-31T06:22:41.5633059Z import argparse 2023-03-31T06:22:41.5633188Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5633194Z 2023-03-31T06:22:41.5633305Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5633615Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5633959Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5634234Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5634328Z args = parser.parse_args() 2023-03-31T06:22:41.5634334Z 2023-03-31T06:22:41.5634422Z if args.benchmark_kernels: 2023-03-31T06:22:41.5634632Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5634723Z else: 2023-03-31T06:22:41.5634837Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5634932Z benchmark_compiled_module() 2023-03-31T06:22:41.5634937Z 2023-03-31T06:22:41.5635005Z if p: 2023-03-31T06:22:41.5635155Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5635253Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5635389Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5635394Z 2023-03-31T06:22:41.5635668Z [2023-03-31 06:21:15,387] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 139 2023-03-31T06:22:41.5635742Z PASSED [ 88%] 2023-03-31T06:22:41.5636128Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_sum_int_cpu [2023-03-31 06:21:15,419] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 140 2023-03-31T06:22:41.5636432Z [2023-03-31 06:21:15,434] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.bool not loaded as mask 2023-03-31T06:22:41.5636727Z [2023-03-31 06:21:15,435] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5637091Z [2023-03-31 06:21:15,435] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5637428Z [2023-03-31 06:21:15,435] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.bool not loaded as mask 2023-03-31T06:22:41.5637720Z [2023-03-31 06:21:15,435] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5638088Z [2023-03-31 06:21:15,435] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5638379Z [2023-03-31 06:21:15,436] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5638657Z [2023-03-31 06:21:15,436] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5638960Z [2023-03-31 06:21:15,438] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5639291Z [2023-03-31 06:21:15,438] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5639595Z [2023-03-31 06:21:15,438] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5639897Z [2023-03-31 06:21:15,438] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by store 2023-03-31T06:22:41.5640184Z [2023-03-31 06:21:15,438] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5640589Z [2023-03-31 06:21:36,717] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/3k/c3kfcfnuqy22ofl4vhsdrmuewhupvmhch3tqw3pic2ictk6wxffo.py 2023-03-31T06:22:41.5640809Z [2023-03-31 06:21:36,717] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5640815Z 2023-03-31T06:22:41.5640915Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5640986Z import torch 2023-03-31T06:22:41.5641044Z import math 2023-03-31T06:22:41.5641116Z import random 2023-03-31T06:22:41.5641186Z import os 2023-03-31T06:22:41.5641260Z import tempfile 2023-03-31T06:22:41.5641378Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5641384Z 2023-03-31T06:22:41.5641504Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5641628Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5641776Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5641808Z 2023-03-31T06:22:41.5641889Z aten = torch.ops.aten 2023-03-31T06:22:41.5642033Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5642129Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5642135Z 2023-03-31T06:22:41.5642139Z 2023-03-31T06:22:41.5642282Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5642505Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5642609Z extern "C" void kernel(long* in_out_ptr0, 2023-03-31T06:22:41.5642704Z const bool* in_ptr0, 2023-03-31T06:22:41.5642777Z long* out_ptr1) 2023-03-31T06:22:41.5642840Z { 2023-03-31T06:22:41.5642928Z auto out_ptr0 = in_out_ptr0; 2023-03-31T06:22:41.5642991Z { 2023-03-31T06:22:41.5643206Z { 2023-03-31T06:22:41.5643284Z long tmp2 = 0; 2023-03-31T06:22:41.5643347Z long tmp3 = 0; 2023-03-31T06:22:41.5643459Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5643529Z { 2023-03-31T06:22:41.5643663Z #pragma omp for reduction(+:tmp2) reduction(+:tmp3) 2023-03-31T06:22:41.5643822Z for(long i0=static_cast(0); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5643891Z { 2023-03-31T06:22:41.5644010Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5644126Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5644193Z tmp2 += tmp1; 2023-03-31T06:22:41.5644327Z tmp3 += tmp1; 2023-03-31T06:22:41.5644394Z } 2023-03-31T06:22:41.5644460Z } 2023-03-31T06:22:41.5644569Z out_ptr0[static_cast(0)] = tmp2; 2023-03-31T06:22:41.5644672Z out_ptr1[static_cast(0)] = tmp3; 2023-03-31T06:22:41.5644724Z } 2023-03-31T06:22:41.5644789Z } 2023-03-31T06:22:41.5644852Z { 2023-03-31T06:22:41.5644967Z auto tmp0 = out_ptr0[static_cast(0)]; 2023-03-31T06:22:41.5645078Z auto tmp3 = out_ptr1[static_cast(0)]; 2023-03-31T06:22:41.5645177Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.5645263Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5645334Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.5645442Z in_out_ptr0[static_cast(0)] = tmp4; 2023-03-31T06:22:41.5645544Z } 2023-03-31T06:22:41.5645608Z } 2023-03-31T06:22:41.5645697Z ''') 2023-03-31T06:22:41.5645703Z 2023-03-31T06:22:41.5645797Z async_compile.wait(globals()) 2023-03-31T06:22:41.5645876Z del async_compile 2023-03-31T06:22:41.5645986Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5646061Z wrapper = ( 2023-03-31T06:22:41.5646145Z ''' 2023-03-31T06:22:41.5646224Z #include 2023-03-31T06:22:41.5646303Z #include 2023-03-31T06:22:41.5646309Z 2023-03-31T06:22:41.5646403Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5646411Z 2023-03-31T06:22:41.5646509Z template 2023-03-31T06:22:41.5646621Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5646707Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5646833Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5646931Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5647059Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5647143Z return kernel_cpp; 2023-03-31T06:22:41.5647207Z } 2023-03-31T06:22:41.5647325Z std::vector call_165(std::vector args) { 2023-03-31T06:22:41.5647407Z at::Tensor arg0_1; 2023-03-31T06:22:41.5647480Z arg0_1 = args[0]; 2023-03-31T06:22:41.5647610Z auto buf0 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5647738Z auto buf1 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5647840Z auto buf2 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.5648176Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/bh/cbhxnrs677eld6stmmeaxrvck3rblcq3euj5farumhwrq74cvoqy.so"); 2023-03-31T06:22:41.5648370Z kernel_cpp_0((long*)(buf2.data_ptr()), (bool*)(arg0_1.data_ptr()), (long*)(buf1.data_ptr())); 2023-03-31T06:22:41.5648432Z arg0_1.reset(); 2023-03-31T06:22:41.5648539Z return std::vector({buf2}); 2023-03-31T06:22:41.5648605Z } 2023-03-31T06:22:41.5648691Z ''' 2023-03-31T06:22:41.5648754Z ) 2023-03-31T06:22:41.5648759Z 2023-03-31T06:22:41.5648842Z module = load_inline( 2023-03-31T06:22:41.5649120Z name='inline_extension_cp7rf5mjxmdow7lrdt3znnpdawtoczcwwkvrt2ifru3knbtnvopo', 2023-03-31T06:22:41.5649194Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5649314Z functions=['call_165'], 2023-03-31T06:22:41.5649699Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5649859Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5650573Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5650583Z 2023-03-31T06:22:41.5650661Z def _wrap_func(f): 2023-03-31T06:22:41.5650735Z def g(args): 2023-03-31T06:22:41.5650842Z return f(args) 2023-03-31T06:22:41.5650913Z return g 2023-03-31T06:22:41.5650998Z call = _wrap_func(module.call_165) 2023-03-31T06:22:41.5651003Z 2023-03-31T06:22:41.5651022Z 2023-03-31T06:22:41.5651105Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5651222Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5651352Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5651557Z arg0_1 = rand_strided({64, }, {1, }, device='cpu', dtype=torch.bool) 2023-03-31T06:22:41.5651669Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5651674Z 2023-03-31T06:22:41.5651679Z 2023-03-31T06:22:41.5651760Z if __name__ == "__main__": 2023-03-31T06:22:41.5651836Z import argparse 2023-03-31T06:22:41.5651954Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5652000Z 2023-03-31T06:22:41.5652099Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5652411Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5652751Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5653031Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5653126Z args = parser.parse_args() 2023-03-31T06:22:41.5653132Z 2023-03-31T06:22:41.5653221Z if args.benchmark_kernels: 2023-03-31T06:22:41.5653409Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5653475Z else: 2023-03-31T06:22:41.5653573Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5653667Z benchmark_compiled_module() 2023-03-31T06:22:41.5653675Z 2023-03-31T06:22:41.5653743Z if p: 2023-03-31T06:22:41.5653892Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5653989Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5654126Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5654132Z 2023-03-31T06:22:41.5654411Z [2023-03-31 06:21:36,718] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 140 2023-03-31T06:22:41.5654707Z [2023-03-31 06:21:36,745] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 141 2023-03-31T06:22:41.5655021Z [2023-03-31 06:21:36,762] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.uint8 not loaded as mask 2023-03-31T06:22:41.5655318Z [2023-03-31 06:21:36,762] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5655686Z [2023-03-31 06:21:36,762] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5655989Z [2023-03-31 06:21:36,763] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.uint8 not loaded as mask 2023-03-31T06:22:41.5656281Z [2023-03-31 06:21:36,763] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5656644Z [2023-03-31 06:21:36,763] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5656936Z [2023-03-31 06:21:36,763] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5657221Z [2023-03-31 06:21:36,764] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5657522Z [2023-03-31 06:21:36,765] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5657838Z [2023-03-31 06:21:36,765] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5658130Z [2023-03-31 06:21:36,765] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5658431Z [2023-03-31 06:21:36,765] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by store 2023-03-31T06:22:41.5658720Z [2023-03-31 06:21:36,766] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5659125Z [2023-03-31 06:21:58,022] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/wh/cwh2oxow6pu7dt6ftmhlq7j5dbqydbfbqfow3ypjuhwb4wdohqrz.py 2023-03-31T06:22:41.5659342Z [2023-03-31 06:21:58,022] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5659377Z 2023-03-31T06:22:41.5659478Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5659551Z import torch 2023-03-31T06:22:41.5659622Z import math 2023-03-31T06:22:41.5659696Z import random 2023-03-31T06:22:41.5659752Z import os 2023-03-31T06:22:41.5659825Z import tempfile 2023-03-31T06:22:41.5659944Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5659950Z 2023-03-31T06:22:41.5660069Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5660196Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5660331Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5660338Z 2023-03-31T06:22:41.5660418Z aten = torch.ops.aten 2023-03-31T06:22:41.5660546Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5660640Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5660645Z 2023-03-31T06:22:41.5660649Z 2023-03-31T06:22:41.5660792Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5661016Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5661121Z extern "C" void kernel(long* in_out_ptr0, 2023-03-31T06:22:41.5661227Z const unsigned char* in_ptr0, 2023-03-31T06:22:41.5661312Z long* out_ptr1) 2023-03-31T06:22:41.5661375Z { 2023-03-31T06:22:41.5661451Z auto out_ptr0 = in_out_ptr0; 2023-03-31T06:22:41.5661514Z { 2023-03-31T06:22:41.5661577Z { 2023-03-31T06:22:41.5661656Z long tmp2 = 0; 2023-03-31T06:22:41.5661760Z long tmp3 = 0; 2023-03-31T06:22:41.5661908Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5661974Z { 2023-03-31T06:22:41.5662093Z #pragma omp for reduction(+:tmp2) reduction(+:tmp3) 2023-03-31T06:22:41.5662249Z for(long i0=static_cast(0); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5662318Z { 2023-03-31T06:22:41.5662525Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5662638Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5662721Z tmp2 += tmp1; 2023-03-31T06:22:41.5662800Z tmp3 += tmp1; 2023-03-31T06:22:41.5662853Z } 2023-03-31T06:22:41.5662918Z } 2023-03-31T06:22:41.5663022Z out_ptr0[static_cast(0)] = tmp2; 2023-03-31T06:22:41.5663125Z out_ptr1[static_cast(0)] = tmp3; 2023-03-31T06:22:41.5663192Z } 2023-03-31T06:22:41.5663258Z } 2023-03-31T06:22:41.5663321Z { 2023-03-31T06:22:41.5663418Z auto tmp0 = out_ptr0[static_cast(0)]; 2023-03-31T06:22:41.5663523Z auto tmp3 = out_ptr1[static_cast(0)]; 2023-03-31T06:22:41.5663621Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.5663707Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5663789Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.5663901Z in_out_ptr0[static_cast(0)] = tmp4; 2023-03-31T06:22:41.5663964Z } 2023-03-31T06:22:41.5664014Z } 2023-03-31T06:22:41.5664099Z ''') 2023-03-31T06:22:41.5664141Z 2023-03-31T06:22:41.5664233Z async_compile.wait(globals()) 2023-03-31T06:22:41.5664307Z del async_compile 2023-03-31T06:22:41.5664428Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5664500Z wrapper = ( 2023-03-31T06:22:41.5664568Z ''' 2023-03-31T06:22:41.5664643Z #include 2023-03-31T06:22:41.5664719Z #include 2023-03-31T06:22:41.5664727Z 2023-03-31T06:22:41.5664825Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5664830Z 2023-03-31T06:22:41.5664922Z template 2023-03-31T06:22:41.5665046Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5665131Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5665255Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5665339Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5665499Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5665577Z return kernel_cpp; 2023-03-31T06:22:41.5665640Z } 2023-03-31T06:22:41.5665773Z std::vector call_166(std::vector args) { 2023-03-31T06:22:41.5665851Z at::Tensor arg0_1; 2023-03-31T06:22:41.5665924Z arg0_1 = args[0]; 2023-03-31T06:22:41.5666041Z auto buf0 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5666165Z auto buf1 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5666268Z auto buf2 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.5666571Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/6n/c6nhoqpigzcq33ym7r3ivxjfs265kv6dw2a4exgun5dssduno7jh.so"); 2023-03-31T06:22:41.5666743Z kernel_cpp_0((long*)(buf2.data_ptr()), (unsigned char*)(arg0_1.data_ptr()), (long*)(buf1.data_ptr())); 2023-03-31T06:22:41.5666819Z arg0_1.reset(); 2023-03-31T06:22:41.5666924Z return std::vector({buf2}); 2023-03-31T06:22:41.5666986Z } 2023-03-31T06:22:41.5667055Z ''' 2023-03-31T06:22:41.5667117Z ) 2023-03-31T06:22:41.5667124Z 2023-03-31T06:22:41.5667201Z module = load_inline( 2023-03-31T06:22:41.5667475Z name='inline_extension_cjsoiu6zykiqormip4kabpd6kedfoxc7s4qi7qpai64lk33v2fbc', 2023-03-31T06:22:41.5667561Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5667681Z functions=['call_166'], 2023-03-31T06:22:41.5668091Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5668262Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5668968Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5668991Z 2023-03-31T06:22:41.5669056Z def _wrap_func(f): 2023-03-31T06:22:41.5669126Z def g(args): 2023-03-31T06:22:41.5669201Z return f(args) 2023-03-31T06:22:41.5669268Z return g 2023-03-31T06:22:41.5669364Z call = _wrap_func(module.call_166) 2023-03-31T06:22:41.5669369Z 2023-03-31T06:22:41.5669374Z 2023-03-31T06:22:41.5669468Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5669589Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5669706Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5669907Z arg0_1 = rand_strided({64, }, {1, }, device='cpu', dtype=torch.uint8) 2023-03-31T06:22:41.5670020Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5670025Z 2023-03-31T06:22:41.5670030Z 2023-03-31T06:22:41.5670110Z if __name__ == "__main__": 2023-03-31T06:22:41.5670190Z import argparse 2023-03-31T06:22:41.5670320Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5670326Z 2023-03-31T06:22:41.5670466Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5670777Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5671116Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5671381Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5671477Z args = parser.parse_args() 2023-03-31T06:22:41.5671482Z 2023-03-31T06:22:41.5671570Z if args.benchmark_kernels: 2023-03-31T06:22:41.5671759Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5671826Z else: 2023-03-31T06:22:41.5671937Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5672061Z benchmark_compiled_module() 2023-03-31T06:22:41.5672067Z 2023-03-31T06:22:41.5672122Z if p: 2023-03-31T06:22:41.5672272Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5672367Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5672504Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5672509Z 2023-03-31T06:22:41.5672796Z [2023-03-31 06:21:58,022] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 141 2023-03-31T06:22:41.5673069Z [2023-03-31 06:21:58,050] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 142 2023-03-31T06:22:41.5673377Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.5673672Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5674045Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5674351Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int32 not supported by load 2023-03-31T06:22:41.5674628Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5675019Z [2023-03-31 06:21:58,066] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: reduction: dtype torch.int64, src_dtype torch.int64, reduction_type sum 2023-03-31T06:22:41.5675335Z [2023-03-31 06:21:58,067] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5675618Z [2023-03-31 06:21:58,067] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: to_dtype: dtype torch.int64 2023-03-31T06:22:41.5675921Z [2023-03-31 06:21:58,069] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5676204Z [2023-03-31 06:21:58,069] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5676507Z [2023-03-31 06:21:58,069] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by load 2023-03-31T06:22:41.5676812Z [2023-03-31 06:21:58,069] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: torch.int64 not supported by store 2023-03-31T06:22:41.5677103Z [2023-03-31 06:21:58,070] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: constant dtype: torch.int32 2023-03-31T06:22:41.5677511Z [2023-03-31 06:22:19,278] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/ze/czerruv4wekbwxhuc3u4deagf4dokuw65vhiqukvneb5rsll5j3n.py 2023-03-31T06:22:41.5677720Z [2023-03-31 06:22:19,279] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5677741Z 2023-03-31T06:22:41.5677858Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5677932Z import torch 2023-03-31T06:22:41.5678003Z import math 2023-03-31T06:22:41.5678074Z import random 2023-03-31T06:22:41.5678142Z import os 2023-03-31T06:22:41.5678217Z import tempfile 2023-03-31T06:22:41.5678323Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5678341Z 2023-03-31T06:22:41.5678451Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5678577Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5678712Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5678717Z 2023-03-31T06:22:41.5678797Z aten = torch.ops.aten 2023-03-31T06:22:41.5678935Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5679031Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5679067Z 2023-03-31T06:22:41.5679071Z 2023-03-31T06:22:41.5679212Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5679436Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5679527Z extern "C" void kernel(long* in_out_ptr0, 2023-03-31T06:22:41.5679617Z const int* in_ptr0, 2023-03-31T06:22:41.5679700Z long* out_ptr1) 2023-03-31T06:22:41.5679765Z { 2023-03-31T06:22:41.5679853Z auto out_ptr0 = in_out_ptr0; 2023-03-31T06:22:41.5679916Z { 2023-03-31T06:22:41.5679970Z { 2023-03-31T06:22:41.5680048Z long tmp2 = 0; 2023-03-31T06:22:41.5680123Z long tmp3 = 0; 2023-03-31T06:22:41.5680234Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5680300Z { 2023-03-31T06:22:41.5680432Z #pragma omp for reduction(+:tmp2) reduction(+:tmp3) 2023-03-31T06:22:41.5680590Z for(long i0=static_cast(0); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5680648Z { 2023-03-31T06:22:41.5680766Z auto tmp0 = in_ptr0[static_cast(i0)]; 2023-03-31T06:22:41.5680876Z auto tmp1 = static_cast(tmp0); 2023-03-31T06:22:41.5680957Z tmp2 += tmp1; 2023-03-31T06:22:41.5681034Z tmp3 += tmp1; 2023-03-31T06:22:41.5681099Z } 2023-03-31T06:22:41.5681166Z } 2023-03-31T06:22:41.5681259Z out_ptr0[static_cast(0)] = tmp2; 2023-03-31T06:22:41.5681427Z out_ptr1[static_cast(0)] = tmp3; 2023-03-31T06:22:41.5681517Z } 2023-03-31T06:22:41.5681580Z } 2023-03-31T06:22:41.5681642Z { 2023-03-31T06:22:41.5681753Z auto tmp0 = out_ptr0[static_cast(0)]; 2023-03-31T06:22:41.5681861Z auto tmp3 = out_ptr1[static_cast(0)]; 2023-03-31T06:22:41.5681946Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.5682035Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5682119Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.5682227Z in_out_ptr0[static_cast(0)] = tmp4; 2023-03-31T06:22:41.5682293Z } 2023-03-31T06:22:41.5682355Z } 2023-03-31T06:22:41.5682437Z ''') 2023-03-31T06:22:41.5682443Z 2023-03-31T06:22:41.5682521Z async_compile.wait(globals()) 2023-03-31T06:22:41.5682596Z del async_compile 2023-03-31T06:22:41.5682719Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5682792Z wrapper = ( 2023-03-31T06:22:41.5682875Z ''' 2023-03-31T06:22:41.5682952Z #include 2023-03-31T06:22:41.5683161Z #include 2023-03-31T06:22:41.5683166Z 2023-03-31T06:22:41.5683247Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5683252Z 2023-03-31T06:22:41.5683346Z template 2023-03-31T06:22:41.5683470Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5683556Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5683684Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5683781Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5683964Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5684031Z return kernel_cpp; 2023-03-31T06:22:41.5684094Z } 2023-03-31T06:22:41.5684229Z std::vector call_167(std::vector args) { 2023-03-31T06:22:41.5684308Z at::Tensor arg0_1; 2023-03-31T06:22:41.5684382Z arg0_1 = args[0]; 2023-03-31T06:22:41.5684511Z auto buf0 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5684640Z auto buf1 = at::empty_strided({}, {}, at::ScalarType::Long); 2023-03-31T06:22:41.5684745Z auto buf2 = buf0; buf0.reset(); // reuse 2023-03-31T06:22:41.5685023Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/le/cle4d5bajynvgx5s7vbf7sxhdm4js2ewq2l5sdlht2odxr57ogqt.so"); 2023-03-31T06:22:41.5685185Z kernel_cpp_0((long*)(buf2.data_ptr()), (int*)(arg0_1.data_ptr()), (long*)(buf1.data_ptr())); 2023-03-31T06:22:41.5685303Z arg0_1.reset(); 2023-03-31T06:22:41.5685410Z return std::vector({buf2}); 2023-03-31T06:22:41.5685479Z } 2023-03-31T06:22:41.5685565Z ''' 2023-03-31T06:22:41.5685628Z ) 2023-03-31T06:22:41.5685633Z 2023-03-31T06:22:41.5685698Z module = load_inline( 2023-03-31T06:22:41.5685979Z name='inline_extension_ccjk5omgiohdvx6lvsuqkpbdlepnuee2lhwirmzbiaijw7dqv3ml', 2023-03-31T06:22:41.5686066Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5686186Z functions=['call_167'], 2023-03-31T06:22:41.5686569Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5686727Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5687437Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5687446Z 2023-03-31T06:22:41.5687523Z def _wrap_func(f): 2023-03-31T06:22:41.5687595Z def g(args): 2023-03-31T06:22:41.5687658Z return f(args) 2023-03-31T06:22:41.5687727Z return g 2023-03-31T06:22:41.5687823Z call = _wrap_func(module.call_167) 2023-03-31T06:22:41.5687828Z 2023-03-31T06:22:41.5687832Z 2023-03-31T06:22:41.5687963Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5688119Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5688249Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5688455Z arg0_1 = rand_strided({64, }, {1, }, device='cpu', dtype=torch.int32) 2023-03-31T06:22:41.5688570Z print_performance(lambda: call([arg0_1])) 2023-03-31T06:22:41.5688575Z 2023-03-31T06:22:41.5688582Z 2023-03-31T06:22:41.5688648Z if __name__ == "__main__": 2023-03-31T06:22:41.5688726Z import argparse 2023-03-31T06:22:41.5688857Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5688865Z 2023-03-31T06:22:41.5688978Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5689286Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5689630Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5689909Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5690004Z args = parser.parse_args() 2023-03-31T06:22:41.5690009Z 2023-03-31T06:22:41.5690101Z if args.benchmark_kernels: 2023-03-31T06:22:41.5690276Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5690346Z else: 2023-03-31T06:22:41.5690459Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5690557Z benchmark_compiled_module() 2023-03-31T06:22:41.5690563Z 2023-03-31T06:22:41.5690662Z if p: 2023-03-31T06:22:41.5690814Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5690910Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5691038Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5691044Z 2023-03-31T06:22:41.5691331Z [2023-03-31 06:22:19,279] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 142 2023-03-31T06:22:41.5691411Z PASSED [ 94%] 2023-03-31T06:22:41.5691809Z inductor/test_cpp_wrapper.py::TestCppWrapper::test_transpose_cpu [2023-03-31 06:22:19,316] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 143 2023-03-31T06:22:41.5692158Z [2023-03-31 06:22:19,335] torch._inductor.codegen.cpp.__schedule: [DEBUG] Disabled vectorization: load with indirect indexing or non-contigous: i0 + 8*i1 2023-03-31T06:22:41.5692598Z [2023-03-31 06:22:40,630] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_jenkins/jr/cjr33uqouhb7gu32dsrcczuqbj6mqngrarlgtyrobkzkxwnuc3ge.py 2023-03-31T06:22:41.5692820Z [2023-03-31 06:22:40,631] torch._inductor.graph.__output_code: [DEBUG] Output code: 2023-03-31T06:22:41.5692826Z 2023-03-31T06:22:41.5692925Z from ctypes import c_void_p, c_long 2023-03-31T06:22:41.5692998Z import torch 2023-03-31T06:22:41.5693058Z import math 2023-03-31T06:22:41.5693132Z import random 2023-03-31T06:22:41.5693204Z import os 2023-03-31T06:22:41.5693281Z import tempfile 2023-03-31T06:22:41.5693400Z from torch._inductor.utils import maybe_profile 2023-03-31T06:22:41.5693406Z 2023-03-31T06:22:41.5693526Z from torch import empty_strided, as_strided, device 2023-03-31T06:22:41.5693652Z from torch._inductor.codecache import AsyncCompile 2023-03-31T06:22:41.5693773Z from torch._inductor.select_algorithm import extern_kernels 2023-03-31T06:22:41.5693795Z 2023-03-31T06:22:41.5693863Z aten = torch.ops.aten 2023-03-31T06:22:41.5694005Z assert_size_stride = torch._C._dynamo.guards.assert_size_stride 2023-03-31T06:22:41.5694102Z async_compile = AsyncCompile() 2023-03-31T06:22:41.5694107Z 2023-03-31T06:22:41.5694112Z 2023-03-31T06:22:41.5694251Z kernel_cpp_0 = async_compile.cpp(''' 2023-03-31T06:22:41.5694471Z #include "/tmp/torchinductor_jenkins/hw/chwr6vy6e6sd25sfh42qtywkuf2emodexm2aomp3lbrcxwznfwyi.h" 2023-03-31T06:22:41.5694609Z extern "C" void kernel(const float* in_ptr0, 2023-03-31T06:22:41.5694729Z const float* in_ptr1, 2023-03-31T06:22:41.5694802Z float* out_ptr0, 2023-03-31T06:22:41.5694888Z float* out_ptr1) 2023-03-31T06:22:41.5694950Z { 2023-03-31T06:22:41.5695050Z #pragma omp parallel num_threads(4) 2023-03-31T06:22:41.5695116Z { 2023-03-31T06:22:41.5695182Z { 2023-03-31T06:22:41.5695265Z #pragma omp for 2023-03-31T06:22:41.5695405Z for(long i0=static_cast(0); i0(1); i0+=static_cast(1)) 2023-03-31T06:22:41.5695470Z { 2023-03-31T06:22:41.5695557Z #pragma GCC ivdep 2023-03-31T06:22:41.5695711Z for(long i1=static_cast(0); i1(1); i1+=static_cast(1)) 2023-03-31T06:22:41.5695782Z { 2023-03-31T06:22:41.5695900Z float tmp0[8*8] __attribute__ ((aligned (8))); 2023-03-31T06:22:41.5696084Z at::vec::transpose_mxn(in_ptr0 + static_cast((8*i0) + (64*i1)), static_cast(8), tmp0, 8); 2023-03-31T06:22:41.5696203Z float tmp4[8*8] __attribute__ ((aligned (8))); 2023-03-31T06:22:41.5696308Z for (long i0_inner = 0; i0_inner < 8; i0_inner++) 2023-03-31T06:22:41.5696377Z { 2023-03-31T06:22:41.5696540Z auto tmp1 = at::vec::Vectorized::loadu(tmp0 + static_cast(8*i0_inner)); 2023-03-31T06:22:41.5696746Z auto tmp2 = at::vec::Vectorized::loadu(in_ptr1 + static_cast((8*i0_inner) + (8*i1) + (64*i0))); 2023-03-31T06:22:41.5696847Z auto tmp3 = tmp1 + tmp2; 2023-03-31T06:22:41.5696971Z tmp3.store(tmp4 + static_cast(8*i0_inner)); 2023-03-31T06:22:41.5697041Z } 2023-03-31T06:22:41.5697222Z at::vec::transpose_mxn(tmp4, 8, out_ptr0 + static_cast((8*i0) + (64*i1)), static_cast(8)); 2023-03-31T06:22:41.5697293Z } 2023-03-31T06:22:41.5697366Z #pragma GCC ivdep 2023-03-31T06:22:41.5697519Z for(long i1=static_cast(8); i1(8); i1+=static_cast(1)) 2023-03-31T06:22:41.5697586Z { 2023-03-31T06:22:41.5697701Z for (long i0_inner = 0; i0_inner < 8; i0_inner++) 2023-03-31T06:22:41.5697798Z { 2023-03-31T06:22:41.5697937Z auto tmp0 = in_ptr0[static_cast(i0_inner + (8*i0) + (8*i1))]; 2023-03-31T06:22:41.5698077Z auto tmp1 = in_ptr1[static_cast(i1 + (8*i0_inner) + (64*i0))]; 2023-03-31T06:22:41.5698163Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.5698292Z out_ptr0[static_cast(i0_inner + (8*i0) + (8*i1))] = tmp2; 2023-03-31T06:22:41.5698359Z } 2023-03-31T06:22:41.5698428Z } 2023-03-31T06:22:41.5698495Z } 2023-03-31T06:22:41.5698577Z #pragma omp for 2023-03-31T06:22:41.5698728Z for(long i0=static_cast(8); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.5698782Z { 2023-03-31T06:22:41.5698866Z #pragma GCC ivdep 2023-03-31T06:22:41.5699018Z for(long i1=static_cast(0); i1(8); i1+=static_cast(1)) 2023-03-31T06:22:41.5699086Z { 2023-03-31T06:22:41.5699209Z auto tmp0 = in_ptr0[static_cast(i0 + (8*i1))]; 2023-03-31T06:22:41.5699330Z auto tmp1 = in_ptr1[static_cast(i1 + (8*i0))]; 2023-03-31T06:22:41.5699426Z auto tmp2 = tmp0 + tmp1; 2023-03-31T06:22:41.5699531Z out_ptr0[static_cast(i0 + (8*i1))] = tmp2; 2023-03-31T06:22:41.5699596Z } 2023-03-31T06:22:41.5699660Z } 2023-03-31T06:22:41.5699724Z } 2023-03-31T06:22:41.5699790Z { 2023-03-31T06:22:41.5699901Z #pragma omp for 2023-03-31T06:22:41.5700078Z for(long i0=static_cast(0); i0(8); i0+=static_cast(1)) 2023-03-31T06:22:41.5700130Z { 2023-03-31T06:22:41.5700289Z auto tmp0 = at::vec::Vectorized::loadu(in_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5700431Z auto tmp1 = at::vec::Vectorized(static_cast(2)); 2023-03-31T06:22:41.5700525Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5700667Z auto tmp3 = at::vec::Vectorized(static_cast(10)); 2023-03-31T06:22:41.5700757Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.5700877Z tmp4.store(out_ptr1 + static_cast(8*i0)); 2023-03-31T06:22:41.5700942Z } 2023-03-31T06:22:41.5701032Z #pragma omp for simd simdlen(4) 2023-03-31T06:22:41.5701184Z for(long i0=static_cast(64); i0(64); i0+=static_cast(1)) 2023-03-31T06:22:41.5701252Z { 2023-03-31T06:22:41.5701366Z auto tmp0 = in_ptr1[static_cast(i0)]; 2023-03-31T06:22:41.5701471Z auto tmp1 = static_cast(2); 2023-03-31T06:22:41.5701560Z auto tmp2 = tmp0 * tmp1; 2023-03-31T06:22:41.5701666Z auto tmp3 = static_cast(10); 2023-03-31T06:22:41.5701741Z auto tmp4 = tmp2 + tmp3; 2023-03-31T06:22:41.5701852Z out_ptr1[static_cast(i0)] = tmp4; 2023-03-31T06:22:41.5701917Z } 2023-03-31T06:22:41.5701981Z } 2023-03-31T06:22:41.5702070Z } 2023-03-31T06:22:41.5702133Z } 2023-03-31T06:22:41.5702216Z ''') 2023-03-31T06:22:41.5702222Z 2023-03-31T06:22:41.5702303Z async_compile.wait(globals()) 2023-03-31T06:22:41.5702466Z del async_compile 2023-03-31T06:22:41.5702596Z from torch.utils.cpp_extension import load_inline 2023-03-31T06:22:41.5702668Z wrapper = ( 2023-03-31T06:22:41.5702752Z ''' 2023-03-31T06:22:41.5702833Z #include 2023-03-31T06:22:41.5702913Z #include 2023-03-31T06:22:41.5702919Z 2023-03-31T06:22:41.5703000Z typedef at::BFloat16 bfloat16; 2023-03-31T06:22:41.5703005Z 2023-03-31T06:22:41.5703099Z template 2023-03-31T06:22:41.5703221Z KernelFunc load_cpp_kernel(const char* so_filename) { 2023-03-31T06:22:41.5703308Z KernelFunc kernel_cpp; 2023-03-31T06:22:41.5703467Z auto kernel_cpp_lib = dlopen(so_filename, RTLD_NOW); 2023-03-31T06:22:41.5703566Z assert(kernel_cpp_lib != nullptr); 2023-03-31T06:22:41.5703695Z *(void **) (&kernel_cpp) = dlsym(kernel_cpp_lib, "kernel"); 2023-03-31T06:22:41.5703761Z return kernel_cpp; 2023-03-31T06:22:41.5703823Z } 2023-03-31T06:22:41.5703953Z std::vector call_168(std::vector args) { 2023-03-31T06:22:41.5704038Z at::Tensor arg0_1, arg1_1; 2023-03-31T06:22:41.5704110Z arg0_1 = args[0]; 2023-03-31T06:22:41.5704183Z arg1_1 = args[1]; 2023-03-31T06:22:41.5704319Z auto buf0 = at::empty_strided({8, 8}, {1, 8}, at::ScalarType::Float); 2023-03-31T06:22:41.5704458Z auto buf1 = at::empty_strided({8, 8}, {1, 8}, at::ScalarType::Float); 2023-03-31T06:22:41.5704762Z static auto kernel_cpp_0 = load_cpp_kernel("/tmp/torchinductor_jenkins/ez/cezyew45rhdo7bekxdrl46cexahsgaysino2ast4ic6ukl6ejcod.so"); 2023-03-31T06:22:41.5704959Z kernel_cpp_0((float*)(arg0_1.data_ptr()), (float*)(arg1_1.data_ptr()), (float*)(buf0.data_ptr()), (float*)(buf1.data_ptr())); 2023-03-31T06:22:41.5705034Z arg0_1.reset(); 2023-03-31T06:22:41.5705108Z arg1_1.reset(); 2023-03-31T06:22:41.5705220Z return std::vector({buf0, buf1}); 2023-03-31T06:22:41.5705282Z } 2023-03-31T06:22:41.5705362Z ''' 2023-03-31T06:22:41.5705411Z ) 2023-03-31T06:22:41.5705429Z 2023-03-31T06:22:41.5705494Z module = load_inline( 2023-03-31T06:22:41.5705772Z name='inline_extension_csjdfz6qxdyb2hxo3eiofc4bvv5rfvpt4wjcwxhmoc7dtkpfzi5p', 2023-03-31T06:22:41.5705884Z cpp_sources=[wrapper], 2023-03-31T06:22:41.5706042Z functions=['call_168'], 2023-03-31T06:22:41.5706421Z extra_cflags=['-std=c++17 -Wno-unused-variable -O3 -ffast-math -fno-finite-math-only -march=native -fopenmp -Wall -D C10_USING_CUSTOM_GENERATED_MACROS'], 2023-03-31T06:22:41.5706580Z extra_ldflags=['-shared -fPIC -lgomp'], 2023-03-31T06:22:41.5707298Z extra_include_paths=['-I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8']) 2023-03-31T06:22:41.5707307Z 2023-03-31T06:22:41.5707382Z def _wrap_func(f): 2023-03-31T06:22:41.5707442Z def g(args): 2023-03-31T06:22:41.5707515Z return f(args) 2023-03-31T06:22:41.5707582Z return g 2023-03-31T06:22:41.5707679Z call = _wrap_func(module.call_168) 2023-03-31T06:22:41.5707686Z 2023-03-31T06:22:41.5707691Z 2023-03-31T06:22:41.5707790Z def benchmark_compiled_module(): 2023-03-31T06:22:41.5707908Z from torch._dynamo.testing import rand_strided 2023-03-31T06:22:41.5708033Z from torch._inductor.utils import print_performance 2023-03-31T06:22:41.5708239Z arg0_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5708433Z arg1_1 = rand_strided({8, 8}, {8, 1}, device='cpu', dtype=torch.float32) 2023-03-31T06:22:41.5708550Z print_performance(lambda: call([arg0_1, arg1_1])) 2023-03-31T06:22:41.5708556Z 2023-03-31T06:22:41.5708595Z 2023-03-31T06:22:41.5708678Z if __name__ == "__main__": 2023-03-31T06:22:41.5708755Z import argparse 2023-03-31T06:22:41.5708886Z from torch._inductor.utils import benchmark_all_kernels 2023-03-31T06:22:41.5708892Z 2023-03-31T06:22:41.5709005Z parser = argparse.ArgumentParser() 2023-03-31T06:22:41.5709317Z parser.add_argument("--benchmark-kernels", "-k", action="store_true", help="Whether to benchmark each individual kernels") 2023-03-31T06:22:41.5709659Z parser.add_argument("--benchmark-all-configs", "-c", action="store_true", help="Whether to benchmark each individual config for a kernel") 2023-03-31T06:22:41.5709935Z parser.add_argument("--profile", "-p", action="store_true", help="Whether to profile the compiled module") 2023-03-31T06:22:41.5710043Z args = parser.parse_args() 2023-03-31T06:22:41.5710049Z 2023-03-31T06:22:41.5710138Z if args.benchmark_kernels: 2023-03-31T06:22:41.5710331Z benchmark_all_kernels('None', args.benchmark_all_configs) 2023-03-31T06:22:41.5710398Z else: 2023-03-31T06:22:41.5710509Z with maybe_profile(args.profile) as p: 2023-03-31T06:22:41.5710603Z benchmark_compiled_module() 2023-03-31T06:22:41.5710609Z 2023-03-31T06:22:41.5710676Z if p: 2023-03-31T06:22:41.5710814Z path = f"{tempfile.gettempdir()}/compiled_module_profile.json" 2023-03-31T06:22:41.5710911Z p.export_chrome_trace(path) 2023-03-31T06:22:41.5711051Z print(f"Chrome trace for the profile is written to {path}") 2023-03-31T06:22:41.5711056Z 2023-03-31T06:22:41.5711337Z [2023-03-31 06:22:40,631] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 143 2023-03-31T06:22:41.5711412Z PASSED [100%] 2023-03-31T06:22:41.5711418Z 2023-03-31T06:22:41.5711798Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cpp_wrapper/inductor.test_cpp_wrapper-28bf4fee1260856b.xml - 2023-03-31T06:22:41.5711917Z ======================= 18 passed in 3070.14s (0:51:10) ======================== 2023-03-31T06:22:41.5712095Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:22:41.5712101Z 2023-03-31T06:22:41.5712527Z ##[endgroup] 2023-03-31T06:22:41.5712901Z FINISHED PRINTING LOG FILE of inductor/test_cpp_wrapper (/var/lib/jenkins/workspace/test/test-reports/inductor-test_cpp_wrapper_vfreyo6s.log) 2023-03-31T06:22:41.5712933Z 2023-03-31T06:22:41.7527352Z Running test_linalg ... [2023-03-31 06:22:41.752358] 2023-03-31T06:22:41.7531240Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_linalg.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:22:41.752841] 2023-03-31T06:24:13.3827637Z 2023-03-31T06:24:13.3828164Z Expand the folded group to see the log file of test_linalg 2023-03-31T06:24:13.3829119Z ##[group]PRINTING LOG FILE of test_linalg (/var/lib/jenkins/workspace/test/test-reports/test_linalg_930rev50.log) 2023-03-31T06:24:13.3830120Z Test results will be stored in test-reports/python-pytest/test_linalg/test_linalg-7b54af76081f9243.xml 2023-03-31T06:24:13.3832374Z ============================= test session starts ============================== 2023-03-31T06:24:13.3832903Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:24:13.3833296Z cachedir: .pytest_cache 2023-03-31T06:24:13.3833760Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:24:13.3834208Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:24:13.3834730Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:24:13.3835156Z collecting ... collected 744 items 2023-03-31T06:24:13.3904130Z Running 744 items in this shard: test/test_linalg.py::TestLinalgCPU::test__int_mm_errors_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_False_cpu, test/test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_True_cpu, test/test_linalg.py::TestLinalgCPU::test_addbmm_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_addbmm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_addbmm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_addbmm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addbmm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addmm_baddbmm_overflow_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_addmm_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_addmm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_addmm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_addmm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addmm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addmm_sizes_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addmm_sizes_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addmv_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_addmv_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_addmv_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_addmv_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addmv_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addmv_rowmajor_colmajor_incx_incy_lda_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addmv_rowmajor_colmajor_incx_incy_lda_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addr_bool_cpu_bool, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int16, test/test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int32, test/test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int8, test/test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_uint8, test/test_linalg.py::TestLinalgCPU::test_addr_type_promotion_cpu, test/test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_baddbmm_nan_input_with_zero_beta_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_bfloat16_accumulation_with_ref_path_cpu, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int16, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int32, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int8, test/test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_uint8, test/test_linalg.py::TestLinalgCPU::test_blas_empty_cpu, test/test_linalg.py::TestLinalgCPU::test_blas_mv_large_input_cpu, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_bmm_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_bmm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_bmm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_bmm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_bmm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_broadcast_batched_matmul_cpu, test/test_linalg.py::TestLinalgCPU::test_broadcast_fused_matmul_cpu, test/test_linalg.py::TestLinalgCPU::test_chain_matmul_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cond_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cond_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cond_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cond_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int16, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int32, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int8, test/test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_uint8, test/test_linalg.py::TestLinalgCPU::test_cross_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cross_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_cross_with_and_without_dim_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_cross_with_and_without_dim_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_det_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_det_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_det_logdet_slogdet_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_det_logdet_slogdet_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_dot_invalid_args_cpu, test/test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eig_check_magma_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eig_numpy_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eig_numpy_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eig_removed_error_cpu, test/test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigh_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigh_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigh_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigh_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigvals_numpy_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigvals_numpy_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_einsum_corner_cases_cpu, test/test_linalg.py::TestLinalgCPU::test_einsum_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_einsum_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_einsum_error_cases_cpu, test/test_linalg.py::TestLinalgCPU::test_einsum_random_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_einsum_random_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_einsum_sublist_format_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_einsum_sublist_format_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_geqrf_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_geqrf_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_geqrf_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_geqrf_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_householder_product_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_householder_product_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_householder_product_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_householder_product_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_householder_product_errors_and_warnings_cpu, test/test_linalg.py::TestLinalgCPU::test_inner_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inner_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_invariance_error_spectral_decompositions_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inverse_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inverse_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inverse_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inverse_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_kron_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_kron_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_kron_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_kron_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lapack_empty_cpu, test/test_linalg.py::TestLinalgCPU::test_large_bmm_backward_cpu, test/test_linalg.py::TestLinalgCPU::test_large_bmm_mm_backward_cpu, test/test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_cross_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_cross_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_cross_with_and_without_dim_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_cross_with_and_without_dim_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_batch_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_batch_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_no_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_utils_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_utils_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_qr_autograd_errors_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_linear_algebra_scalar_raises_cpu, test/test_linalg.py::TestLinalgCPU::test_lobpcg_basic_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lobpcg_ortho_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lobpcg_scipy_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lobpcg_torchscript_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lstsq_removed_error_cpu, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_lu_unpack_check_input_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matmul_45724_cpu, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_matrix_norm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_norm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_power_negative_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_power_negative_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_power_non_negative_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_power_non_negative_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_rtol_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_matrix_rank_removed_error_cpu, test/test_linalg.py::TestLinalgCPU::test_mm_bmm_non_memory_dense_cpu, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_int32, test/test_linalg.py::TestLinalgCPU::test_mm_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_multi_dot_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_multi_dot_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_multi_dot_errors_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_bfloat16_and_half_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_norm_bfloat16_and_half_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_norm_complex_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_norm_complex_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_norm_complex_old_cpu, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_norm_errors_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_errors_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_norm_extreme_values_cpu, test/test_linalg.py::TestLinalgCPU::test_norm_fastpaths_cpu, test/test_linalg.py::TestLinalgCPU::test_norm_fro_2_equivalence_old_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_fused_type_promotion_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_norm_fused_type_promotion_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_norm_old_cpu, test/test_linalg.py::TestLinalgCPU::test_norm_old_nan_propagation_cpu, test/test_linalg.py::TestLinalgCPU::test_norm_vector_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_vector_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_nuclear_norm_axes_small_brute_force_old_cpu, test/test_linalg.py::TestLinalgCPU::test_nuclear_norm_exceptions_old_cpu, test/test_linalg.py::TestLinalgCPU::test_nuclear_norm_out_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_nuclear_norm_out_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_many_batches_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_ormqr_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_ormqr_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_ormqr_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_ormqr_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_bool, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_int16, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_int32, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_int64, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_int8, test/test_linalg.py::TestLinalgCPU::test_outer_cpu_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_ger_addr_legacy_tests_cpu, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_uint8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_bfloat16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_bool, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_complex128, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_complex64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int16, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int32, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int64, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int8, test/test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_uint8, test/test_linalg.py::TestLinalgCPU::test_pca_lowrank_cpu, test/test_linalg.py::TestLinalgCPU::test_permute_matmul_cpu, test/test_linalg.py::TestLinalgCPU::test_pinv_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_pinv_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_pinv_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_pinv_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_pinverse_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_pinverse_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_pinverse_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_pinverse_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_preferred_linalg_library_cpu, test/test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_qr_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_qr_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_qr_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_qr_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_qr_error_cases_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_renorm_cpu, test/test_linalg.py::TestLinalgCPU::test_renorm_ps_cpu, test/test_linalg.py::TestLinalgCPU::test_slogdet_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_slogdet_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_slogdet_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_slogdet_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_solve_removed_error_cpu, test/test_linalg.py::TestLinalgCPU::test_strided_mm_bmm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_strided_mm_bmm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_svd_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_svd_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_svd_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_svd_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_svd_lowrank_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_symeig_removed_error_cpu, test/test_linalg.py::TestLinalgCPU::test_tensordot_cpu, test/test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_tensorsolve_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_vdot_invalid_args_cpu, test/test_linalg.py::TestLinalgCPU::test_vdot_vs_numpy_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_vdot_vs_numpy_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_bfloat16, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_complex128, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_complex64, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float16, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float32, test/test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float64, test/test_linalg.py::TestLinalgCPU::test_vector_norm_dim_tuple_arg_cpu, test/test_linalg.py::TestLinalgCPU::test_vector_norm_extreme_values_cpu 2023-03-31T06:24:13.3966442Z 2023-03-31T06:24:13.3966654Z test_linalg.py::TestLinalgCPU::test__int_mm_errors_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3967067Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3967518Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3967969Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3968401Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3968829Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3969391Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 0%] 2023-03-31T06:24:13.3969825Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3970268Z test_linalg.py::TestLinalgCPU::test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3970690Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3971550Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3971979Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3972459Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3972881Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 1%] 2023-03-31T06:24:13.3973314Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 2%] 2023-03-31T06:24:13.3973740Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_False_cpu SKIPPED (Only runs on cuda) [ 2%] 2023-03-31T06:24:13.3974171Z test_linalg.py::TestLinalgCPU::test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_True_cpu SKIPPED (Only runs on cuda) [ 2%] 2023-03-31T06:24:13.3974534Z test_linalg.py::TestLinalgCPU::test_addbmm_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:24:13.3974871Z test_linalg.py::TestLinalgCPU::test_addbmm_cpu_complex128 PASSED [ 2%] 2023-03-31T06:24:13.3975199Z test_linalg.py::TestLinalgCPU::test_addbmm_cpu_complex64 PASSED [ 2%] 2023-03-31T06:24:13.3975523Z test_linalg.py::TestLinalgCPU::test_addbmm_cpu_float32 PASSED [ 2%] 2023-03-31T06:24:13.3975828Z test_linalg.py::TestLinalgCPU::test_addbmm_cpu_float64 PASSED [ 2%] 2023-03-31T06:24:13.3976149Z test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:24:13.3976531Z test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_float32 PASSED [ 3%] 2023-03-31T06:24:13.3976891Z test_linalg.py::TestLinalgCPU::test_addmm_activation_cpu_float64 PASSED [ 3%] 2023-03-31T06:24:13.3977255Z test_linalg.py::TestLinalgCPU::test_addmm_baddbmm_overflow_cpu_float16 SKIPPED (Only runs on cuda) [ 3%] 2023-03-31T06:24:13.3977608Z test_linalg.py::TestLinalgCPU::test_addmm_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:24:13.3977931Z test_linalg.py::TestLinalgCPU::test_addmm_cpu_complex128 PASSED [ 3%] 2023-03-31T06:24:13.3978237Z test_linalg.py::TestLinalgCPU::test_addmm_cpu_complex64 PASSED [ 3%] 2023-03-31T06:24:13.3978549Z test_linalg.py::TestLinalgCPU::test_addmm_cpu_float32 PASSED [ 4%] 2023-03-31T06:24:13.3978864Z test_linalg.py::TestLinalgCPU::test_addmm_cpu_float64 PASSED [ 4%] 2023-03-31T06:24:13.3979167Z test_linalg.py::TestLinalgCPU::test_addmm_sizes_cpu_float32 PASSED [ 4%] 2023-03-31T06:24:13.3979491Z test_linalg.py::TestLinalgCPU::test_addmm_sizes_cpu_float64 PASSED [ 4%] 2023-03-31T06:24:13.3979804Z test_linalg.py::TestLinalgCPU::test_addmv_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:24:13.3980117Z test_linalg.py::TestLinalgCPU::test_addmv_cpu_complex128 PASSED [ 4%] 2023-03-31T06:24:13.3980420Z test_linalg.py::TestLinalgCPU::test_addmv_cpu_complex64 PASSED [ 4%] 2023-03-31T06:24:13.3980727Z test_linalg.py::TestLinalgCPU::test_addmv_cpu_float32 PASSED [ 4%] 2023-03-31T06:24:13.3981038Z test_linalg.py::TestLinalgCPU::test_addmv_cpu_float64 PASSED [ 5%] 2023-03-31T06:24:13.3981407Z test_linalg.py::TestLinalgCPU::test_addmv_rowmajor_colmajor_incx_incy_lda_cpu_float32 PASSED [ 5%] 2023-03-31T06:24:13.3981789Z test_linalg.py::TestLinalgCPU::test_addmv_rowmajor_colmajor_incx_incy_lda_cpu_float64 PASSED [ 5%] 2023-03-31T06:24:13.3982138Z test_linalg.py::TestLinalgCPU::test_addr_bool_cpu_bool PASSED [ 5%] 2023-03-31T06:24:13.3982477Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:24:13.3982821Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_complex128 PASSED [ 5%] 2023-03-31T06:24:13.3983262Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_complex64 PASSED [ 5%] 2023-03-31T06:24:13.3983621Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float16 PASSED [ 6%] 2023-03-31T06:24:13.3984006Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float32 PASSED [ 6%] 2023-03-31T06:24:13.3984350Z test_linalg.py::TestLinalgCPU::test_addr_float_and_complex_cpu_float64 PASSED [ 6%] 2023-03-31T06:24:13.3984690Z test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int16 PASSED [ 6%] 2023-03-31T06:24:13.3985017Z test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int32 PASSED [ 6%] 2023-03-31T06:24:13.3985324Z test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int64 PASSED [ 6%] 2023-03-31T06:24:13.3985643Z test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_int8 PASSED [ 6%] 2023-03-31T06:24:13.3985961Z test_linalg.py::TestLinalgCPU::test_addr_integral_cpu_uint8 PASSED [ 6%] 2023-03-31T06:24:13.3986291Z test_linalg.py::TestLinalgCPU::test_addr_type_promotion_cpu PASSED [ 7%] 2023-03-31T06:24:13.3986598Z test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_bfloat16 PASSED [ 7%] 2023-03-31T06:24:13.3986913Z test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_complex128 PASSED [ 7%] 2023-03-31T06:24:13.3987234Z test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_complex64 PASSED [ 7%] 2023-03-31T06:24:13.3987540Z test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_float32 PASSED [ 7%] 2023-03-31T06:24:13.3987850Z test_linalg.py::TestLinalgCPU::test_baddbmm_cpu_float64 PASSED [ 7%] 2023-03-31T06:24:13.3988197Z test_linalg.py::TestLinalgCPU::test_baddbmm_nan_input_with_zero_beta_cpu_float32 PASSED [ 7%] 2023-03-31T06:24:13.3988605Z test_linalg.py::TestLinalgCPU::test_bfloat16_accumulation_with_ref_path_cpu PASSED [ 8%] 2023-03-31T06:24:13.3988981Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:24:13.3989332Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_complex128 PASSED [ 8%] 2023-03-31T06:24:13.3989687Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_complex64 PASSED [ 8%] 2023-03-31T06:24:13.3990025Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_float32 PASSED [ 8%] 2023-03-31T06:24:13.3990375Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_float64 PASSED [ 8%] 2023-03-31T06:24:13.3990724Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int16 PASSED [ 8%] 2023-03-31T06:24:13.3991067Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int32 PASSED [ 9%] 2023-03-31T06:24:13.3991392Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int64 PASSED [ 9%] 2023-03-31T06:24:13.3991733Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_int8 PASSED [ 9%] 2023-03-31T06:24:13.3992072Z test_linalg.py::TestLinalgCPU::test_blas_alpha_beta_empty_cpu_uint8 PASSED [ 9%] 2023-03-31T06:24:13.3992388Z test_linalg.py::TestLinalgCPU::test_blas_empty_cpu PASSED [ 9%] 2023-03-31T06:24:13.3992710Z test_linalg.py::TestLinalgCPU::test_blas_mv_large_input_cpu PASSED [ 9%] 2023-03-31T06:24:13.3993035Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:24:13.3993362Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_complex128 PASSED [ 9%] 2023-03-31T06:24:13.3993709Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_complex64 PASSED [ 10%] 2023-03-31T06:24:13.3994030Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float16 PASSED [ 10%] 2023-03-31T06:24:13.3994350Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float32 PASSED [ 10%] 2023-03-31T06:24:13.3994649Z test_linalg.py::TestLinalgCPU::test_blas_nan_out_cpu_float64 PASSED [ 10%] 2023-03-31T06:24:13.3994963Z test_linalg.py::TestLinalgCPU::test_bmm_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:24:13.3995279Z test_linalg.py::TestLinalgCPU::test_bmm_cpu_complex128 PASSED [ 10%] 2023-03-31T06:24:13.3995592Z test_linalg.py::TestLinalgCPU::test_bmm_cpu_complex64 PASSED [ 10%] 2023-03-31T06:24:13.3995888Z test_linalg.py::TestLinalgCPU::test_bmm_cpu_float32 PASSED [ 11%] 2023-03-31T06:24:13.3996250Z test_linalg.py::TestLinalgCPU::test_bmm_cpu_float64 PASSED [ 11%] 2023-03-31T06:24:13.3996577Z test_linalg.py::TestLinalgCPU::test_broadcast_batched_matmul_cpu PASSED [ 11%] 2023-03-31T06:24:13.3996892Z test_linalg.py::TestLinalgCPU::test_broadcast_fused_matmul_cpu PASSED [ 11%] 2023-03-31T06:24:13.3997217Z test_linalg.py::TestLinalgCPU::test_chain_matmul_cpu_float64 PASSED [ 11%] 2023-03-31T06:24:13.3997535Z test_linalg.py::TestLinalgCPU::test_cholesky_cpu_complex128 PASSED [ 11%] 2023-03-31T06:24:13.3997857Z test_linalg.py::TestLinalgCPU::test_cholesky_cpu_complex64 PASSED [ 11%] 2023-03-31T06:24:13.3998161Z test_linalg.py::TestLinalgCPU::test_cholesky_cpu_float32 PASSED [ 11%] 2023-03-31T06:24:13.3998470Z test_linalg.py::TestLinalgCPU::test_cholesky_cpu_float64 PASSED [ 12%] 2023-03-31T06:24:13.3998815Z test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_complex128 PASSED [ 12%] 2023-03-31T06:24:13.3999174Z test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_complex64 PASSED [ 12%] 2023-03-31T06:24:13.3999543Z test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_float32 PASSED [ 12%] 2023-03-31T06:24:13.3999907Z test_linalg.py::TestLinalgCPU::test_cholesky_errors_and_warnings_cpu_float64 PASSED [ 12%] 2023-03-31T06:24:13.4000247Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_complex128 PASSED [ 12%] 2023-03-31T06:24:13.4000561Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_complex64 PASSED [ 12%] 2023-03-31T06:24:13.4000914Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_float32 PASSED [ 13%] 2023-03-31T06:24:13.4001260Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_cpu_float64 PASSED [ 13%] 2023-03-31T06:24:13.4001580Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_complex128 PASSED [ 13%] 2023-03-31T06:24:13.4001924Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_complex64 PASSED [ 13%] 2023-03-31T06:24:13.4002264Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_float32 PASSED [ 13%] 2023-03-31T06:24:13.4002600Z test_linalg.py::TestLinalgCPU::test_cholesky_ex_non_pd_cpu_float64 PASSED [ 13%] 2023-03-31T06:24:13.4002931Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_complex128 PASSED [ 13%] 2023-03-31T06:24:13.4003498Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_complex64 PASSED [ 13%] 2023-03-31T06:24:13.4003838Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_float32 PASSED [ 14%] 2023-03-31T06:24:13.4004163Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_cpu_float64 PASSED [ 14%] 2023-03-31T06:24:13.4004530Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_complex128 PASSED [ 14%] 2023-03-31T06:24:13.4004922Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_complex64 PASSED [ 14%] 2023-03-31T06:24:13.4005308Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_float32 PASSED [ 14%] 2023-03-31T06:24:13.4005676Z test_linalg.py::TestLinalgCPU::test_cholesky_inverse_errors_and_warnings_cpu_float64 PASSED [ 14%] 2023-03-31T06:24:13.4006139Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_complex128 PASSED [ 14%] 2023-03-31T06:24:13.4006534Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_complex64 PASSED [ 15%] 2023-03-31T06:24:13.4006906Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_float32 PASSED [ 15%] 2023-03-31T06:24:13.4007288Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_broadcasting_cpu_float64 PASSED [ 15%] 2023-03-31T06:24:13.4007662Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_complex128 PASSED [ 15%] 2023-03-31T06:24:13.4008021Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_complex64 PASSED [ 15%] 2023-03-31T06:24:13.4008359Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_float32 PASSED [ 15%] 2023-03-31T06:24:13.4008708Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_cpu_float64 PASSED [ 15%] 2023-03-31T06:24:13.4009211Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 15%] 2023-03-31T06:24:13.4009726Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 16%] 2023-03-31T06:24:13.4010222Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 16%] 2023-03-31T06:24:13.4010718Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_batched_many_batches_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 16%] 2023-03-31T06:24:13.4011137Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_complex128 PASSED [ 16%] 2023-03-31T06:24:13.4011479Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_complex64 PASSED [ 16%] 2023-03-31T06:24:13.4011801Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_float32 PASSED [ 16%] 2023-03-31T06:24:13.4012128Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_cpu_float64 PASSED [ 16%] 2023-03-31T06:24:13.4012491Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_complex128 PASSED [ 17%] 2023-03-31T06:24:13.4012866Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_complex64 PASSED [ 17%] 2023-03-31T06:24:13.4013319Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_float32 PASSED [ 17%] 2023-03-31T06:24:13.4013741Z test_linalg.py::TestLinalgCPU::test_cholesky_solve_out_errors_and_warnings_cpu_float64 PASSED [ 17%] 2023-03-31T06:24:13.4014076Z test_linalg.py::TestLinalgCPU::test_cond_cpu_complex128 2023-03-31T06:24:13.4014354Z Intel MKL ERROR: Parameter 4 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4014523Z 2023-03-31T06:24:13.4014655Z Intel MKL ERROR: Parameter 5 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4014894Z PASSED [ 17%] 2023-03-31T06:24:13.4015120Z test_linalg.py::TestLinalgCPU::test_cond_cpu_complex64 2023-03-31T06:24:13.4015409Z Intel MKL ERROR: Parameter 4 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4015574Z 2023-03-31T06:24:13.4015706Z Intel MKL ERROR: Parameter 5 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4015937Z PASSED [ 17%] 2023-03-31T06:24:13.4016162Z test_linalg.py::TestLinalgCPU::test_cond_cpu_float32 2023-03-31T06:24:13.4016448Z Intel MKL ERROR: Parameter 4 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4016610Z 2023-03-31T06:24:13.4016739Z Intel MKL ERROR: Parameter 5 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4016961Z PASSED [ 17%] 2023-03-31T06:24:13.4017198Z test_linalg.py::TestLinalgCPU::test_cond_cpu_float64 2023-03-31T06:24:13.4017479Z Intel MKL ERROR: Parameter 4 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4017640Z 2023-03-31T06:24:13.4017769Z Intel MKL ERROR: Parameter 5 was incorrect on entry to DLASCL. 2023-03-31T06:24:13.4017987Z PASSED [ 18%] 2023-03-31T06:24:13.4018300Z test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_complex128 PASSED [ 18%] 2023-03-31T06:24:13.4018664Z test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_complex64 PASSED [ 18%] 2023-03-31T06:24:13.4019008Z test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_float32 PASSED [ 18%] 2023-03-31T06:24:13.4019363Z test_linalg.py::TestLinalgCPU::test_cond_errors_and_warnings_cpu_float64 PASSED [ 18%] 2023-03-31T06:24:13.4019725Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_bfloat16 PASSED [ 18%] 2023-03-31T06:24:13.4020101Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_complex128 PASSED [ 18%] 2023-03-31T06:24:13.4020462Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_complex64 PASSED [ 18%] 2023-03-31T06:24:13.4020827Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_float32 PASSED [ 19%] 2023-03-31T06:24:13.4021237Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_float64 PASSED [ 19%] 2023-03-31T06:24:13.4021590Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int16 PASSED [ 19%] 2023-03-31T06:24:13.4021950Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int32 PASSED [ 19%] 2023-03-31T06:24:13.4022307Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int64 PASSED [ 19%] 2023-03-31T06:24:13.4022663Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_int8 PASSED [ 19%] 2023-03-31T06:24:13.4023012Z test_linalg.py::TestLinalgCPU::test_corner_cases_of_cublasltmatmul_cpu_uint8 PASSED [ 19%] 2023-03-31T06:24:13.4023430Z test_linalg.py::TestLinalgCPU::test_cross_cpu_complex64 PASSED [ 20%] 2023-03-31T06:24:13.4023754Z test_linalg.py::TestLinalgCPU::test_cross_cpu_float32 PASSED [ 20%] 2023-03-31T06:24:13.4024090Z test_linalg.py::TestLinalgCPU::test_cross_with_and_without_dim_cpu_complex64 PASSED [ 20%] 2023-03-31T06:24:13.4024451Z test_linalg.py::TestLinalgCPU::test_cross_with_and_without_dim_cpu_float32 PASSED [ 20%] 2023-03-31T06:24:13.4024792Z test_linalg.py::TestLinalgCPU::test_det_cpu_complex128 PASSED [ 20%] 2023-03-31T06:24:13.4025112Z test_linalg.py::TestLinalgCPU::test_det_cpu_float64 PASSED [ 20%] 2023-03-31T06:24:13.4025438Z test_linalg.py::TestLinalgCPU::test_det_logdet_slogdet_batched_cpu_float64 PASSED [ 20%] 2023-03-31T06:24:13.4025825Z test_linalg.py::TestLinalgCPU::test_det_logdet_slogdet_cpu_float64 PASSED [ 20%] 2023-03-31T06:24:13.4026189Z test_linalg.py::TestLinalgCPU::test_dot_invalid_args_cpu PASSED [ 21%] 2023-03-31T06:24:13.4026503Z test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:24:13.4026839Z test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_complex64 PASSED [ 21%] 2023-03-31T06:24:13.4027162Z test_linalg.py::TestLinalgCPU::test_dot_vs_numpy_cpu_float32 PASSED [ 21%] 2023-03-31T06:24:13.4027568Z test_linalg.py::TestLinalgCPU::test_eig_check_magma_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 21%] 2023-03-31T06:24:13.4027988Z test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_complex128 SKIPPED (Only runs on cuda) [ 21%] 2023-03-31T06:24:13.4028385Z test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_complex64 SKIPPED (Only runs on cuda) [ 21%] 2023-03-31T06:24:13.4028777Z test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_float32 SKIPPED (Only runs on cuda) [ 22%] 2023-03-31T06:24:13.4029165Z test_linalg.py::TestLinalgCPU::test_eig_compare_backends_cpu_float64 SKIPPED (Only runs on cuda) [ 22%] 2023-03-31T06:24:13.4029524Z test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_complex128 PASSED [ 22%] 2023-03-31T06:24:13.4029879Z test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_complex64 PASSED [ 22%] 2023-03-31T06:24:13.4030232Z test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_float32 PASSED [ 22%] 2023-03-31T06:24:13.4030605Z test_linalg.py::TestLinalgCPU::test_eig_errors_and_warnings_cpu_float64 PASSED [ 22%] 2023-03-31T06:24:13.4030942Z test_linalg.py::TestLinalgCPU::test_eig_numpy_cpu_complex128 PASSED [ 22%] 2023-03-31T06:24:13.4031269Z test_linalg.py::TestLinalgCPU::test_eig_numpy_cpu_float64 PASSED [ 22%] 2023-03-31T06:24:13.4031592Z test_linalg.py::TestLinalgCPU::test_eig_removed_error_cpu PASSED [ 23%] 2023-03-31T06:24:13.4031903Z test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_complex128 PASSED [ 23%] 2023-03-31T06:24:13.4032223Z test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_complex64 PASSED [ 23%] 2023-03-31T06:24:13.4032547Z test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_float32 PASSED [ 23%] 2023-03-31T06:24:13.4032852Z test_linalg.py::TestLinalgCPU::test_eig_with_nan_cpu_float64 PASSED [ 23%] 2023-03-31T06:24:13.4033205Z test_linalg.py::TestLinalgCPU::test_eigh_cpu_complex128 PASSED [ 23%] 2023-03-31T06:24:13.4033517Z test_linalg.py::TestLinalgCPU::test_eigh_cpu_complex64 PASSED [ 23%] 2023-03-31T06:24:13.4033827Z test_linalg.py::TestLinalgCPU::test_eigh_cpu_float32 PASSED [ 24%] 2023-03-31T06:24:13.4034123Z test_linalg.py::TestLinalgCPU::test_eigh_cpu_float64 PASSED [ 24%] 2023-03-31T06:24:13.4034457Z test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_complex128 PASSED [ 24%] 2023-03-31T06:24:13.4034816Z test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_complex64 PASSED [ 24%] 2023-03-31T06:24:13.4035166Z test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_float32 PASSED [ 24%] 2023-03-31T06:24:13.4035515Z test_linalg.py::TestLinalgCPU::test_eigh_errors_and_warnings_cpu_float64 PASSED [ 24%] 2023-03-31T06:24:13.4035862Z test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_complex128 PASSED [ 24%] 2023-03-31T06:24:13.4036212Z test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_complex64 PASSED [ 25%] 2023-03-31T06:24:13.4036530Z test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_float32 PASSED [ 25%] 2023-03-31T06:24:13.4036853Z test_linalg.py::TestLinalgCPU::test_eigh_lower_uplo_cpu_float64 PASSED [ 25%] 2023-03-31T06:24:13.4037224Z test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_complex128 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T06:24:13.4037619Z test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_complex64 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T06:24:13.4038059Z test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_float32 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T06:24:13.4038495Z test_linalg.py::TestLinalgCPU::test_eigvals_compare_backends_cpu_float64 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T06:24:13.4038868Z test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_complex128 PASSED [ 25%] 2023-03-31T06:24:13.4039221Z test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_complex64 PASSED [ 26%] 2023-03-31T06:24:13.4039588Z test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_float32 PASSED [ 26%] 2023-03-31T06:24:13.4039950Z test_linalg.py::TestLinalgCPU::test_eigvals_errors_and_warnings_cpu_float64 PASSED [ 26%] 2023-03-31T06:24:13.4040297Z test_linalg.py::TestLinalgCPU::test_eigvals_numpy_cpu_complex128 PASSED [ 26%] 2023-03-31T06:24:13.4040614Z test_linalg.py::TestLinalgCPU::test_eigvals_numpy_cpu_float64 PASSED [ 26%] 2023-03-31T06:24:13.4040942Z test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_complex128 PASSED [ 26%] 2023-03-31T06:24:13.4041272Z test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_complex64 PASSED [ 26%] 2023-03-31T06:24:13.4041579Z test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_float32 PASSED [ 27%] 2023-03-31T06:24:13.4041899Z test_linalg.py::TestLinalgCPU::test_eigvalsh_cpu_float64 PASSED [ 27%] 2023-03-31T06:24:13.4042245Z test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_complex128 PASSED [ 27%] 2023-03-31T06:24:13.4042621Z test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_complex64 PASSED [ 27%] 2023-03-31T06:24:13.4043164Z test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_float32 PASSED [ 27%] 2023-03-31T06:24:13.4043598Z test_linalg.py::TestLinalgCPU::test_eigvalsh_errors_and_warnings_cpu_float64 PASSED [ 27%] 2023-03-31T06:24:13.4043945Z test_linalg.py::TestLinalgCPU::test_einsum_corner_cases_cpu PASSED [ 27%] 2023-03-31T06:24:13.4044262Z test_linalg.py::TestLinalgCPU::test_einsum_cpu_complex128 PASSED [ 27%] 2023-03-31T06:24:13.4044585Z test_linalg.py::TestLinalgCPU::test_einsum_cpu_float64 PASSED [ 28%] 2023-03-31T06:24:13.4044905Z test_linalg.py::TestLinalgCPU::test_einsum_error_cases_cpu PASSED [ 28%] 2023-03-31T06:24:13.4045233Z test_linalg.py::TestLinalgCPU::test_einsum_random_cpu_complex128 PASSED [ 28%] 2023-03-31T06:24:13.4045550Z test_linalg.py::TestLinalgCPU::test_einsum_random_cpu_float64 PASSED [ 28%] 2023-03-31T06:24:13.4045960Z test_linalg.py::TestLinalgCPU::test_einsum_sublist_format_cpu_complex128 PASSED [ 28%] 2023-03-31T06:24:13.4046316Z test_linalg.py::TestLinalgCPU::test_einsum_sublist_format_cpu_float64 PASSED [ 28%] 2023-03-31T06:24:13.4046642Z test_linalg.py::TestLinalgCPU::test_geqrf_cpu_complex128 PASSED [ 28%] 2023-03-31T06:24:13.4046964Z test_linalg.py::TestLinalgCPU::test_geqrf_cpu_complex64 PASSED [ 29%] 2023-03-31T06:24:13.4047279Z test_linalg.py::TestLinalgCPU::test_geqrf_cpu_float32 PASSED [ 29%] 2023-03-31T06:24:13.4047601Z test_linalg.py::TestLinalgCPU::test_geqrf_cpu_float64 PASSED [ 29%] 2023-03-31T06:24:13.4047929Z test_linalg.py::TestLinalgCPU::test_householder_product_cpu_complex128 PASSED [ 29%] 2023-03-31T06:24:13.4048291Z test_linalg.py::TestLinalgCPU::test_householder_product_cpu_complex64 PASSED [ 29%] 2023-03-31T06:24:13.4048641Z test_linalg.py::TestLinalgCPU::test_householder_product_cpu_float32 PASSED [ 29%] 2023-03-31T06:24:13.4048976Z test_linalg.py::TestLinalgCPU::test_householder_product_cpu_float64 PASSED [ 29%] 2023-03-31T06:24:13.4049343Z test_linalg.py::TestLinalgCPU::test_householder_product_errors_and_warnings_cpu PASSED [ 29%] 2023-03-31T06:24:13.4049690Z test_linalg.py::TestLinalgCPU::test_inner_cpu_complex64 PASSED [ 30%] 2023-03-31T06:24:13.4050007Z test_linalg.py::TestLinalgCPU::test_inner_cpu_float32 PASSED [ 30%] 2023-03-31T06:24:13.4050331Z test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_complex128 PASSED [ 30%] 2023-03-31T06:24:13.4050730Z test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_complex64 PASSED [ 30%] 2023-03-31T06:24:13.4051120Z test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_float32 PASSED [ 30%] 2023-03-31T06:24:13.4051457Z test_linalg.py::TestLinalgCPU::test_inv_errors_and_warnings_cpu_float64 PASSED [ 30%] 2023-03-31T06:24:13.4051800Z test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_complex128 PASSED [ 30%] 2023-03-31T06:24:13.4052143Z test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_complex64 PASSED [ 31%] 2023-03-31T06:24:13.4052486Z test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_float32 PASSED [ 31%] 2023-03-31T06:24:13.4052807Z test_linalg.py::TestLinalgCPU::test_inv_ex_info_device_cpu_float64 PASSED [ 31%] 2023-03-31T06:24:13.4053136Z test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_complex128 PASSED [ 31%] 2023-03-31T06:24:13.4053472Z test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_complex64 PASSED [ 31%] 2023-03-31T06:24:13.4053790Z test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_float32 PASSED [ 31%] 2023-03-31T06:24:13.4054117Z test_linalg.py::TestLinalgCPU::test_inv_ex_singular_cpu_float64 PASSED [ 31%] 2023-03-31T06:24:13.4054485Z test_linalg.py::TestLinalgCPU::test_invariance_error_spectral_decompositions_cpu_complex128 PASSED [ 31%] 2023-03-31T06:24:13.4054850Z test_linalg.py::TestLinalgCPU::test_inverse_cpu_complex128 PASSED [ 32%] 2023-03-31T06:24:13.4055165Z test_linalg.py::TestLinalgCPU::test_inverse_cpu_complex64 PASSED [ 32%] 2023-03-31T06:24:13.4055526Z test_linalg.py::TestLinalgCPU::test_inverse_cpu_float32 PASSED [ 32%] 2023-03-31T06:24:13.4055844Z test_linalg.py::TestLinalgCPU::test_inverse_cpu_float64 PASSED [ 32%] 2023-03-31T06:24:13.4056153Z test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_complex128 PASSED [ 32%] 2023-03-31T06:24:13.4056482Z test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_complex64 PASSED [ 32%] 2023-03-31T06:24:13.4056809Z test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_float32 PASSED [ 32%] 2023-03-31T06:24:13.4057133Z test_linalg.py::TestLinalgCPU::test_inverse_errors_cpu_float64 PASSED [ 33%] 2023-03-31T06:24:13.4057457Z test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_complex128 PASSED [ 33%] 2023-03-31T06:24:13.4057805Z test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_complex64 PASSED [ 33%] 2023-03-31T06:24:13.4058193Z test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_float32 PASSED [ 33%] 2023-03-31T06:24:13.4058523Z test_linalg.py::TestLinalgCPU::test_inverse_errors_large_cpu_float64 PASSED [ 33%] 2023-03-31T06:24:13.4058941Z test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 33%] 2023-03-31T06:24:13.4059421Z test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 33%] 2023-03-31T06:24:13.4059900Z test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 34%] 2023-03-31T06:24:13.4060362Z test_linalg.py::TestLinalgCPU::test_inverse_many_batches_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 34%] 2023-03-31T06:24:13.4060753Z test_linalg.py::TestLinalgCPU::test_kron_cpu_complex128 PASSED [ 34%] 2023-03-31T06:24:13.4061082Z test_linalg.py::TestLinalgCPU::test_kron_cpu_complex64 PASSED [ 34%] 2023-03-31T06:24:13.4061404Z test_linalg.py::TestLinalgCPU::test_kron_cpu_float32 PASSED [ 34%] 2023-03-31T06:24:13.4061702Z test_linalg.py::TestLinalgCPU::test_kron_cpu_float64 PASSED [ 34%] 2023-03-31T06:24:13.4062018Z test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_complex128 PASSED [ 34%] 2023-03-31T06:24:13.4062347Z test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_complex64 PASSED [ 34%] 2023-03-31T06:24:13.4062699Z test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_float32 PASSED [ 35%] 2023-03-31T06:24:13.4063029Z test_linalg.py::TestLinalgCPU::test_kron_empty_cpu_float64 PASSED [ 35%] 2023-03-31T06:24:13.4063478Z test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_complex128 PASSED [ 35%] 2023-03-31T06:24:13.4063838Z test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_complex64 PASSED [ 35%] 2023-03-31T06:24:13.4064185Z test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_float32 PASSED [ 35%] 2023-03-31T06:24:13.4064543Z test_linalg.py::TestLinalgCPU::test_kron_errors_and_warnings_cpu_float64 PASSED [ 35%] 2023-03-31T06:24:13.4064875Z test_linalg.py::TestLinalgCPU::test_lapack_empty_cpu PASSED [ 35%] 2023-03-31T06:24:13.4065236Z test_linalg.py::TestLinalgCPU::test_large_bmm_backward_cpu SKIPPED (Insufficient cuda memory) [ 36%] 2023-03-31T06:24:13.4065619Z test_linalg.py::TestLinalgCPU::test_large_bmm_mm_backward_cpu SKIPPED (Insufficient cuda memory) [ 36%] 2023-03-31T06:24:13.4065988Z test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_complex128 PASSED [ 36%] 2023-03-31T06:24:13.4066326Z test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_complex64 PASSED [ 36%] 2023-03-31T06:24:13.4066637Z test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_float32 PASSED [ 36%] 2023-03-31T06:24:13.4066954Z test_linalg.py::TestLinalgCPU::test_ldl_factor_cpu_float64 PASSED [ 36%] 2023-03-31T06:24:13.4067278Z test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_complex128 PASSED [ 36%] 2023-03-31T06:24:13.4067637Z test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_complex64 PASSED [ 36%] 2023-03-31T06:24:13.4067943Z test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_float32 PASSED [ 37%] 2023-03-31T06:24:13.4068256Z test_linalg.py::TestLinalgCPU::test_ldl_solve_cpu_float64 PASSED [ 37%] 2023-03-31T06:24:13.4068583Z test_linalg.py::TestLinalgCPU::test_linalg_cross_cpu_complex64 PASSED [ 37%] 2023-03-31T06:24:13.4068900Z test_linalg.py::TestLinalgCPU::test_linalg_cross_cpu_float32 PASSED [ 37%] 2023-03-31T06:24:13.4069247Z test_linalg.py::TestLinalgCPU::test_linalg_cross_with_and_without_dim_cpu_complex64 PASSED [ 37%] 2023-03-31T06:24:13.4069619Z test_linalg.py::TestLinalgCPU::test_linalg_cross_with_and_without_dim_cpu_float32 PASSED [ 37%] 2023-03-31T06:24:13.4069993Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_complex128 PASSED [ 37%] 2023-03-31T06:24:13.4070392Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_complex64 PASSED [ 38%] 2023-03-31T06:24:13.4070764Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_float32 PASSED [ 38%] 2023-03-31T06:24:13.4071133Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_batch_broadcasting_cpu_float64 PASSED [ 38%] 2023-03-31T06:24:13.4071467Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_complex128 PASSED [ 38%] 2023-03-31T06:24:13.4071797Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_complex64 PASSED [ 38%] 2023-03-31T06:24:13.4072121Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_float32 PASSED [ 38%] 2023-03-31T06:24:13.4072445Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_cpu_float64 PASSED [ 38%] 2023-03-31T06:24:13.4072772Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_complex128 PASSED [ 38%] 2023-03-31T06:24:13.4073132Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_complex64 PASSED [ 39%] 2023-03-31T06:24:13.4073489Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_float32 PASSED [ 39%] 2023-03-31T06:24:13.4073830Z test_linalg.py::TestLinalgCPU::test_linalg_lstsq_input_checks_cpu_float64 PASSED [ 39%] 2023-03-31T06:24:13.4074179Z test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_complex128 PASSED [ 39%] 2023-03-31T06:24:13.4074527Z test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_complex64 PASSED [ 39%] 2023-03-31T06:24:13.4074900Z test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_float32 PASSED [ 39%] 2023-03-31T06:24:13.4075259Z test_linalg.py::TestLinalgCPU::test_linalg_lu_cpu_errors_cpu_float64 PASSED [ 39%] 2023-03-31T06:24:13.4075601Z test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_complex128 PASSED [ 40%] 2023-03-31T06:24:13.4075940Z test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_complex64 PASSED [ 40%] 2023-03-31T06:24:13.4076275Z test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_float32 PASSED [ 40%] 2023-03-31T06:24:13.4076595Z test_linalg.py::TestLinalgCPU::test_linalg_lu_family_cpu_float64 PASSED [ 40%] 2023-03-31T06:24:13.4076926Z test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_complex128 PASSED [ 40%] 2023-03-31T06:24:13.4077257Z test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_complex64 PASSED [ 40%] 2023-03-31T06:24:13.4077573Z test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_float32 PASSED [ 40%] 2023-03-31T06:24:13.4077898Z test_linalg.py::TestLinalgCPU::test_linalg_lu_solve_cpu_float64 PASSED [ 40%] 2023-03-31T06:24:13.4078319Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 41%] 2023-03-31T06:24:13.4078821Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 41%] 2023-03-31T06:24:13.4079295Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 41%] 2023-03-31T06:24:13.4079811Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_analytic_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 41%] 2023-03-31T06:24:13.4080229Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_batch_cpu_float32 PASSED [ 41%] 2023-03-31T06:24:13.4080579Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_batch_cpu_float64 PASSED [ 41%] 2023-03-31T06:24:13.4080926Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_complex128 PASSED [ 41%] 2023-03-31T06:24:13.4081307Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_complex64 PASSED [ 42%] 2023-03-31T06:24:13.4081675Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_float32 PASSED [ 42%] 2023-03-31T06:24:13.4082031Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_boundary_cases_cpu_float64 PASSED [ 42%] 2023-03-31T06:24:13.4082456Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_complex128 PASSED [ 42%] 2023-03-31T06:24:13.4082845Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_complex64 PASSED [ 42%] 2023-03-31T06:24:13.4083361Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_float32 PASSED [ 42%] 2023-03-31T06:24:13.4083728Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_compare_with_taylor_cpu_float64 PASSED [ 42%] 2023-03-31T06:24:13.4084110Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_no_warnings_cpu_complex64 PASSED [ 43%] 2023-03-31T06:24:13.4084473Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_utils_cpu_complex64 PASSED [ 43%] 2023-03-31T06:24:13.4084827Z test_linalg.py::TestLinalgCPU::test_linalg_matrix_exp_utils_cpu_float32 PASSED [ 43%] 2023-03-31T06:24:13.4085167Z test_linalg.py::TestLinalgCPU::test_linalg_qr_autograd_errors_cpu_float32 PASSED [ 43%] 2023-03-31T06:24:13.4085544Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_complex128 PASSED [ 43%] 2023-03-31T06:24:13.4085939Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_complex64 PASSED [ 43%] 2023-03-31T06:24:13.4086314Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_float32 PASSED [ 43%] 2023-03-31T06:24:13.4086699Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_broadcasting_cpu_float64 PASSED [ 43%] 2023-03-31T06:24:13.4087067Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_complex128 PASSED [ 44%] 2023-03-31T06:24:13.4087480Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_complex64 PASSED [ 44%] 2023-03-31T06:24:13.4087859Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_float32 PASSED [ 44%] 2023-03-31T06:24:13.4088209Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_cpu_float64 PASSED [ 44%] 2023-03-31T06:24:13.4088595Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_complex128 SKIPPED (Only runs on cuda) [ 44%] 2023-03-31T06:24:13.4089012Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_complex64 SKIPPED (Only runs on cuda) [ 44%] 2023-03-31T06:24:13.4089414Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_float32 SKIPPED (Only runs on cuda) [ 44%] 2023-03-31T06:24:13.4089819Z test_linalg.py::TestLinalgCPU::test_linalg_solve_triangular_large_cpu_float64 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T06:24:13.4090197Z test_linalg.py::TestLinalgCPU::test_linear_algebra_scalar_raises_cpu PASSED [ 45%] 2023-03-31T06:24:13.4090525Z test_linalg.py::TestLinalgCPU::test_lobpcg_basic_cpu_float64 PASSED [ 45%] 2023-03-31T06:24:13.4090854Z test_linalg.py::TestLinalgCPU::test_lobpcg_ortho_cpu_float64 PASSED [ 45%] 2023-03-31T06:24:13.4091179Z test_linalg.py::TestLinalgCPU::test_lobpcg_scipy_cpu_float64 PASSED [ 45%] 2023-03-31T06:24:13.4091514Z test_linalg.py::TestLinalgCPU::test_lobpcg_torchscript_cpu_float64 PASSED [ 45%] 2023-03-31T06:24:13.4091839Z test_linalg.py::TestLinalgCPU::test_lstsq_removed_error_cpu PASSED [ 45%] 2023-03-31T06:24:13.4092232Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_complex128 PASSED [ 45%] 2023-03-31T06:24:13.4092607Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_complex64 PASSED [ 46%] 2023-03-31T06:24:13.4092961Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_float32 PASSED [ 46%] 2023-03-31T06:24:13.4093330Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_broadcasting_cpu_float64 PASSED [ 46%] 2023-03-31T06:24:13.4093685Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_complex128 PASSED [ 46%] 2023-03-31T06:24:13.4094029Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_complex64 PASSED [ 46%] 2023-03-31T06:24:13.4094351Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_float32 PASSED [ 46%] 2023-03-31T06:24:13.4094682Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_cpu_float64 PASSED [ 46%] 2023-03-31T06:24:13.4095148Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 47%] 2023-03-31T06:24:13.4095643Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 47%] 2023-03-31T06:24:13.4096124Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 47%] 2023-03-31T06:24:13.4096617Z test_linalg.py::TestLinalgCPU::test_lu_solve_batched_many_batches_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 47%] 2023-03-31T06:24:13.4097019Z test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_complex128 PASSED [ 47%] 2023-03-31T06:24:13.4097349Z test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_complex64 PASSED [ 47%] 2023-03-31T06:24:13.4097658Z test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_float32 PASSED [ 47%] 2023-03-31T06:24:13.4097978Z test_linalg.py::TestLinalgCPU::test_lu_solve_cpu_float64 PASSED [ 47%] 2023-03-31T06:24:13.4098347Z test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_complex128 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T06:24:13.4098732Z test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_complex64 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T06:24:13.4099128Z test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_float32 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T06:24:13.4099551Z test_linalg.py::TestLinalgCPU::test_lu_solve_large_matrices_cpu_float64 SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T06:24:13.4099944Z test_linalg.py::TestLinalgCPU::test_lu_unpack_check_input_cpu_float64 PASSED [ 48%] 2023-03-31T06:24:13.4100279Z test_linalg.py::TestLinalgCPU::test_matmul_45724_cpu SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T06:24:13.4100640Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_complex64 PASSED [ 48%] 2023-03-31T06:24:13.4101011Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_float32 PASSED [ 49%] 2023-03-31T06:24:13.4101357Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_1d_Nd_cpu_int64 PASSED [ 49%] 2023-03-31T06:24:13.4101717Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_complex64 PASSED [ 49%] 2023-03-31T06:24:13.4102079Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_float32 PASSED [ 49%] 2023-03-31T06:24:13.4102437Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_2d_Nd_cpu_int64 PASSED [ 49%] 2023-03-31T06:24:13.4102784Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_complex64 PASSED [ 49%] 2023-03-31T06:24:13.4103227Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_float32 PASSED [ 49%] 2023-03-31T06:24:13.4103590Z test_linalg.py::TestLinalgCPU::test_matmul_small_brute_force_3d_Nd_cpu_int64 PASSED [ 50%] 2023-03-31T06:24:13.4103933Z test_linalg.py::TestLinalgCPU::test_matrix_norm_cpu_float32 PASSED [ 50%] 2023-03-31T06:24:13.4104293Z test_linalg.py::TestLinalgCPU::test_matrix_norm_cpu_float64 PASSED [ 50%] 2023-03-31T06:24:13.4104639Z test_linalg.py::TestLinalgCPU::test_matrix_power_negative_cpu_complex128 PASSED [ 50%] 2023-03-31T06:24:13.4104990Z test_linalg.py::TestLinalgCPU::test_matrix_power_negative_cpu_float64 PASSED [ 50%] 2023-03-31T06:24:13.4105331Z test_linalg.py::TestLinalgCPU::test_matrix_power_non_negative_cpu_complex128 PASSED [ 50%] 2023-03-31T06:24:13.4105688Z test_linalg.py::TestLinalgCPU::test_matrix_power_non_negative_cpu_float64 PASSED [ 50%] 2023-03-31T06:24:13.4106032Z test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_complex128 PASSED [ 50%] 2023-03-31T06:24:13.4106370Z test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_complex64 PASSED [ 51%] 2023-03-31T06:24:13.4106691Z test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_float32 PASSED [ 51%] 2023-03-31T06:24:13.4107086Z test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_cpu_float64 PASSED [ 51%] 2023-03-31T06:24:13.4107427Z test_linalg.py::TestLinalgCPU::test_matrix_rank_atol_rtol_cpu_float64 PASSED [ 51%] 2023-03-31T06:24:13.4107758Z test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_complex128 PASSED [ 51%] 2023-03-31T06:24:13.4108101Z test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_complex64 PASSED [ 51%] 2023-03-31T06:24:13.4108444Z test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_float32 PASSED [ 51%] 2023-03-31T06:24:13.4108776Z test_linalg.py::TestLinalgCPU::test_matrix_rank_basic_cpu_float64 PASSED [ 52%] 2023-03-31T06:24:13.4109092Z test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_complex128 PASSED [ 52%] 2023-03-31T06:24:13.4109418Z test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_complex64 PASSED [ 52%] 2023-03-31T06:24:13.4109747Z test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_float32 PASSED [ 52%] 2023-03-31T06:24:13.4110057Z test_linalg.py::TestLinalgCPU::test_matrix_rank_cpu_float64 PASSED [ 52%] 2023-03-31T06:24:13.4110388Z test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_complex128 PASSED [ 52%] 2023-03-31T06:24:13.4110730Z test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_complex64 PASSED [ 52%] 2023-03-31T06:24:13.4111060Z test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_float32 PASSED [ 52%] 2023-03-31T06:24:13.4111375Z test_linalg.py::TestLinalgCPU::test_matrix_rank_empty_cpu_float64 PASSED [ 53%] 2023-03-31T06:24:13.4111772Z test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_complex128 PASSED [ 53%] 2023-03-31T06:24:13.4112185Z test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_complex64 PASSED [ 53%] 2023-03-31T06:24:13.4112545Z test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_float32 PASSED [ 53%] 2023-03-31T06:24:13.4112914Z test_linalg.py::TestLinalgCPU::test_matrix_rank_out_errors_and_warnings_cpu_float64 PASSED [ 53%] 2023-03-31T06:24:13.4113268Z test_linalg.py::TestLinalgCPU::test_matrix_rank_removed_error_cpu PASSED [ 53%] 2023-03-31T06:24:13.4113599Z test_linalg.py::TestLinalgCPU::test_mm_bmm_non_memory_dense_cpu PASSED [ 53%] 2023-03-31T06:24:13.4113976Z test_linalg.py::TestLinalgCPU::test_mm_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4114416Z test_linalg.py::TestLinalgCPU::test_mm_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4114849Z test_linalg.py::TestLinalgCPU::test_mm_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4115273Z test_linalg.py::TestLinalgCPU::test_mm_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4115690Z test_linalg.py::TestLinalgCPU::test_mm_cpu_int32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4116110Z test_linalg.py::TestLinalgCPU::test_mm_cpu_int64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 54%] 2023-03-31T06:24:13.4116523Z test_linalg.py::TestLinalgCPU::test_multi_dot_cpu_complex128 PASSED [ 54%] 2023-03-31T06:24:13.4116852Z test_linalg.py::TestLinalgCPU::test_multi_dot_cpu_float64 PASSED [ 54%] 2023-03-31T06:24:13.4117163Z test_linalg.py::TestLinalgCPU::test_multi_dot_errors_cpu_float32 PASSED [ 55%] 2023-03-31T06:24:13.4117500Z test_linalg.py::TestLinalgCPU::test_norm_bfloat16_and_half_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:24:13.4117849Z test_linalg.py::TestLinalgCPU::test_norm_bfloat16_and_half_cpu_float16 PASSED [ 55%] 2023-03-31T06:24:13.4118176Z test_linalg.py::TestLinalgCPU::test_norm_complex_cpu_complex128 PASSED [ 55%] 2023-03-31T06:24:13.4118507Z test_linalg.py::TestLinalgCPU::test_norm_complex_cpu_complex64 PASSED [ 55%] 2023-03-31T06:24:13.4118833Z test_linalg.py::TestLinalgCPU::test_norm_complex_old_cpu PASSED [ 55%] 2023-03-31T06:24:13.4119185Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:24:13.4119498Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_complex128 PASSED [ 56%] 2023-03-31T06:24:13.4119825Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_complex64 PASSED [ 56%] 2023-03-31T06:24:13.4120150Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float16 PASSED [ 56%] 2023-03-31T06:24:13.4120454Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float32 PASSED [ 56%] 2023-03-31T06:24:13.4120776Z test_linalg.py::TestLinalgCPU::test_norm_dtype_cpu_float64 PASSED [ 56%] 2023-03-31T06:24:13.4121093Z test_linalg.py::TestLinalgCPU::test_norm_errors_cpu_float32 PASSED [ 56%] 2023-03-31T06:24:13.4121412Z test_linalg.py::TestLinalgCPU::test_norm_errors_cpu_float64 PASSED [ 56%] 2023-03-31T06:24:13.4121716Z test_linalg.py::TestLinalgCPU::test_norm_extreme_values_cpu PASSED [ 56%] 2023-03-31T06:24:13.4122030Z test_linalg.py::TestLinalgCPU::test_norm_fastpaths_cpu PASSED [ 57%] 2023-03-31T06:24:13.4122370Z test_linalg.py::TestLinalgCPU::test_norm_fro_2_equivalence_old_cpu_float32 PASSED [ 57%] 2023-03-31T06:24:13.4122739Z test_linalg.py::TestLinalgCPU::test_norm_fused_type_promotion_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 57%] 2023-03-31T06:24:13.4123311Z test_linalg.py::TestLinalgCPU::test_norm_fused_type_promotion_cpu_float16 SKIPPED (Only runs on cuda) [ 57%] 2023-03-31T06:24:13.4123677Z test_linalg.py::TestLinalgCPU::test_norm_matrix_cpu_float32 PASSED [ 57%] 2023-03-31T06:24:13.4124067Z test_linalg.py::TestLinalgCPU::test_norm_matrix_cpu_float64 PASSED [ 57%] 2023-03-31T06:24:13.4124463Z test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_complex128 PASSED [ 57%] 2023-03-31T06:24:13.4124847Z test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_complex64 PASSED [ 58%] 2023-03-31T06:24:13.4125215Z test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_float32 PASSED [ 58%] 2023-03-31T06:24:13.4125570Z test_linalg.py::TestLinalgCPU::test_norm_matrix_degenerate_shapes_cpu_float64 PASSED [ 58%] 2023-03-31T06:24:13.4125910Z test_linalg.py::TestLinalgCPU::test_norm_old_cpu PASSED [ 58%] 2023-03-31T06:24:13.4126235Z test_linalg.py::TestLinalgCPU::test_norm_old_nan_propagation_cpu PASSED [ 58%] 2023-03-31T06:24:13.4126567Z test_linalg.py::TestLinalgCPU::test_norm_vector_cpu_float32 PASSED [ 58%] 2023-03-31T06:24:13.4126874Z test_linalg.py::TestLinalgCPU::test_norm_vector_cpu_float64 PASSED [ 58%] 2023-03-31T06:24:13.4127228Z test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_complex128 PASSED [ 59%] 2023-03-31T06:24:13.4127601Z test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_complex64 PASSED [ 59%] 2023-03-31T06:24:13.4127958Z test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_float32 PASSED [ 59%] 2023-03-31T06:24:13.4128320Z test_linalg.py::TestLinalgCPU::test_norm_vector_degenerate_shapes_cpu_float64 PASSED [ 59%] 2023-03-31T06:24:13.4128683Z test_linalg.py::TestLinalgCPU::test_nuclear_norm_axes_small_brute_force_old_cpu PASSED [ 59%] 2023-03-31T06:24:13.4129088Z test_linalg.py::TestLinalgCPU::test_nuclear_norm_exceptions_old_cpu PASSED [ 59%] 2023-03-31T06:24:13.4129413Z test_linalg.py::TestLinalgCPU::test_nuclear_norm_out_cpu_float32 PASSED [ 59%] 2023-03-31T06:24:13.4129739Z test_linalg.py::TestLinalgCPU::test_nuclear_norm_out_cpu_float64 PASSED [ 59%] 2023-03-31T06:24:13.4130086Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_complex128 PASSED [ 60%] 2023-03-31T06:24:13.4130440Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_complex64 PASSED [ 60%] 2023-03-31T06:24:13.4130774Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_float32 PASSED [ 60%] 2023-03-31T06:24:13.4131115Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_cpu_float64 PASSED [ 60%] 2023-03-31T06:24:13.4131542Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_many_batches_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 60%] 2023-03-31T06:24:13.4132026Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_complex128 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T06:24:13.4132437Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_complex64 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T06:24:13.4132836Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_float32 SKIPPED (Only runs on cuda) [ 61%] 2023-03-31T06:24:13.4133234Z test_linalg.py::TestLinalgCPU::test_old_cholesky_batched_upper_cpu_float64 SKIPPED (Only runs on cuda) [ 61%] 2023-03-31T06:24:13.4133584Z test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_complex128 PASSED [ 61%] 2023-03-31T06:24:13.4133919Z test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_complex64 PASSED [ 61%] 2023-03-31T06:24:13.4134248Z test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_float32 PASSED [ 61%] 2023-03-31T06:24:13.4134573Z test_linalg.py::TestLinalgCPU::test_old_cholesky_cpu_float64 PASSED [ 61%] 2023-03-31T06:24:13.4134898Z test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_complex128 PASSED [ 61%] 2023-03-31T06:24:13.4135244Z test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_complex64 PASSED [ 61%] 2023-03-31T06:24:13.4135588Z test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_float32 PASSED [ 62%] 2023-03-31T06:24:13.4135909Z test_linalg.py::TestLinalgCPU::test_old_cholesky_empty_cpu_float64 PASSED [ 62%] 2023-03-31T06:24:13.4136278Z test_linalg.py::TestLinalgCPU::test_ormqr_cpu_complex128 PASSED [ 62%] 2023-03-31T06:24:13.4136626Z test_linalg.py::TestLinalgCPU::test_ormqr_cpu_complex64 PASSED [ 62%] 2023-03-31T06:24:13.4136927Z test_linalg.py::TestLinalgCPU::test_ormqr_cpu_float32 PASSED [ 62%] 2023-03-31T06:24:13.4137238Z test_linalg.py::TestLinalgCPU::test_ormqr_cpu_float64 PASSED [ 62%] 2023-03-31T06:24:13.4137575Z test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_complex128 PASSED [ 62%] 2023-03-31T06:24:13.4137936Z test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_complex64 PASSED [ 63%] 2023-03-31T06:24:13.4138284Z test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_float32 PASSED [ 63%] 2023-03-31T06:24:13.4138639Z test_linalg.py::TestLinalgCPU::test_ormqr_errors_and_warnings_cpu_float64 PASSED [ 63%] 2023-03-31T06:24:13.4138979Z test_linalg.py::TestLinalgCPU::test_outer_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:24:13.4139298Z test_linalg.py::TestLinalgCPU::test_outer_cpu_bool PASSED [ 63%] 2023-03-31T06:24:13.4139604Z test_linalg.py::TestLinalgCPU::test_outer_cpu_complex128 PASSED [ 63%] 2023-03-31T06:24:13.4139923Z test_linalg.py::TestLinalgCPU::test_outer_cpu_complex64 PASSED [ 63%] 2023-03-31T06:24:13.4140236Z test_linalg.py::TestLinalgCPU::test_outer_cpu_float16 PASSED [ 63%] 2023-03-31T06:24:13.4140536Z test_linalg.py::TestLinalgCPU::test_outer_cpu_float32 PASSED [ 64%] 2023-03-31T06:24:13.4140847Z test_linalg.py::TestLinalgCPU::test_outer_cpu_float64 PASSED [ 64%] 2023-03-31T06:24:13.4141188Z test_linalg.py::TestLinalgCPU::test_outer_cpu_int16 PASSED [ 64%] 2023-03-31T06:24:13.4141490Z test_linalg.py::TestLinalgCPU::test_outer_cpu_int32 PASSED [ 64%] 2023-03-31T06:24:13.4141784Z test_linalg.py::TestLinalgCPU::test_outer_cpu_int64 PASSED [ 64%] 2023-03-31T06:24:13.4142088Z test_linalg.py::TestLinalgCPU::test_outer_cpu_int8 PASSED [ 64%] 2023-03-31T06:24:13.4142397Z test_linalg.py::TestLinalgCPU::test_outer_cpu_uint8 PASSED [ 64%] 2023-03-31T06:24:13.4142713Z test_linalg.py::TestLinalgCPU::test_outer_ger_addr_legacy_tests_cpu PASSED [ 65%] 2023-03-31T06:24:13.4143067Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_bfloat16 PASSED [ 65%] 2023-03-31T06:24:13.4143516Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_bool PASSED [ 65%] 2023-03-31T06:24:13.4143922Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_complex128 PASSED [ 65%] 2023-03-31T06:24:13.4144280Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_complex64 PASSED [ 65%] 2023-03-31T06:24:13.4144647Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float16 PASSED [ 65%] 2023-03-31T06:24:13.4145010Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float32 PASSED [ 65%] 2023-03-31T06:24:13.4145358Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_float64 PASSED [ 65%] 2023-03-31T06:24:13.4145714Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int16 PASSED [ 66%] 2023-03-31T06:24:13.4146062Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int32 PASSED [ 66%] 2023-03-31T06:24:13.4146415Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int64 PASSED [ 66%] 2023-03-31T06:24:13.4146760Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_int8 PASSED [ 66%] 2023-03-31T06:24:13.4147115Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bfloat16_uint8 PASSED [ 66%] 2023-03-31T06:24:13.4147469Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_bfloat16 PASSED [ 66%] 2023-03-31T06:24:13.4147809Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_bool PASSED [ 66%] 2023-03-31T06:24:13.4148170Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_complex128 PASSED [ 67%] 2023-03-31T06:24:13.4148568Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_complex64 PASSED [ 67%] 2023-03-31T06:24:13.4148955Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float16 PASSED [ 67%] 2023-03-31T06:24:13.4149297Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float32 PASSED [ 67%] 2023-03-31T06:24:13.4149653Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_float64 PASSED [ 67%] 2023-03-31T06:24:13.4150008Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int16 PASSED [ 67%] 2023-03-31T06:24:13.4150347Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int32 PASSED [ 67%] 2023-03-31T06:24:13.4150694Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int64 PASSED [ 68%] 2023-03-31T06:24:13.4151043Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_int8 PASSED [ 68%] 2023-03-31T06:24:13.4151398Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_bool_uint8 PASSED [ 68%] 2023-03-31T06:24:13.4151745Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_bfloat16 PASSED [ 68%] 2023-03-31T06:24:13.4152116Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_bool PASSED [ 68%] 2023-03-31T06:24:13.4152490Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_complex128 PASSED [ 68%] 2023-03-31T06:24:13.4152867Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_complex64 PASSED [ 68%] 2023-03-31T06:24:13.4153228Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float16 PASSED [ 68%] 2023-03-31T06:24:13.4153633Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float32 PASSED [ 69%] 2023-03-31T06:24:13.4154002Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_float64 PASSED [ 69%] 2023-03-31T06:24:13.4154351Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int16 PASSED [ 69%] 2023-03-31T06:24:13.4154717Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int32 PASSED [ 69%] 2023-03-31T06:24:13.4155077Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int64 PASSED [ 69%] 2023-03-31T06:24:13.4155442Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_int8 PASSED [ 69%] 2023-03-31T06:24:13.4155787Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex128_uint8 PASSED [ 69%] 2023-03-31T06:24:13.4156197Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_bfloat16 PASSED [ 70%] 2023-03-31T06:24:13.4156558Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_bool PASSED [ 70%] 2023-03-31T06:24:13.4156916Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_complex128 PASSED [ 70%] 2023-03-31T06:24:13.4157286Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_complex64 PASSED [ 70%] 2023-03-31T06:24:13.4157652Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float16 PASSED [ 70%] 2023-03-31T06:24:13.4158017Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float32 PASSED [ 70%] 2023-03-31T06:24:13.4158362Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_float64 PASSED [ 70%] 2023-03-31T06:24:13.4158723Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int16 PASSED [ 70%] 2023-03-31T06:24:13.4159079Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int32 PASSED [ 71%] 2023-03-31T06:24:13.4159434Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int64 PASSED [ 71%] 2023-03-31T06:24:13.4159773Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_int8 PASSED [ 71%] 2023-03-31T06:24:13.4160129Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_complex64_uint8 PASSED [ 71%] 2023-03-31T06:24:13.4160483Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_bfloat16 PASSED [ 71%] 2023-03-31T06:24:13.4160873Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_bool PASSED [ 71%] 2023-03-31T06:24:13.4161263Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_complex128 PASSED [ 71%] 2023-03-31T06:24:13.4161623Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_complex64 PASSED [ 72%] 2023-03-31T06:24:13.4161979Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float16 PASSED [ 72%] 2023-03-31T06:24:13.4162318Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float32 PASSED [ 72%] 2023-03-31T06:24:13.4162674Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_float64 PASSED [ 72%] 2023-03-31T06:24:13.4163196Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int16 PASSED [ 72%] 2023-03-31T06:24:13.4163613Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int32 PASSED [ 72%] 2023-03-31T06:24:13.4163969Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int64 PASSED [ 72%] 2023-03-31T06:24:13.4164329Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_int8 PASSED [ 72%] 2023-03-31T06:24:13.4164684Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float16_uint8 PASSED [ 73%] 2023-03-31T06:24:13.4165025Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_bfloat16 PASSED [ 73%] 2023-03-31T06:24:13.4165384Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_bool PASSED [ 73%] 2023-03-31T06:24:13.4165750Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_complex128 PASSED [ 73%] 2023-03-31T06:24:13.4166193Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_complex64 PASSED [ 73%] 2023-03-31T06:24:13.4166547Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float16 PASSED [ 73%] 2023-03-31T06:24:13.4166902Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float32 PASSED [ 73%] 2023-03-31T06:24:13.4167260Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_float64 PASSED [ 74%] 2023-03-31T06:24:13.4167601Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int16 PASSED [ 74%] 2023-03-31T06:24:13.4167956Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int32 PASSED [ 74%] 2023-03-31T06:24:13.4168309Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int64 PASSED [ 74%] 2023-03-31T06:24:13.4168719Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_int8 PASSED [ 74%] 2023-03-31T06:24:13.4169063Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float32_uint8 PASSED [ 74%] 2023-03-31T06:24:13.4169420Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_bfloat16 PASSED [ 74%] 2023-03-31T06:24:13.4169780Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_bool PASSED [ 75%] 2023-03-31T06:24:13.4170132Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_complex128 PASSED [ 75%] 2023-03-31T06:24:13.4170504Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_complex64 PASSED [ 75%] 2023-03-31T06:24:13.4170870Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float16 PASSED [ 75%] 2023-03-31T06:24:13.4171228Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float32 PASSED [ 75%] 2023-03-31T06:24:13.4171571Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_float64 PASSED [ 75%] 2023-03-31T06:24:13.4171923Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int16 PASSED [ 75%] 2023-03-31T06:24:13.4172279Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int32 PASSED [ 75%] 2023-03-31T06:24:13.4172615Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int64 PASSED [ 76%] 2023-03-31T06:24:13.4172967Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_int8 PASSED [ 76%] 2023-03-31T06:24:13.4173356Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_float64_uint8 PASSED [ 76%] 2023-03-31T06:24:13.4173748Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_bfloat16 PASSED [ 76%] 2023-03-31T06:24:13.4174086Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_bool PASSED [ 76%] 2023-03-31T06:24:13.4174438Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_complex128 PASSED [ 76%] 2023-03-31T06:24:13.4174799Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_complex64 PASSED [ 76%] 2023-03-31T06:24:13.4175154Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float16 PASSED [ 77%] 2023-03-31T06:24:13.4175490Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float32 PASSED [ 77%] 2023-03-31T06:24:13.4175840Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_float64 PASSED [ 77%] 2023-03-31T06:24:13.4176189Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int16 PASSED [ 77%] 2023-03-31T06:24:13.4176526Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int32 PASSED [ 77%] 2023-03-31T06:24:13.4176876Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int64 PASSED [ 77%] 2023-03-31T06:24:13.4177223Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_int8 PASSED [ 77%] 2023-03-31T06:24:13.4177568Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int16_uint8 PASSED [ 77%] 2023-03-31T06:24:13.4177906Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_bfloat16 PASSED [ 78%] 2023-03-31T06:24:13.4178297Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_bool PASSED [ 78%] 2023-03-31T06:24:13.4178657Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_complex128 PASSED [ 78%] 2023-03-31T06:24:13.4179006Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_complex64 PASSED [ 78%] 2023-03-31T06:24:13.4179359Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float16 PASSED [ 78%] 2023-03-31T06:24:13.4179712Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float32 PASSED [ 78%] 2023-03-31T06:24:13.4180064Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_float64 PASSED [ 78%] 2023-03-31T06:24:13.4180397Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int16 PASSED [ 79%] 2023-03-31T06:24:13.4180740Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int32 PASSED [ 79%] 2023-03-31T06:24:13.4181121Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int64 PASSED [ 79%] 2023-03-31T06:24:13.4181458Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_int8 PASSED [ 79%] 2023-03-31T06:24:13.4181806Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int32_uint8 PASSED [ 79%] 2023-03-31T06:24:13.4182154Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_bfloat16 PASSED [ 79%] 2023-03-31T06:24:13.4182509Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_bool PASSED [ 79%] 2023-03-31T06:24:13.4182850Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_complex128 PASSED [ 79%] 2023-03-31T06:24:13.4183293Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_complex64 PASSED [ 80%] 2023-03-31T06:24:13.4183660Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float16 PASSED [ 80%] 2023-03-31T06:24:13.4184015Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float32 PASSED [ 80%] 2023-03-31T06:24:13.4184355Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_float64 PASSED [ 80%] 2023-03-31T06:24:13.4184708Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int16 PASSED [ 80%] 2023-03-31T06:24:13.4185059Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int32 PASSED [ 80%] 2023-03-31T06:24:13.4185392Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int64 PASSED [ 80%] 2023-03-31T06:24:13.4185780Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_int8 PASSED [ 81%] 2023-03-31T06:24:13.4186160Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int64_uint8 PASSED [ 81%] 2023-03-31T06:24:13.4186512Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_bfloat16 PASSED [ 81%] 2023-03-31T06:24:13.4186850Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_bool PASSED [ 81%] 2023-03-31T06:24:13.4187206Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_complex128 PASSED [ 81%] 2023-03-31T06:24:13.4187571Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_complex64 PASSED [ 81%] 2023-03-31T06:24:13.4187910Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float16 PASSED [ 81%] 2023-03-31T06:24:13.4188261Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float32 PASSED [ 81%] 2023-03-31T06:24:13.4188620Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_float64 PASSED [ 82%] 2023-03-31T06:24:13.4188969Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int16 PASSED [ 82%] 2023-03-31T06:24:13.4189306Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int32 PASSED [ 82%] 2023-03-31T06:24:13.4189648Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int64 PASSED [ 82%] 2023-03-31T06:24:13.4189990Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_int8 PASSED [ 82%] 2023-03-31T06:24:13.4190324Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_int8_uint8 PASSED [ 82%] 2023-03-31T06:24:13.4190715Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_bfloat16 PASSED [ 82%] 2023-03-31T06:24:13.4191069Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_bool PASSED [ 83%] 2023-03-31T06:24:13.4191421Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_complex128 PASSED [ 83%] 2023-03-31T06:24:13.4191774Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_complex64 PASSED [ 83%] 2023-03-31T06:24:13.4192136Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float16 PASSED [ 83%] 2023-03-31T06:24:13.4192490Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float32 PASSED [ 83%] 2023-03-31T06:24:13.4192828Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_float64 PASSED [ 83%] 2023-03-31T06:24:13.4193177Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int16 PASSED [ 83%] 2023-03-31T06:24:13.4193557Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int32 PASSED [ 84%] 2023-03-31T06:24:13.4193902Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int64 PASSED [ 84%] 2023-03-31T06:24:13.4194235Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_int8 PASSED [ 84%] 2023-03-31T06:24:13.4194585Z test_linalg.py::TestLinalgCPU::test_outer_type_promotion_cpu_uint8_uint8 PASSED [ 84%] 2023-03-31T06:24:13.4194922Z test_linalg.py::TestLinalgCPU::test_pca_lowrank_cpu PASSED [ 84%] 2023-03-31T06:24:13.4195243Z test_linalg.py::TestLinalgCPU::test_permute_matmul_cpu PASSED [ 84%] 2023-03-31T06:24:13.4195555Z test_linalg.py::TestLinalgCPU::test_pinv_cpu_complex128 PASSED [ 84%] 2023-03-31T06:24:13.4195869Z test_linalg.py::TestLinalgCPU::test_pinv_cpu_complex64 PASSED [ 84%] 2023-03-31T06:24:13.4196186Z test_linalg.py::TestLinalgCPU::test_pinv_cpu_float32 PASSED [ 85%] 2023-03-31T06:24:13.4196490Z test_linalg.py::TestLinalgCPU::test_pinv_cpu_float64 PASSED [ 85%] 2023-03-31T06:24:13.4196835Z test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_complex128 PASSED [ 85%] 2023-03-31T06:24:13.4197198Z test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_complex64 PASSED [ 85%] 2023-03-31T06:24:13.4197548Z test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_float32 PASSED [ 85%] 2023-03-31T06:24:13.4197878Z test_linalg.py::TestLinalgCPU::test_pinv_errors_and_warnings_cpu_float64 PASSED [ 85%] 2023-03-31T06:24:13.4198252Z test_linalg.py::TestLinalgCPU::test_pinverse_cpu_complex128 PASSED [ 85%] 2023-03-31T06:24:13.4198605Z test_linalg.py::TestLinalgCPU::test_pinverse_cpu_complex64 PASSED [ 86%] 2023-03-31T06:24:13.4198912Z test_linalg.py::TestLinalgCPU::test_pinverse_cpu_float32 PASSED [ 86%] 2023-03-31T06:24:13.4199227Z test_linalg.py::TestLinalgCPU::test_pinverse_cpu_float64 PASSED [ 86%] 2023-03-31T06:24:13.4199585Z test_linalg.py::TestLinalgCPU::test_preferred_linalg_library_cpu SKIPPED (Only runs on cuda) [ 86%] 2023-03-31T06:24:13.4199945Z test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_complex128 PASSED [ 86%] 2023-03-31T06:24:13.4200259Z test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_complex64 PASSED [ 86%] 2023-03-31T06:24:13.4200578Z test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_float32 PASSED [ 86%] 2023-03-31T06:24:13.4200899Z test_linalg.py::TestLinalgCPU::test_qr_batched_cpu_float64 PASSED [ 86%] 2023-03-31T06:24:13.4201207Z test_linalg.py::TestLinalgCPU::test_qr_cpu_complex128 PASSED [ 87%] 2023-03-31T06:24:13.4201523Z test_linalg.py::TestLinalgCPU::test_qr_cpu_complex64 PASSED [ 87%] 2023-03-31T06:24:13.4201834Z test_linalg.py::TestLinalgCPU::test_qr_cpu_float32 PASSED [ 87%] 2023-03-31T06:24:13.4202141Z test_linalg.py::TestLinalgCPU::test_qr_cpu_float64 PASSED [ 87%] 2023-03-31T06:24:13.4202448Z test_linalg.py::TestLinalgCPU::test_qr_error_cases_cpu_float32 PASSED [ 87%] 2023-03-31T06:24:13.4202839Z test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_complex128 PASSED [ 87%] 2023-03-31T06:24:13.4203365Z test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_complex64 PASSED [ 87%] 2023-03-31T06:24:13.4203676Z test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_float32 PASSED [ 88%] 2023-03-31T06:24:13.4203993Z test_linalg.py::TestLinalgCPU::test_qr_vs_numpy_cpu_float64 PASSED [ 88%] 2023-03-31T06:24:13.4204305Z test_linalg.py::TestLinalgCPU::test_renorm_cpu PASSED [ 88%] 2023-03-31T06:24:13.4204613Z test_linalg.py::TestLinalgCPU::test_renorm_ps_cpu PASSED [ 88%] 2023-03-31T06:24:13.4204914Z test_linalg.py::TestLinalgCPU::test_slogdet_cpu_complex128 PASSED [ 88%] 2023-03-31T06:24:13.4205233Z test_linalg.py::TestLinalgCPU::test_slogdet_cpu_complex64 PASSED [ 88%] 2023-03-31T06:24:13.4205550Z test_linalg.py::TestLinalgCPU::test_slogdet_cpu_float32 PASSED [ 88%] 2023-03-31T06:24:13.4205902Z test_linalg.py::TestLinalgCPU::test_slogdet_cpu_float64 PASSED [ 88%] 2023-03-31T06:24:13.4206249Z test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_complex128 PASSED [ 89%] 2023-03-31T06:24:13.4206619Z test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_complex64 PASSED [ 89%] 2023-03-31T06:24:13.4206982Z test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_float32 PASSED [ 89%] 2023-03-31T06:24:13.4207324Z test_linalg.py::TestLinalgCPU::test_slogdet_errors_and_warnings_cpu_float64 PASSED [ 89%] 2023-03-31T06:24:13.4207686Z test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_complex128 PASSED [ 89%] 2023-03-31T06:24:13.4208053Z test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_complex64 PASSED [ 89%] 2023-03-31T06:24:13.4208403Z test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_float32 PASSED [ 89%] 2023-03-31T06:24:13.4208767Z test_linalg.py::TestLinalgCPU::test_solve_batched_broadcasting_cpu_float64 PASSED [ 90%] 2023-03-31T06:24:13.4209109Z test_linalg.py::TestLinalgCPU::test_solve_cpu_complex128 PASSED [ 90%] 2023-03-31T06:24:13.4209430Z test_linalg.py::TestLinalgCPU::test_solve_cpu_complex64 PASSED [ 90%] 2023-03-31T06:24:13.4209787Z test_linalg.py::TestLinalgCPU::test_solve_cpu_float32 PASSED [ 90%] 2023-03-31T06:24:13.4210178Z test_linalg.py::TestLinalgCPU::test_solve_cpu_float64 PASSED [ 90%] 2023-03-31T06:24:13.4210705Z test_linalg.py::TestLinalgCPU::test_solve_removed_error_cpu PASSED [ 90%] 2023-03-31T06:24:13.4211088Z test_linalg.py::TestLinalgCPU::test_strided_mm_bmm_cpu_float32 PASSED [ 90%] 2023-03-31T06:24:13.4211482Z test_linalg.py::TestLinalgCPU::test_strided_mm_bmm_cpu_float64 PASSED [ 90%] 2023-03-31T06:24:13.4211857Z test_linalg.py::TestLinalgCPU::test_svd_cpu_complex128 PASSED [ 91%] 2023-03-31T06:24:13.4212228Z test_linalg.py::TestLinalgCPU::test_svd_cpu_complex64 PASSED [ 91%] 2023-03-31T06:24:13.4212584Z test_linalg.py::TestLinalgCPU::test_svd_cpu_float32 PASSED [ 91%] 2023-03-31T06:24:13.4212995Z test_linalg.py::TestLinalgCPU::test_svd_cpu_float64 PASSED [ 91%] 2023-03-31T06:24:13.4213366Z test_linalg.py::TestLinalgCPU::test_svd_lowrank_cpu_float64 PASSED [ 91%] 2023-03-31T06:24:13.4213721Z test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_complex128 PASSED [ 91%] 2023-03-31T06:24:13.4226708Z test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_complex64 PASSED [ 91%] 2023-03-31T06:24:13.4227083Z test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_float32 PASSED [ 92%] 2023-03-31T06:24:13.4227430Z test_linalg.py::TestLinalgCPU::test_svd_memory_allocation_cpu_float64 PASSED [ 92%] 2023-03-31T06:24:13.4227771Z test_linalg.py::TestLinalgCPU::test_symeig_removed_error_cpu PASSED [ 92%] 2023-03-31T06:24:13.4228100Z test_linalg.py::TestLinalgCPU::test_tensordot_cpu PASSED [ 92%] 2023-03-31T06:24:13.4228421Z test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_complex128 PASSED [ 92%] 2023-03-31T06:24:13.4228900Z test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_complex64 PASSED [ 92%] 2023-03-31T06:24:13.4229230Z test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_float32 PASSED [ 92%] 2023-03-31T06:24:13.4229542Z test_linalg.py::TestLinalgCPU::test_tensorinv_cpu_float64 PASSED [ 93%] 2023-03-31T06:24:13.4229871Z test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_complex128 PASSED [ 93%] 2023-03-31T06:24:13.4230210Z test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_complex64 PASSED [ 93%] 2023-03-31T06:24:13.4230548Z test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_float32 PASSED [ 93%] 2023-03-31T06:24:13.4230863Z test_linalg.py::TestLinalgCPU::test_tensorinv_empty_cpu_float64 PASSED [ 93%] 2023-03-31T06:24:13.4231208Z test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_complex128 PASSED [ 93%] 2023-03-31T06:24:13.4231623Z test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_complex64 PASSED [ 93%] 2023-03-31T06:24:13.4231979Z test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_float32 PASSED [ 93%] 2023-03-31T06:24:13.4232341Z test_linalg.py::TestLinalgCPU::test_tensorinv_errors_and_warnings_cpu_float64 PASSED [ 94%] 2023-03-31T06:24:13.4232704Z test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_complex128 PASSED [ 94%] 2023-03-31T06:24:13.4233064Z test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_complex64 PASSED [ 94%] 2023-03-31T06:24:13.4233411Z test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_float32 PASSED [ 94%] 2023-03-31T06:24:13.4233762Z test_linalg.py::TestLinalgCPU::test_tensorinv_singular_input_cpu_float64 PASSED [ 94%] 2023-03-31T06:24:13.4234106Z test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_complex128 PASSED [ 94%] 2023-03-31T06:24:13.4234426Z test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_complex64 PASSED [ 94%] 2023-03-31T06:24:13.4234761Z test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_float32 PASSED [ 95%] 2023-03-31T06:24:13.4235086Z test_linalg.py::TestLinalgCPU::test_tensorsolve_cpu_float64 PASSED [ 95%] 2023-03-31T06:24:13.4235424Z test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_complex128 PASSED [ 95%] 2023-03-31T06:24:13.4235757Z test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_complex64 PASSED [ 95%] 2023-03-31T06:24:13.4236094Z test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_float32 PASSED [ 95%] 2023-03-31T06:24:13.4236470Z test_linalg.py::TestLinalgCPU::test_tensorsolve_empty_cpu_float64 PASSED [ 95%] 2023-03-31T06:24:13.4236845Z test_linalg.py::TestLinalgCPU::test_tensorsolve_errors_and_warnings_cpu_float32 PASSED [ 95%] 2023-03-31T06:24:13.4237234Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_complex128 PASSED [ 95%] 2023-03-31T06:24:13.4237627Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_complex64 PASSED [ 96%] 2023-03-31T06:24:13.4238020Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_float32 PASSED [ 96%] 2023-03-31T06:24:13.4238395Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_broadcasting_cpu_float64 PASSED [ 96%] 2023-03-31T06:24:13.4238770Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_complex128 PASSED [ 96%] 2023-03-31T06:24:13.4239128Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_complex64 PASSED [ 96%] 2023-03-31T06:24:13.4239485Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_float32 PASSED [ 96%] 2023-03-31T06:24:13.4239823Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_cpu_float64 PASSED [ 96%] 2023-03-31T06:24:13.4240264Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 97%] 2023-03-31T06:24:13.4240782Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 97%] 2023-03-31T06:24:13.4241332Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 97%] 2023-03-31T06:24:13.4241830Z test_linalg.py::TestLinalgCPU::test_triangular_solve_batched_many_batches_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 97%] 2023-03-31T06:24:13.4242253Z test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_complex128 PASSED [ 97%] 2023-03-31T06:24:13.4242602Z test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_complex64 PASSED [ 97%] 2023-03-31T06:24:13.4242939Z test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_float32 PASSED [ 97%] 2023-03-31T06:24:13.4243493Z test_linalg.py::TestLinalgCPU::test_triangular_solve_cpu_float64 PASSED [ 97%] 2023-03-31T06:24:13.4243865Z test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_complex128 PASSED [ 98%] 2023-03-31T06:24:13.4244337Z test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_complex64 PASSED [ 98%] 2023-03-31T06:24:13.4244714Z test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_float32 PASSED [ 98%] 2023-03-31T06:24:13.4245101Z test_linalg.py::TestLinalgCPU::test_triangular_solve_out_errors_and_warnings_cpu_float64 PASSED [ 98%] 2023-03-31T06:24:13.4245457Z test_linalg.py::TestLinalgCPU::test_vdot_invalid_args_cpu PASSED [ 98%] 2023-03-31T06:24:13.4245790Z test_linalg.py::TestLinalgCPU::test_vdot_vs_numpy_cpu_complex64 PASSED [ 98%] 2023-03-31T06:24:13.4246109Z test_linalg.py::TestLinalgCPU::test_vdot_vs_numpy_cpu_float32 PASSED [ 98%] 2023-03-31T06:24:13.4246429Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:24:13.4246757Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_complex128 PASSED [ 99%] 2023-03-31T06:24:13.4247074Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_complex64 PASSED [ 99%] 2023-03-31T06:24:13.4247388Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float16 PASSED [ 99%] 2023-03-31T06:24:13.4247704Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float32 PASSED [ 99%] 2023-03-31T06:24:13.4248020Z test_linalg.py::TestLinalgCPU::test_vector_norm_cpu_float64 PASSED [ 99%] 2023-03-31T06:24:13.4248333Z test_linalg.py::TestLinalgCPU::test_vector_norm_dim_tuple_arg_cpu PASSED [ 99%] 2023-03-31T06:24:13.4248702Z test_linalg.py::TestLinalgCPU::test_vector_norm_extreme_values_cpu PASSED [100%] 2023-03-31T06:24:13.4248930Z 2023-03-31T06:24:13.4249383Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_linalg/test_linalg-7b54af76081f9243.xml - 2023-03-31T06:24:13.4249731Z ================== 672 passed, 72 skipped in 89.34s (0:01:29) ================== 2023-03-31T06:24:13.4250051Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:24:13.4250260Z 2023-03-31T06:24:13.4250657Z ##[endgroup] 2023-03-31T06:24:13.4251055Z FINISHED PRINTING LOG FILE of test_linalg (/var/lib/jenkins/workspace/test/test-reports/test_linalg_930rev50.log) 2023-03-31T06:24:13.4251263Z 2023-03-31T06:24:13.4251429Z Running test_autograd ... [2023-03-31 06:24:13.385092] 2023-03-31T06:24:13.4252045Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_autograd.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:24:13.385425] 2023-03-31T06:24:33.6200447Z 2023-03-31T06:24:33.6201325Z Expand the folded group to see the log file of test_autograd 2023-03-31T06:24:33.6204122Z ##[group]PRINTING LOG FILE of test_autograd (/var/lib/jenkins/workspace/test/test-reports/test_autograd_b7ivmmwj.log) 2023-03-31T06:24:33.6205138Z Test results will be stored in test-reports/python-pytest/test_autograd/test_autograd-e1c0d42718cf0b95.xml 2023-03-31T06:24:33.6205980Z ============================= test session starts ============================== 2023-03-31T06:24:33.6207398Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:24:33.6207928Z cachedir: .pytest_cache 2023-03-31T06:24:33.6208615Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:24:33.6209195Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:24:33.6209884Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:24:33.6210412Z collecting ... collected 574 items 2023-03-31T06:24:33.6272795Z Running 574 items in this shard: test/test_autograd.py::TestAutograd::test_access_saved_tensor_twice_without_recomputation_works, test/test_autograd.py::TestAutograd::test_accumulate_grad, test/test_autograd.py::TestAutograd::test_accumulate_grad_posthooks_can_observe_tensor_prehook, test/test_autograd.py::TestAutograd::test_accumulate_grad_tensor_reference, test/test_autograd.py::TestAutograd::test_accumulate_grad_with_zero_numel_grad, test/test_autograd.py::TestAutograd::test_anomaly_assign_parent_cleanup, test/test_autograd.py::TestAutograd::test_anomaly_detect_nan, test/test_autograd.py::TestAutograd::test_anomaly_grad_warnings, test/test_autograd.py::TestAutograd::test_anomaly_mode_no_check_nan, test/test_autograd.py::TestAutograd::test_attribute_deletion, test/test_autograd.py::TestAutograd::test_autograd_inplace_view_of_view, test/test_autograd.py::TestAutograd::test_autograd_inplace_views_creation_meta, test/test_autograd.py::TestAutograd::test_autograd_inplace_views_cross_dtype, test/test_autograd.py::TestAutograd::test_autograd_multiple_views_python, test/test_autograd.py::TestAutograd::test_autograd_node_isinstance, test/test_autograd.py::TestAutograd::test_autograd_python_custom_function_inplace, test/test_autograd.py::TestAutograd::test_autograd_simple_views_python, test/test_autograd.py::TestAutograd::test_autograd_views_codegen, test/test_autograd.py::TestAutograd::test_backward, test/test_autograd.py::TestAutograd::test_backward_badcalls, test/test_autograd.py::TestAutograd::test_backward_copy, test/test_autograd.py::TestAutograd::test_backward_create_graph_warns, test/test_autograd.py::TestAutograd::test_backward_no_grad, test/test_autograd.py::TestAutograd::test_backward_twice_retained_graph_with_saved_values, test/test_autograd.py::TestAutograd::test_backward_twice_retained_graph_without_saved_values, test/test_autograd.py::TestAutograd::test_backward_twice_with_saved_values, test/test_autograd.py::TestAutograd::test_backward_twice_without_saved_values, test/test_autograd.py::TestAutograd::test_backward_with_inputs, test/test_autograd.py::TestAutograd::test_backward_with_nonleaf_inputs, test/test_autograd.py::TestAutograd::test_calculate_shape_util, test/test_autograd.py::TestAutograd::test_callback_adds_callback, test/test_autograd.py::TestAutograd::test_cant_create_saved_tensors, test/test_autograd.py::TestAutograd::test_checkpoint_valid_reset_on_error, test/test_autograd.py::TestAutograd::test_checkpointing, test/test_autograd.py::TestAutograd::test_checkpointing_non_reentrant_autocast_cpu, test/test_autograd.py::TestAutograd::test_checkpointing_non_reentrant_autocast_gpu, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_arbitrary_input_output, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_correct_grad, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_custom_function_works, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_dataparallel, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_False, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_True, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_input_requires_grad_False, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_input_requires_grad_True, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_memory_savings, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_parameter_used_in_an_out, test/test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_with_context_fn, test/test_autograd.py::TestAutograd::test_copy_slices_graph_task_updates, test/test_autograd.py::TestAutograd::test_create_graph_and_full_backward_hook_cycle, test/test_autograd.py::TestAutograd::test_current_graph_task_execution_order, test/test_autograd.py::TestAutograd::test_current_graph_task_id, test/test_autograd.py::TestAutograd::test_current_node, test/test_autograd.py::TestAutograd::test_custom_autograd_no_early_free, test/test_autograd.py::TestAutograd::test_custom_autograd_repeated_grad_grad, test/test_autograd.py::TestAutograd::test_custom_function_cycle, test/test_autograd.py::TestAutograd::test_custom_function_error, test/test_autograd.py::TestAutograd::test_custom_function_exception, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_forward_is_no_op, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_inplace_checks, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_non_differentiable, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_non_tensor_before_tensor_args, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_view_checks, test/test_autograd.py::TestAutograd::test_custom_function_forward_mode_wrong_formula, test/test_autograd.py::TestAutograd::test_custom_function_local_inplace, test/test_autograd.py::TestAutograd::test_custom_function_mark_dirty_not_differentiable, test/test_autograd.py::TestAutograd::test_custom_function_no_tensors, test/test_autograd.py::TestAutograd::test_custom_function_non_tensor_inputs_outputs, test/test_autograd.py::TestAutograd::test_custom_function_return_view_in_nograd, test/test_autograd.py::TestAutograd::test_custom_function_save_for_forward, test/test_autograd.py::TestAutograd::test_custom_function_saved_tensors, test/test_autograd.py::TestAutograd::test_custom_function_setup_context_multi_input, test/test_autograd.py::TestAutograd::test_custom_function_setup_context_multi_output, test/test_autograd.py::TestAutograd::test_custom_function_setup_context_simple, test/test_autograd.py::TestAutograd::test_custom_function_vmap_defaults, test/test_autograd.py::TestAutograd::test_deep_reentrant, test/test_autograd.py::TestAutograd::test_default_saved_variable_hooks_double_backward, test/test_autograd.py::TestAutograd::test_dep_nograd, test/test_autograd.py::TestAutograd::test_dependent_backward, test/test_autograd.py::TestAutograd::test_detach, test/test_autograd.py::TestAutograd::test_detach_base, test/test_autograd.py::TestAutograd::test_detach_then_inplace_raises_in_autograd, test/test_autograd.py::TestAutograd::test_diagonal_expanded_v, test/test_autograd.py::TestAutograd::test_dir, test/test_autograd.py::TestAutograd::test_disabling_saved_tensor_hooks, test/test_autograd.py::TestAutograd::test_disabling_saved_tensor_hooks_nested, test/test_autograd.py::TestAutograd::test_dont_materialize_grads, test/test_autograd.py::TestAutograd::test_duplicate_backward_root, test/test_autograd.py::TestAutograd::test_first_grad_fn_access_in_no_grad_mode, test/test_autograd.py::TestAutograd::test_free_deep_graph, test/test_autograd.py::TestAutograd::test_free_deep_graph_complicated, test/test_autograd.py::TestAutograd::test_free_deep_graph_pyfunction, test/test_autograd.py::TestAutograd::test_full_backward_hook_double_backward, test/test_autograd.py::TestAutograd::test_function, test/test_autograd.py::TestAutograd::test_function_returns_input, test/test_autograd.py::TestAutograd::test_function_returns_undefined_tensor, test/test_autograd.py::TestAutograd::test_gc_in_destructor, test/test_autograd.py::TestAutograd::test_grad, test/test_autograd.py::TestAutograd::test_grad_badcalls, test/test_autograd.py::TestAutograd::test_grad_batched_grad, test/test_autograd.py::TestAutograd::test_grad_empty_inputs, test/test_autograd.py::TestAutograd::test_grad_fn_attr_bindings, test/test_autograd.py::TestAutograd::test_grad_fn_badcalls, test/test_autograd.py::TestAutograd::test_grad_fn_prehooks, test/test_autograd.py::TestAutograd::test_grad_fn_prehooks_multiple_outputs, test/test_autograd.py::TestAutograd::test_grad_fn_prehooks_remove_hooks, test/test_autograd.py::TestAutograd::test_grad_materialize_grads, test/test_autograd.py::TestAutograd::test_grad_mode_class_decoration, test/test_autograd.py::TestAutograd::test_grad_mode_restored_reentrant, test/test_autograd.py::TestAutograd::test_grad_nonleaf, test/test_autograd.py::TestAutograd::test_grad_nonleaf_many_outputs, test/test_autograd.py::TestAutograd::test_grad_nonleaf_register_hook, test/test_autograd.py::TestAutograd::test_grad_unreachable, test/test_autograd.py::TestAutograd::test_grad_unreachable_discovery, test/test_autograd.py::TestAutograd::test_gradcheck_backward_mul_by_grad_output, test/test_autograd.py::TestAutograd::test_gradcheck_check_batched_grad, test/test_autograd.py::TestAutograd::test_gradcheck_check_forward_or_backward_only, test/test_autograd.py::TestAutograd::test_gradcheck_check_no_differentiable_outputs, test/test_autograd.py::TestAutograd::test_gradcheck_complex_non_complex_outputs, test/test_autograd.py::TestAutograd::test_gradcheck_custom_error, test/test_autograd.py::TestAutograd::test_gradcheck_dense_and_sparse_inputs, test/test_autograd.py::TestAutograd::test_gradcheck_forward_ad, test/test_autograd.py::TestAutograd::test_gradcheck_forward_ad_batched_grad, test/test_autograd.py::TestAutograd::test_gradcheck_forward_ad_respects_requires_grad, test/test_autograd.py::TestAutograd::test_gradcheck_forward_ad_runs_with_no_requires_grad, test/test_autograd.py::TestAutograd::test_gradcheck_get_analytical_jacobian, test/test_autograd.py::TestAutograd::test_gradcheck_get_numerical_jacobian, test/test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_bsc, test/test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_bsr, test/test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_coo, test/test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_csc, test/test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_csr, test/test_autograd.py::TestAutograd::test_gradcheck_jacobian_mismatch, test/test_autograd.py::TestAutograd::test_gradcheck_multiple_mkldnn_inputs, test/test_autograd.py::TestAutograd::test_gradcheck_nondeterministic, test/test_autograd.py::TestAutograd::test_gradcheck_output_shape_or_dtype_depend_on_values, test/test_autograd.py::TestAutograd::test_gradcheck_single_input, test/test_autograd.py::TestAutograd::test_gradcheck_test_outputs, test/test_autograd.py::TestAutograd::test_gradcheck_undefined_grad, test/test_autograd.py::TestAutograd::test_gradcheck_validates_input_mkldnn, test/test_autograd.py::TestAutograd::test_gradcheck_validates_inputs, test/test_autograd.py::TestAutograd::test_graph_save_on_cpu, test/test_autograd.py::TestAutograd::test_graph_save_on_cpu_cuda, test/test_autograd.py::TestAutograd::test_hessian_vector, test/test_autograd.py::TestAutograd::test_hook_edge_case_when_called_with_grad, test/test_autograd.py::TestAutograd::test_hook_none, test/test_autograd.py::TestAutograd::test_hook_with_no_name, test/test_autograd.py::TestAutograd::test_hooks, test/test_autograd.py::TestAutograd::test_hooks_cpp, test/test_autograd.py::TestAutograd::test_increment_version, test/test_autograd.py::TestAutograd::test_index_backward_does_not_save_tensor, test/test_autograd.py::TestAutograd::test_indexing, test/test_autograd.py::TestAutograd::test_indexing_duplicates, test/test_autograd.py::TestAutograd::test_inplace, test/test_autograd.py::TestAutograd::test_inplace_not_requires_grad, test/test_autograd.py::TestAutograd::test_inplace_on_view_backward, test/test_autograd.py::TestAutograd::test_inplace_on_view_leaf_errors, test/test_autograd.py::TestAutograd::test_inplace_on_view_saved_output, test/test_autograd.py::TestAutograd::test_inplace_on_view_weak_grad_fn, test/test_autograd.py::TestAutograd::test_input_buffer_accum, test/test_autograd.py::TestAutograd::test_integer_outputs, test/test_autograd.py::TestAutograd::test_invalid_gradients, test/test_autograd.py::TestAutograd::test_isolated_node, test/test_autograd.py::TestAutograd::test_leaf_assignment, test/test_autograd.py::TestAutograd::test_legacy_function_deprecation_exception, test/test_autograd.py::TestAutograd::test_lobpcg, test/test_autograd.py::TestAutograd::test_mark_non_differentiable, test/test_autograd.py::TestAutograd::test_mark_non_differentiable_mixed, test/test_autograd.py::TestAutograd::test_mark_non_differentiable_none, test/test_autograd.py::TestAutograd::test_materialize_grads, test/test_autograd.py::TestAutograd::test_multi_backward, test/test_autograd.py::TestAutograd::test_multi_backward_no_grad, test/test_autograd.py::TestAutograd::test_multi_grad_hooks, test/test_autograd.py::TestAutograd::test_named_tensor_for_complex_views, test/test_autograd.py::TestAutograd::test_naughty_anomaly_access, test/test_autograd.py::TestAutograd::test_naughty_autograd_function_attribute_access, test/test_autograd.py::TestAutograd::test_naughty_autograd_function_stashing_ctx, test/test_autograd.py::TestAutograd::test_nested_anomaly_detect_nan, test/test_autograd.py::TestAutograd::test_nested_anomaly_printstack_cleanup, test/test_autograd.py::TestAutograd::test_next_functions, test/test_autograd.py::TestAutograd::test_no_grad, test/test_autograd.py::TestAutograd::test_no_grad_assignment, test/test_autograd.py::TestAutograd::test_no_grad_copy, test/test_autograd.py::TestAutograd::test_no_grad_copy_sparse, test/test_autograd.py::TestAutograd::test_no_grad_input, test/test_autograd.py::TestAutograd::test_no_grad_modifies_version, test/test_autograd.py::TestAutograd::test_no_grad_python_function, test/test_autograd.py::TestAutograd::test_no_requires_grad_inplace, test/test_autograd.py::TestAutograd::test_no_unnecessary_save, test/test_autograd.py::TestAutograd::test_no_unnecessary_unwrapping, test/test_autograd.py::TestAutograd::test_not_implemented_fwad, test/test_autograd.py::TestAutograd::test_not_implemented_grad, test/test_autograd.py::TestAutograd::test_numpy_requires_grad, test/test_autograd.py::TestAutograd::test_once_differentiable, test/test_autograd.py::TestAutograd::test_out_variant_raises_when_inputs_require_grad, test/test_autograd.py::TestAutograd::test_pack_hook_with_inplace_modification_should_fail, test/test_autograd.py::TestAutograd::test_pickle, test/test_autograd.py::TestAutograd::test_pow_zero_tensor_gradient, test/test_autograd.py::TestAutograd::test_power_function, test/test_autograd.py::TestAutograd::test_prehook_ordering, test/test_autograd.py::TestAutograd::test_profiler, test/test_autograd.py::TestAutograd::test_profiler_aggregation_fake, test/test_autograd.py::TestAutograd::test_profiler_aggregation_lstm, test/test_autograd.py::TestAutograd::test_profiler_aggregation_table, test/test_autograd.py::TestAutograd::test_profiler_function_event_avg, test/test_autograd.py::TestAutograd::test_profiler_propagation, test/test_autograd.py::TestAutograd::test_profiler_seq_nr, test/test_autograd.py::TestAutograd::test_profiler_shapes, test/test_autograd.py::TestAutograd::test_profiler_unboxed_only, test/test_autograd.py::TestAutograd::test_pynode_destruction_deadlock, test/test_autograd.py::TestAutograd::test_record_function, test/test_autograd.py::TestAutograd::test_record_function_callbacks, test/test_autograd.py::TestAutograd::test_record_function_legacy, test/test_autograd.py::TestAutograd::test_record_function_multithreaded, test/test_autograd.py::TestAutograd::test_reentrant_child_error, test/test_autograd.py::TestAutograd::test_reentrant_priority, test/test_autograd.py::TestAutograd::test_reentrant_with_callbacks_both_depths, test/test_autograd.py::TestAutograd::test_reentrant_with_callbacks_depth_0, test/test_autograd.py::TestAutograd::test_reentrant_with_callbacks_depth_1, test/test_autograd.py::TestAutograd::test_reentrant_with_leaf_variable_hook, test/test_autograd.py::TestAutograd::test_reentrant_with_non_leaf_variable_hook, test/test_autograd.py::TestAutograd::test_requires_grad, test/test_autograd.py::TestAutograd::test_requires_grad_, test/test_autograd.py::TestAutograd::test_requires_grad_inplace, test/test_autograd.py::TestAutograd::test_retain_grad, test/test_autograd.py::TestAutograd::test_retain_grad_cycle, test/test_autograd.py::TestAutograd::test_retain_grad_inplace, test/test_autograd.py::TestAutograd::test_retain_grad_inplace_over_view, test/test_autograd.py::TestAutograd::test_retains_grad_can_always_observe_tensor_prehook, test/test_autograd.py::TestAutograd::test_retains_grad_inplace_multiple_outputs, test/test_autograd.py::TestAutograd::test_return_duplicate, test/test_autograd.py::TestAutograd::test_return_duplicate_inplace, test/test_autograd.py::TestAutograd::test_return_leaf, test/test_autograd.py::TestAutograd::test_return_leaf_inplace, test/test_autograd.py::TestAutograd::test_save_none_for_backward, test/test_autograd.py::TestAutograd::test_save_on_cpu_and_checkpoint, test/test_autograd.py::TestAutograd::test_save_output_nr, test/test_autograd.py::TestAutograd::test_saved_tensor_hooks_custom_error_propagaation, test/test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_default_hooks, test/test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_hooks, test/test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_saved_original_with_default_hooks, test/test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_saved_original_with_hooks, test/test_autograd.py::TestAutograd::test_saved_variable_saved_original_inplace_detach, test/test_autograd.py::TestAutograd::test_saved_variable_version_counter, test/test_autograd.py::TestAutograd::test_saved_variables_deprecated, test/test_autograd.py::TestAutograd::test_saving_variable_to_disk, test/test_autograd.py::TestAutograd::test_select_expanded_v, test/test_autograd.py::TestAutograd::test_select_sum, test/test_autograd.py::TestAutograd::test_set_data_preserve_pyobj, test/test_autograd.py::TestAutograd::test_set_data_self_requires_grad, test/test_autograd.py::TestAutograd::test_set_data_tensorimpl_type, test/test_autograd.py::TestAutograd::test_set_grad_coroutines, test/test_autograd.py::TestAutograd::test_set_grad_coroutines_benign_exceptions, test/test_autograd.py::TestAutograd::test_set_grad_coroutines_critical_exceptions, test/test_autograd.py::TestAutograd::test_set_grad_coroutines_exit, test/test_autograd.py::TestAutograd::test_set_grad_enabled, test/test_autograd.py::TestAutograd::test_set_grad_generator_functions, test/test_autograd.py::TestAutograd::test_set_grad_generator_functions_recursive, test/test_autograd.py::TestAutograd::test_setitem, test/test_autograd.py::TestAutograd::test_setitem_mask, test/test_autograd.py::TestAutograd::test_setting_default_saved_variable_hooks_twice_should_not_fail, test/test_autograd.py::TestAutograd::test_setting_default_saved_variable_hooks_twice_should_use_inner, test/test_autograd.py::TestAutograd::test_shape, test/test_autograd.py::TestAutograd::test_sharded_grad, test/test_autograd.py::TestAutograd::test_simple_reentrant, test/test_autograd.py::TestAutograd::test_slice_expanded_v, test/test_autograd.py::TestAutograd::test_sparse_gather_both_scalar, test/test_autograd.py::TestAutograd::test_sparse_gather_dim0, test/test_autograd.py::TestAutograd::test_sparse_gather_dim1, test/test_autograd.py::TestAutograd::test_sparse_gather_dim_neg, test/test_autograd.py::TestAutograd::test_sparse_gather_ind_scalar, test/test_autograd.py::TestAutograd::test_sparse_gather_x_scalar, test/test_autograd.py::TestAutograd::test_sparse_mm_backward, test/test_autograd.py::TestAutograd::test_tensor_grad_warnings, test/test_autograd.py::TestAutograd::test_tensor_hooks_inplace, test/test_autograd.py::TestAutograd::test_tensor_hooks_inplace_multiple_outputs, test/test_autograd.py::TestAutograd::test_tensor_hooks_inplace_over_view, test/test_autograd.py::TestAutograd::test_thread_shutdown, test/test_autograd.py::TestAutograd::test_to_sparse_backward, test/test_autograd.py::TestAutograd::test_too_many_grads, test/test_autograd.py::TestAutograd::test_type_conversions, test/test_autograd.py::TestAutograd::test_unrelated_inputs, test/test_autograd.py::TestAutograd::test_unsafe_set_version_counter, test/test_autograd.py::TestAutograd::test_unused_output, test/test_autograd.py::TestAutograd::test_var_mean_differentiable, test/test_autograd.py::TestAutograd::test_variable_traverse, test/test_autograd.py::TestAutograd::test_version_counter, test/test_autograd.py::TestAutograd::test_view_func_replay, test/test_autograd.py::TestAutograd::test_view_replay_enabled, test/test_autograd.py::TestAutograd::test_volatile_deprecated, test/test_autograd.py::TestAutograd::test_will_engine_execute_node, test/test_autograd.py::TestAutograd::test_wrapped_number_saved_variable_hooks, test/test_autograd.py::TestAutogradForwardModeBatchedGrad::test_inplace_on_view_not_same_layout, test/test_autograd.py::TestAutogradForwardModeBatchedGrad::test_inplace_on_view_same_layout, test/test_autograd.py::TestAutogradForwardModeBatchedGrad::test_metadata_check_for_storage_numel_skipped, test/test_autograd.py::TestAutogradForwardModeBatchedGrad::test_out_of_place_basic, test/test_autograd.py::TestAutogradForwardModeBatchedGrad::test_out_of_place_not_same_layout, test/test_autograd.py::TestAutogradForwardMode::test_advanced_packing_unpacking, test/test_autograd.py::TestAutogradForwardMode::test_backward_graph_destruction, test/test_autograd.py::TestAutogradForwardMode::test_basic_packing_unpacking, test/test_autograd.py::TestAutogradForwardMode::test_codegen_ignores_undefined_outputs, test/test_autograd.py::TestAutogradForwardMode::test_create_new_zeros_with_same_meta, test/test_autograd.py::TestAutogradForwardMode::test_default_level, test/test_autograd.py::TestAutogradForwardMode::test_detach_view_tracking, test/test_autograd.py::TestAutogradForwardMode::test_forward_level_cleanup, test/test_autograd.py::TestAutogradForwardMode::test_fwd_grad_enabled, test/test_autograd.py::TestAutogradForwardMode::test_grad_cleanup, test/test_autograd.py::TestAutogradForwardMode::test_make_dual_forbid_integral_dtype, test/test_autograd.py::TestAutogradForwardMode::test_make_dual_inference_tensor_in_inference_mode, test/test_autograd.py::TestAutogradForwardMode::test_make_dual_torch_dispatch, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_check_conj, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_checks_ignores_size_zero, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_checks_storage_numel, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_ignore_storage_offset_for_zero_numel_tensor, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_when_primal_has_conj_bit, test/test_autograd.py::TestAutogradForwardMode::test_metadata_check_when_primal_has_neg_bit, test/test_autograd.py::TestAutogradForwardMode::test_nested_level, test/test_autograd.py::TestAutogradForwardMode::test_non_differentiable, test/test_autograd.py::TestAutogradForwardMode::test_out_variant, test/test_autograd.py::TestAutogradForwardMode::test_print, test/test_autograd.py::TestAutogradForwardMode::test_set_fw_grad_having_own_fw_grad_at_same_level, test/test_autograd.py::TestAutogradForwardMode::test_set_fwd_grad_enabled, test/test_autograd.py::TestAutogradForwardMode::test_size_check, test/test_autograd.py::TestAutogradForwardMode::test_view_inplace_always_creates_a_view, test/test_autograd.py::TestAutogradForwardMode::test_view_inplace_differentiable_views, test/test_autograd.py::TestAutogradForwardMode::test_view_inplace_non_differentiable_views, test/test_autograd.py::TestAllowMutationOnSaved::test_backward_out_of_context, test/test_autograd.py::TestAllowMutationOnSaved::test_basic, test/test_autograd.py::TestAllowMutationOnSaved::test_disallow_nesting, test/test_autograd.py::TestAllowMutationOnSaved::test_double_backward, test/test_autograd.py::TestAllowMutationOnSaved::test_save_base_and_modify_view, test/test_autograd.py::TestAllowMutationOnSaved::test_save_view_modify_base, test/test_autograd.py::TestAllowMutationOnSaved::test_saved_but_not_anymore, test/test_autograd.py::TestAllowMutationOnSaved::test_saved_same_tensor_different_versions, test/test_autograd.py::TestAllowMutationOnSaved::test_saved_same_tensor_many_times, test/test_autograd.py::TestAllowMutationOnSaved::test_views, test/test_autograd.py::TestAllowMutationOnSaved::test_with_math_views, test/test_autograd.py::TestAllowMutationOnSaved::test_with_out_variant, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_context_manager, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_decorator, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_existing_autograd_session, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_handle_direct_view_on_rebase, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_handle_indirect_view_on_rebase, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_functional_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_inplace_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_view_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_functional_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_inplace_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_view_op, test/test_autograd.py::TestAutogradInferenceMode::test_inference_mode_tensor_creation, test/test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_functional_op, test/test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_inplace_op, test/test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_view_op, test/test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_inplace_output_in_inference_mode, test/test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_inplace_output_in_normal_mode, test/test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_view_output_in_inference_mode, test/test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_view_output_in_normal_mode, test/test_autograd.py::TestMultithreadAutograd::test_cat_stack_r_to_c, test/test_autograd.py::TestMultithreadAutograd::test_dataparallel_saved_tensors_hooks, test/test_autograd.py::TestMultithreadAutograd::test_fork_join_in_middle, test/test_autograd.py::TestMultithreadAutograd::test_multi_grad_hooks, test/test_autograd.py::TestMultithreadAutograd::test_multithreaded_exception_propagation, test/test_autograd.py::TestMultithreadAutograd::test_preserve_backtrace, test/test_autograd.py::TestMultithreadAutograd::test_python_thread_in_middle, test/test_autograd.py::TestMultithreadAutograd::test_simple_backward, test/test_autograd.py::TestMultithreadAutograd::test_simple_backward_same_input, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_early_stop_True, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_kwargs_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_kwargs_early_stop_True, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_True, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_reentrant_backwards_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_reentrant_backwards_early_stop_True, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_same_graph_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_same_graph_early_stop_True, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_set_early_stop, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_set_early_stop_no_recompution_needed, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_two_children_early_stop_False, test/test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_two_children_early_stop_True, test/test_autograd.py::TestAutogradComplex::test_view_func_for_complex_views, test/test_autograd.py::TestAutogradComplex::test_view_with_multi_output, test/test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_cuda_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_cuda_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_False_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_False_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_True_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_True_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_vectorize_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_vectorize_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_False_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_False_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_True_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_True_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_match_vhp_hvp_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_match_vhp_hvp_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_output_vectorized_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_output_vectorized_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_False_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_False_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_True_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_True_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_multi_input_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_multi_input_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_simple_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_simple_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_unrelated_outputs_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_unrelated_outputs_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_raises_no_warnings_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_raises_no_warnings_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_create_graph_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_create_graph_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_err_check_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_err_check_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_scalar_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_hvp_scalar_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_False_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_False_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_True_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_True_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_vectorize_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_vectorize_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_False_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_False_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_True_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_True_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_match_vjp_jvp_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_match_vjp_jvp_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_output_vectorized_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_output_vectorized_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_vectorized_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_vectorized_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_devices_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_devices_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_dtype_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_dtype_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_multi_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_multi_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_simple_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_simple_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_unrelated_outputs_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_unrelated_outputs_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_zero_dim_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_zero_dim_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_raises_no_warnings_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_raises_no_warnings_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_create_graph_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_create_graph_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_err_check_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_err_check_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_scalar_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_jvp_scalar_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_create_graph_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_create_graph_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_err_check_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_err_check_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_scalar_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vhp_scalar_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_create_graph_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_create_graph_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_err_check_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_err_check_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_err_check_strict_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_err_check_strict_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_no_grad_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_no_grad_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_output_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_output_logging_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_scalar_base_tensor, test/test_autograd.py::TestAutogradFunctional::test_vjp_scalar_logging_tensor, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_advanced_indexing_backwards_large_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_advanced_indexing_backwards_memory_format_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_backward_device_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_complex_scalar_backward_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_copy__cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_forward_ad_broadcasting_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_forward_ad_same_layout_copies_grad_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_r_to_c_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_cross_device_reentrant_autograd_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_free_unneeded_tensor_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_grad_assignment_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_gradcheck_input_output_different_device_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_multiple_output_view_of_view_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_base_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_view_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_view_of_view_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_gradcheck_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_makes_base_require_grad_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_modify_base_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multi_output_safe_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multi_output_unsafe_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multiple_outputs_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_non_contig_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_of_multiple_output_view_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_of_view_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_python_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_then_no_grad_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_inputbuffer_add_multidevice_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_min_max_median_backprops_to_all_values_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_mv_grad_stride_0_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_non_differentiable_ops_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_parameter_resize_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_pin_memory_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_pow_real_negative_base_complex_exponent_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_profiler_emit_itt_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_profiler_emit_nvtx_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_pyscalar_conversions_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_reentrant_parent_error_on_cpu_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_requires_grad_factory_cpu_float32, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_requires_grad_factory_cpu_float64, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_resize_version_bump_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_rnn_backward_to_input_but_not_parameters_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_scatter_index_reduce_amin_amax_backprops_to_all_values_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_scatter_index_reduce_prod_gradgrad_error_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_float32, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_float64, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int16, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int32, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int64, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int8, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_simple_reentrant_cross_device_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_backward_cpu_complex128, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_backward_cpu_float64, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_ctor_getter_backward_cpu_complex128, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_ctor_getter_backward_cpu_float64, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_mask_autograd_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_strided_leaf_grad_layout_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_to_r_to_c_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_unused_output_device_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_warning_in_backward_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_where_functional_cpu, test/test_autograd.py::TestAutogradDeviceTypeCPU::test_where_scalar_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_autograd_composite_implicit_and_dispatch_registration_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_autograd_multiple_dispatch_registrations_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_backward_single_threaded_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_backward_tls_stash_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_foward_mode_AD_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_per_dispatch_key_input_saving_cpu, test/test_autograd.py::TestAutogradMultipleDispatchCPU::test_view_copy_cpu 2023-03-31T06:24:33.6328957Z 2023-03-31T06:24:33.6329564Z test_autograd.py::TestAutograd::test_access_saved_tensor_twice_without_recomputation_works PASSED [ 0%] 2023-03-31T06:24:33.6329953Z test_autograd.py::TestAutograd::test_accumulate_grad PASSED [ 0%] 2023-03-31T06:24:33.6330316Z test_autograd.py::TestAutograd::test_accumulate_grad_posthooks_can_observe_tensor_prehook PASSED [ 0%] 2023-03-31T06:24:33.6330690Z test_autograd.py::TestAutograd::test_accumulate_grad_tensor_reference PASSED [ 0%] 2023-03-31T06:24:33.6331045Z test_autograd.py::TestAutograd::test_accumulate_grad_with_zero_numel_grad PASSED [ 0%] 2023-03-31T06:24:33.6331394Z test_autograd.py::TestAutograd::test_anomaly_assign_parent_cleanup PASSED [ 1%] 2023-03-31T06:24:33.6331837Z test_autograd.py::TestAutograd::test_anomaly_detect_nan PASSED [ 1%] 2023-03-31T06:24:33.6332203Z test_autograd.py::TestAutograd::test_anomaly_grad_warnings PASSED [ 1%] 2023-03-31T06:24:33.6332528Z test_autograd.py::TestAutograd::test_anomaly_mode_no_check_nan PASSED [ 1%] 2023-03-31T06:24:33.6332839Z test_autograd.py::TestAutograd::test_attribute_deletion PASSED [ 1%] 2023-03-31T06:24:33.6333178Z test_autograd.py::TestAutograd::test_autograd_inplace_view_of_view PASSED [ 1%] 2023-03-31T06:24:33.6333533Z test_autograd.py::TestAutograd::test_autograd_inplace_views_creation_meta PASSED [ 2%] 2023-03-31T06:24:33.6333886Z test_autograd.py::TestAutograd::test_autograd_inplace_views_cross_dtype PASSED [ 2%] 2023-03-31T06:24:33.6334227Z test_autograd.py::TestAutograd::test_autograd_multiple_views_python PASSED [ 2%] 2023-03-31T06:24:33.6334563Z test_autograd.py::TestAutograd::test_autograd_node_isinstance PASSED [ 2%] 2023-03-31T06:24:33.6334914Z test_autograd.py::TestAutograd::test_autograd_python_custom_function_inplace PASSED [ 2%] 2023-03-31T06:24:33.6335254Z test_autograd.py::TestAutograd::test_autograd_simple_views_python PASSED [ 2%] 2023-03-31T06:24:33.6335586Z test_autograd.py::TestAutograd::test_autograd_views_codegen PASSED [ 3%] 2023-03-31T06:24:33.6335899Z test_autograd.py::TestAutograd::test_backward PASSED [ 3%] 2023-03-31T06:24:33.6336215Z test_autograd.py::TestAutograd::test_backward_badcalls PASSED [ 3%] 2023-03-31T06:24:33.6336514Z test_autograd.py::TestAutograd::test_backward_copy PASSED [ 3%] 2023-03-31T06:24:33.6336874Z test_autograd.py::TestAutograd::test_backward_create_graph_warns PASSED [ 3%] 2023-03-31T06:24:33.6337194Z test_autograd.py::TestAutograd::test_backward_no_grad PASSED [ 4%] 2023-03-31T06:24:33.6337530Z test_autograd.py::TestAutograd::test_backward_twice_retained_graph_with_saved_values PASSED [ 4%] 2023-03-31T06:24:33.6337916Z test_autograd.py::TestAutograd::test_backward_twice_retained_graph_without_saved_values PASSED [ 4%] 2023-03-31T06:24:33.6338283Z test_autograd.py::TestAutograd::test_backward_twice_with_saved_values PASSED [ 4%] 2023-03-31T06:24:33.6338633Z test_autograd.py::TestAutograd::test_backward_twice_without_saved_values PASSED [ 4%] 2023-03-31T06:24:33.6338958Z test_autograd.py::TestAutograd::test_backward_with_inputs PASSED [ 4%] 2023-03-31T06:24:33.6339342Z test_autograd.py::TestAutograd::test_backward_with_nonleaf_inputs PASSED [ 5%] 2023-03-31T06:24:33.6339675Z test_autograd.py::TestAutograd::test_calculate_shape_util PASSED [ 5%] 2023-03-31T06:24:33.6339997Z test_autograd.py::TestAutograd::test_callback_adds_callback PASSED [ 5%] 2023-03-31T06:24:33.6340309Z test_autograd.py::TestAutograd::test_cant_create_saved_tensors PASSED [ 5%] 2023-03-31T06:24:33.6340647Z test_autograd.py::TestAutograd::test_checkpoint_valid_reset_on_error PASSED [ 5%] 2023-03-31T06:24:33.6341051Z test_autograd.py::TestAutograd::test_checkpointing SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 5%] 2023-03-31T06:24:33.6341450Z test_autograd.py::TestAutograd::test_checkpointing_non_reentrant_autocast_cpu PASSED [ 6%] 2023-03-31T06:24:33.6341858Z test_autograd.py::TestAutograd::test_checkpointing_non_reentrant_autocast_gpu SKIPPED (Test requires CUDA bf16 support) [ 6%] 2023-03-31T06:24:33.6342278Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_arbitrary_input_output PASSED [ 6%] 2023-03-31T06:24:33.6342674Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_correct_grad PASSED [ 6%] 2023-03-31T06:24:33.6343124Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_custom_function_works PASSED [ 6%] 2023-03-31T06:24:33.6343513Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_dataparallel PASSED [ 6%] 2023-03-31T06:24:33.6343966Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_False PASSED [ 7%] 2023-03-31T06:24:33.6344428Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_True PASSED [ 7%] 2023-03-31T06:24:33.6344904Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_input_requires_grad_False SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 7%] 2023-03-31T06:24:33.6345436Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_input_requires_grad_True SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 7%] 2023-03-31T06:24:33.6345916Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_memory_savings SKIPPED (Test requires CUDA) [ 7%] 2023-03-31T06:24:33.6346336Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_parameter_used_in_an_out PASSED [ 8%] 2023-03-31T06:24:33.6346720Z test_autograd.py::TestAutograd::test_checkpointing_without_reentrant_with_context_fn PASSED [ 8%] 2023-03-31T06:24:33.6347089Z test_autograd.py::TestAutograd::test_copy_slices_graph_task_updates PASSED [ 8%] 2023-03-31T06:24:33.6347450Z test_autograd.py::TestAutograd::test_create_graph_and_full_backward_hook_cycle PASSED [ 8%] 2023-03-31T06:24:33.6347799Z test_autograd.py::TestAutograd::test_current_graph_task_execution_order PASSED [ 8%] 2023-03-31T06:24:33.6348144Z test_autograd.py::TestAutograd::test_current_graph_task_id PASSED [ 8%] 2023-03-31T06:24:33.6348465Z test_autograd.py::TestAutograd::test_current_node PASSED [ 9%] 2023-03-31T06:24:33.6348828Z test_autograd.py::TestAutograd::test_custom_autograd_no_early_free PASSED [ 9%] 2023-03-31T06:24:33.6349163Z test_autograd.py::TestAutograd::test_custom_autograd_repeated_grad_grad PASSED [ 9%] 2023-03-31T06:24:33.6349502Z test_autograd.py::TestAutograd::test_custom_function_cycle PASSED [ 9%] 2023-03-31T06:24:33.6349826Z test_autograd.py::TestAutograd::test_custom_function_error PASSED [ 9%] 2023-03-31T06:24:33.6350150Z test_autograd.py::TestAutograd::test_custom_function_exception PASSED [ 9%] 2023-03-31T06:24:33.6350511Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_forward_is_no_op PASSED [ 10%] 2023-03-31T06:24:33.6350886Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_inplace_checks PASSED [ 10%] 2023-03-31T06:24:33.6351268Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_non_differentiable PASSED [ 10%] 2023-03-31T06:24:33.6351687Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_non_tensor_before_tensor_args PASSED [ 10%] 2023-03-31T06:24:33.6352074Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_view_checks PASSED [ 10%] 2023-03-31T06:24:33.6352445Z test_autograd.py::TestAutograd::test_custom_function_forward_mode_wrong_formula PASSED [ 10%] 2023-03-31T06:24:33.6352802Z test_autograd.py::TestAutograd::test_custom_function_local_inplace PASSED [ 11%] 2023-03-31T06:24:33.6353156Z test_autograd.py::TestAutograd::test_custom_function_mark_dirty_not_differentiable PASSED [ 11%] 2023-03-31T06:24:33.6353518Z test_autograd.py::TestAutograd::test_custom_function_no_tensors PASSED [ 11%] 2023-03-31T06:24:33.6353874Z test_autograd.py::TestAutograd::test_custom_function_non_tensor_inputs_outputs PASSED [ 11%] 2023-03-31T06:24:33.6354222Z test_autograd.py::TestAutograd::test_custom_function_return_view_in_nograd PASSED [ 11%] 2023-03-31T06:24:33.6354577Z test_autograd.py::TestAutograd::test_custom_function_save_for_forward PASSED [ 12%] 2023-03-31T06:24:33.6354927Z test_autograd.py::TestAutograd::test_custom_function_saved_tensors PASSED [ 12%] 2023-03-31T06:24:33.6355287Z test_autograd.py::TestAutograd::test_custom_function_setup_context_multi_input PASSED [ 12%] 2023-03-31T06:24:33.6355646Z test_autograd.py::TestAutograd::test_custom_function_setup_context_multi_output PASSED [ 12%] 2023-03-31T06:24:33.6356011Z test_autograd.py::TestAutograd::test_custom_function_setup_context_simple PASSED [ 12%] 2023-03-31T06:24:33.6356399Z test_autograd.py::TestAutograd::test_custom_function_vmap_defaults PASSED [ 12%] 2023-03-31T06:24:33.6356765Z test_autograd.py::TestAutograd::test_deep_reentrant PASSED [ 13%] 2023-03-31T06:24:33.6357119Z test_autograd.py::TestAutograd::test_default_saved_variable_hooks_double_backward PASSED [ 13%] 2023-03-31T06:24:33.6357463Z test_autograd.py::TestAutograd::test_dep_nograd PASSED [ 13%] 2023-03-31T06:24:33.6357785Z test_autograd.py::TestAutograd::test_dependent_backward PASSED [ 13%] 2023-03-31T06:24:33.6358088Z test_autograd.py::TestAutograd::test_detach PASSED [ 13%] 2023-03-31T06:24:33.6358399Z test_autograd.py::TestAutograd::test_detach_base PASSED [ 13%] 2023-03-31T06:24:33.6358741Z test_autograd.py::TestAutograd::test_detach_then_inplace_raises_in_autograd PASSED [ 14%] 2023-03-31T06:24:33.6359072Z test_autograd.py::TestAutograd::test_diagonal_expanded_v PASSED [ 14%] 2023-03-31T06:24:33.6359385Z test_autograd.py::TestAutograd::test_dir PASSED [ 14%] 2023-03-31T06:24:33.6359702Z test_autograd.py::TestAutograd::test_disabling_saved_tensor_hooks PASSED [ 14%] 2023-03-31T06:24:33.6360049Z test_autograd.py::TestAutograd::test_disabling_saved_tensor_hooks_nested PASSED [ 14%] 2023-03-31T06:24:33.6360381Z test_autograd.py::TestAutograd::test_dont_materialize_grads PASSED [ 14%] 2023-03-31T06:24:33.6360706Z test_autograd.py::TestAutograd::test_duplicate_backward_root PASSED [ 15%] 2023-03-31T06:24:33.6361045Z test_autograd.py::TestAutograd::test_first_grad_fn_access_in_no_grad_mode PASSED [ 15%] 2023-03-31T06:24:33.6361402Z test_autograd.py::TestAutograd::test_free_deep_graph PASSED [ 15%] 2023-03-31T06:24:33.6361731Z test_autograd.py::TestAutograd::test_free_deep_graph_complicated PASSED [ 15%] 2023-03-31T06:24:33.6362060Z test_autograd.py::TestAutograd::test_free_deep_graph_pyfunction PASSED [ 15%] 2023-03-31T06:24:33.6362400Z test_autograd.py::TestAutograd::test_full_backward_hook_double_backward PASSED [ 16%] 2023-03-31T06:24:33.6362718Z test_autograd.py::TestAutograd::test_function PASSED [ 16%] 2023-03-31T06:24:33.6363211Z test_autograd.py::TestAutograd::test_function_returns_input PASSED [ 16%] 2023-03-31T06:24:33.6363568Z test_autograd.py::TestAutograd::test_function_returns_undefined_tensor PASSED [ 16%] 2023-03-31T06:24:33.6363894Z test_autograd.py::TestAutograd::test_gc_in_destructor PASSED [ 16%] 2023-03-31T06:24:33.6364282Z test_autograd.py::TestAutograd::test_grad PASSED [ 16%] 2023-03-31T06:24:33.6364583Z test_autograd.py::TestAutograd::test_grad_badcalls PASSED [ 17%] 2023-03-31T06:24:33.6364901Z test_autograd.py::TestAutograd::test_grad_batched_grad PASSED [ 17%] 2023-03-31T06:24:33.6365204Z test_autograd.py::TestAutograd::test_grad_empty_inputs PASSED [ 17%] 2023-03-31T06:24:33.6365520Z test_autograd.py::TestAutograd::test_grad_fn_attr_bindings PASSED [ 17%] 2023-03-31T06:24:33.6365837Z test_autograd.py::TestAutograd::test_grad_fn_badcalls PASSED [ 17%] 2023-03-31T06:24:33.6366140Z test_autograd.py::TestAutograd::test_grad_fn_prehooks PASSED [ 17%] 2023-03-31T06:24:33.6366478Z test_autograd.py::TestAutograd::test_grad_fn_prehooks_multiple_outputs PASSED [ 18%] 2023-03-31T06:24:33.6366826Z test_autograd.py::TestAutograd::test_grad_fn_prehooks_remove_hooks PASSED [ 18%] 2023-03-31T06:24:33.6367168Z test_autograd.py::TestAutograd::test_grad_materialize_grads PASSED [ 18%] 2023-03-31T06:24:33.6367481Z test_autograd.py::TestAutograd::test_grad_mode_class_decoration PASSED [ 18%] 2023-03-31T06:24:33.6367806Z test_autograd.py::TestAutograd::test_grad_mode_restored_reentrant PASSED [ 18%] 2023-03-31T06:24:33.6368127Z test_autograd.py::TestAutograd::test_grad_nonleaf PASSED [ 18%] 2023-03-31T06:24:33.6368429Z test_autograd.py::TestAutograd::test_grad_nonleaf_many_outputs PASSED [ 19%] 2023-03-31T06:24:33.6368794Z test_autograd.py::TestAutograd::test_grad_nonleaf_register_hook PASSED [ 19%] 2023-03-31T06:24:33.6369144Z test_autograd.py::TestAutograd::test_grad_unreachable PASSED [ 19%] 2023-03-31T06:24:33.6369463Z test_autograd.py::TestAutograd::test_grad_unreachable_discovery PASSED [ 19%] 2023-03-31T06:24:33.6369800Z test_autograd.py::TestAutograd::test_gradcheck_backward_mul_by_grad_output PASSED [ 19%] 2023-03-31T06:24:33.6370148Z test_autograd.py::TestAutograd::test_gradcheck_check_batched_grad PASSED [ 20%] 2023-03-31T06:24:33.6370500Z test_autograd.py::TestAutograd::test_gradcheck_check_forward_or_backward_only PASSED [ 20%] 2023-03-31T06:24:33.6370861Z test_autograd.py::TestAutograd::test_gradcheck_check_no_differentiable_outputs PASSED [ 20%] 2023-03-31T06:24:33.6371230Z test_autograd.py::TestAutograd::test_gradcheck_complex_non_complex_outputs PASSED [ 20%] 2023-03-31T06:24:33.6371576Z test_autograd.py::TestAutograd::test_gradcheck_custom_error PASSED [ 20%] 2023-03-31T06:24:33.6371924Z test_autograd.py::TestAutograd::test_gradcheck_dense_and_sparse_inputs PASSED [ 20%] 2023-03-31T06:24:33.6372254Z test_autograd.py::TestAutograd::test_gradcheck_forward_ad PASSED [ 21%] 2023-03-31T06:24:33.6372593Z test_autograd.py::TestAutograd::test_gradcheck_forward_ad_batched_grad PASSED [ 21%] 2023-03-31T06:24:33.6372955Z test_autograd.py::TestAutograd::test_gradcheck_forward_ad_respects_requires_grad PASSED [ 21%] 2023-03-31T06:24:33.6373319Z test_autograd.py::TestAutograd::test_gradcheck_forward_ad_runs_with_no_requires_grad PASSED [ 21%] 2023-03-31T06:24:33.6373735Z test_autograd.py::TestAutograd::test_gradcheck_get_analytical_jacobian PASSED [ 21%] 2023-03-31T06:24:33.6374084Z test_autograd.py::TestAutograd::test_gradcheck_get_numerical_jacobian PASSED [ 21%] 2023-03-31T06:24:33.6374440Z test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_bsc PASSED [ 22%] 2023-03-31T06:24:33.6374786Z test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_bsr PASSED [ 22%] 2023-03-31T06:24:33.6375149Z test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_coo PASSED [ 22%] 2023-03-31T06:24:33.6375503Z test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_csc PASSED [ 22%] 2023-03-31T06:24:33.6375853Z test_autograd.py::TestAutograd::test_gradcheck_input_layout_torch_sparse_csr PASSED [ 22%] 2023-03-31T06:24:33.6376184Z test_autograd.py::TestAutograd::test_gradcheck_jacobian_mismatch PASSED [ 22%] 2023-03-31T06:24:33.6376559Z test_autograd.py::TestAutograd::test_gradcheck_multiple_mkldnn_inputs PASSED [ 23%] 2023-03-31T06:24:33.6376910Z test_autograd.py::TestAutograd::test_gradcheck_nondeterministic PASSED [ 23%] 2023-03-31T06:24:33.6377264Z test_autograd.py::TestAutograd::test_gradcheck_output_shape_or_dtype_depend_on_values PASSED [ 23%] 2023-03-31T06:24:33.6377625Z test_autograd.py::TestAutograd::test_gradcheck_single_input PASSED [ 23%] 2023-03-31T06:24:33.6377952Z test_autograd.py::TestAutograd::test_gradcheck_test_outputs PASSED [ 23%] 2023-03-31T06:24:33.6378281Z test_autograd.py::TestAutograd::test_gradcheck_undefined_grad PASSED [ 24%] 2023-03-31T06:24:33.6378614Z test_autograd.py::TestAutograd::test_gradcheck_validates_input_mkldnn PASSED [ 24%] 2023-03-31T06:24:33.6378957Z test_autograd.py::TestAutograd::test_gradcheck_validates_inputs PASSED [ 24%] 2023-03-31T06:24:33.6379283Z test_autograd.py::TestAutograd::test_graph_save_on_cpu PASSED [ 24%] 2023-03-31T06:24:33.6379616Z test_autograd.py::TestAutograd::test_graph_save_on_cpu_cuda SKIPPED (test requires CUDA) [ 24%] 2023-03-31T06:24:33.6379961Z test_autograd.py::TestAutograd::test_hessian_vector PASSED [ 24%] 2023-03-31T06:24:33.6380293Z test_autograd.py::TestAutograd::test_hook_edge_case_when_called_with_grad PASSED [ 25%] 2023-03-31T06:24:33.6380624Z test_autograd.py::TestAutograd::test_hook_none PASSED [ 25%] 2023-03-31T06:24:33.6380917Z test_autograd.py::TestAutograd::test_hook_with_no_name PASSED [ 25%] 2023-03-31T06:24:33.6381257Z test_autograd.py::TestAutograd::test_hooks PASSED [ 25%] 2023-03-31T06:24:33.6381598Z test_autograd.py::TestAutograd::test_hooks_cpp PASSED [ 25%] 2023-03-31T06:24:33.6381895Z test_autograd.py::TestAutograd::test_increment_version PASSED [ 25%] 2023-03-31T06:24:33.6382225Z test_autograd.py::TestAutograd::test_index_backward_does_not_save_tensor PASSED [ 26%] 2023-03-31T06:24:33.6382556Z test_autograd.py::TestAutograd::test_indexing PASSED [ 26%] 2023-03-31T06:24:33.6382940Z test_autograd.py::TestAutograd::test_indexing_duplicates PASSED [ 26%] 2023-03-31T06:24:33.6383247Z test_autograd.py::TestAutograd::test_inplace PASSED [ 26%] 2023-03-31T06:24:33.6383563Z test_autograd.py::TestAutograd::test_inplace_not_requires_grad PASSED [ 26%] 2023-03-31T06:24:33.6383889Z test_autograd.py::TestAutograd::test_inplace_on_view_backward PASSED [ 27%] 2023-03-31T06:24:33.6384202Z test_autograd.py::TestAutograd::test_inplace_on_view_leaf_errors PASSED [ 27%] 2023-03-31T06:24:33.6384534Z test_autograd.py::TestAutograd::test_inplace_on_view_saved_output PASSED [ 27%] 2023-03-31T06:24:33.6384863Z test_autograd.py::TestAutograd::test_inplace_on_view_weak_grad_fn PASSED [ 27%] 2023-03-31T06:24:33.6385180Z test_autograd.py::TestAutograd::test_input_buffer_accum PASSED [ 27%] 2023-03-31T06:24:33.6385480Z test_autograd.py::TestAutograd::test_integer_outputs PASSED [ 27%] 2023-03-31T06:24:33.6385792Z test_autograd.py::TestAutograd::test_invalid_gradients PASSED [ 28%] 2023-03-31T06:24:33.6386159Z test_autograd.py::TestAutograd::test_isolated_node PASSED [ 28%] 2023-03-31T06:24:33.6386461Z test_autograd.py::TestAutograd::test_leaf_assignment PASSED [ 28%] 2023-03-31T06:24:33.6386798Z test_autograd.py::TestAutograd::test_legacy_function_deprecation_exception PASSED [ 28%] 2023-03-31T06:24:33.6387817Z test_autograd.py::TestAutograd::test_lobpcg SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/80338 for platform(s) linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 28%] 2023-03-31T06:24:33.6388440Z test_autograd.py::TestAutograd::test_mark_non_differentiable PASSED [ 28%] 2023-03-31T06:24:33.6388780Z test_autograd.py::TestAutograd::test_mark_non_differentiable_mixed PASSED [ 29%] 2023-03-31T06:24:33.6389143Z test_autograd.py::TestAutograd::test_mark_non_differentiable_none PASSED [ 29%] 2023-03-31T06:24:33.6389468Z test_autograd.py::TestAutograd::test_materialize_grads PASSED [ 29%] 2023-03-31T06:24:33.6389784Z test_autograd.py::TestAutograd::test_multi_backward PASSED [ 29%] 2023-03-31T06:24:33.6390087Z test_autograd.py::TestAutograd::test_multi_backward_no_grad PASSED [ 29%] 2023-03-31T06:24:33.6390411Z test_autograd.py::TestAutograd::test_multi_grad_hooks PASSED [ 29%] 2023-03-31T06:24:33.6390738Z test_autograd.py::TestAutograd::test_named_tensor_for_complex_views PASSED [ 30%] 2023-03-31T06:24:33.6391076Z test_autograd.py::TestAutograd::test_naughty_anomaly_access XFAIL [ 30%] 2023-03-31T06:24:33.6391409Z test_autograd.py::TestAutograd::test_naughty_autograd_function_attribute_access PASSED [ 30%] 2023-03-31T06:24:33.6391774Z test_autograd.py::TestAutograd::test_naughty_autograd_function_stashing_ctx PASSED [ 30%] 2023-03-31T06:24:33.6392118Z test_autograd.py::TestAutograd::test_nested_anomaly_detect_nan PASSED [ 30%] 2023-03-31T06:24:33.6392451Z test_autograd.py::TestAutograd::test_nested_anomaly_printstack_cleanup PASSED [ 31%] 2023-03-31T06:24:33.6392788Z test_autograd.py::TestAutograd::test_next_functions PASSED [ 31%] 2023-03-31T06:24:33.6393101Z test_autograd.py::TestAutograd::test_no_grad PASSED [ 31%] 2023-03-31T06:24:33.6393409Z test_autograd.py::TestAutograd::test_no_grad_assignment PASSED [ 31%] 2023-03-31T06:24:33.6393742Z test_autograd.py::TestAutograd::test_no_grad_copy PASSED [ 31%] 2023-03-31T06:24:33.6394082Z test_autograd.py::TestAutograd::test_no_grad_copy_sparse PASSED [ 31%] 2023-03-31T06:24:33.6394395Z test_autograd.py::TestAutograd::test_no_grad_input PASSED [ 32%] 2023-03-31T06:24:33.6394697Z test_autograd.py::TestAutograd::test_no_grad_modifies_version PASSED [ 32%] 2023-03-31T06:24:33.6395020Z test_autograd.py::TestAutograd::test_no_grad_python_function PASSED [ 32%] 2023-03-31T06:24:33.6395342Z test_autograd.py::TestAutograd::test_no_requires_grad_inplace PASSED [ 32%] 2023-03-31T06:24:33.6395661Z test_autograd.py::TestAutograd::test_no_unnecessary_save PASSED [ 32%] 2023-03-31T06:24:33.6395970Z test_autograd.py::TestAutograd::test_no_unnecessary_unwrapping PASSED [ 32%] 2023-03-31T06:24:33.6396291Z test_autograd.py::TestAutograd::test_not_implemented_fwad PASSED [ 33%] 2023-03-31T06:24:33.6396612Z test_autograd.py::TestAutograd::test_not_implemented_grad PASSED [ 33%] 2023-03-31T06:24:33.6396916Z test_autograd.py::TestAutograd::test_numpy_requires_grad PASSED [ 33%] 2023-03-31T06:24:33.6397231Z test_autograd.py::TestAutograd::test_once_differentiable PASSED [ 33%] 2023-03-31T06:24:33.6397574Z test_autograd.py::TestAutograd::test_out_variant_raises_when_inputs_require_grad PASSED [ 33%] 2023-03-31T06:24:33.6397952Z test_autograd.py::TestAutograd::test_pack_hook_with_inplace_modification_should_fail PASSED [ 33%] 2023-03-31T06:24:33.6398326Z test_autograd.py::TestAutograd::test_pickle PASSED [ 34%] 2023-03-31T06:24:33.6398639Z test_autograd.py::TestAutograd::test_pow_zero_tensor_gradient PASSED [ 34%] 2023-03-31T06:24:33.6398954Z test_autograd.py::TestAutograd::test_power_function PASSED [ 34%] 2023-03-31T06:24:33.6399258Z test_autograd.py::TestAutograd::test_prehook_ordering PASSED [ 34%] 2023-03-31T06:24:33.6399794Z test_autograd.py::TestAutograd::test_profiler STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6400309Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6400783Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6401062Z PASSED [ 34%] 2023-03-31T06:24:33.6401392Z test_autograd.py::TestAutograd::test_profiler_aggregation_fake PASSED [ 35%] 2023-03-31T06:24:33.6401944Z test_autograd.py::TestAutograd::test_profiler_aggregation_lstm STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6402467Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6402920Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6403406Z PASSED [ 35%] 2023-03-31T06:24:33.6403897Z test_autograd.py::TestAutograd::test_profiler_aggregation_table STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6404409Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6404879Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6405162Z PASSED [ 35%] 2023-03-31T06:24:33.6405435Z test_autograd.py::TestAutograd::test_profiler_function_event_avg PASSED [ 35%] 2023-03-31T06:24:33.6405966Z test_autograd.py::TestAutograd::test_profiler_propagation STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6406490Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6407039Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6407343Z PASSED [ 35%] 2023-03-31T06:24:33.6407810Z test_autograd.py::TestAutograd::test_profiler_seq_nr STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6408317Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6408785Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6409054Z PASSED [ 35%] 2023-03-31T06:24:33.6409528Z test_autograd.py::TestAutograd::test_profiler_shapes STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6410041Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6410508Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6410779Z PASSED [ 36%] 2023-03-31T06:24:33.6411258Z test_autograd.py::TestAutograd::test_profiler_unboxed_only STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6411772Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6412227Z STAGE:2023-03-31 06:24:24 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6412508Z PASSED [ 36%] 2023-03-31T06:24:33.6412832Z test_autograd.py::TestAutograd::test_pynode_destruction_deadlock PASSED [ 36%] 2023-03-31T06:24:33.6413365Z test_autograd.py::TestAutograd::test_record_function STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6413857Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6414327Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6414788Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6415243Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6415689Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6416021Z PASSED [ 36%] 2023-03-31T06:24:33.6416506Z test_autograd.py::TestAutograd::test_record_function_callbacks STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6417013Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6417475Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6417754Z PASSED [ 36%] 2023-03-31T06:24:33.6418225Z test_autograd.py::TestAutograd::test_record_function_legacy STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:311] Completed Stage: Warm Up 2023-03-31T06:24:33.6418722Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:317] Completed Stage: Collection 2023-03-31T06:24:33.6419179Z STAGE:2023-03-31 06:24:25 5620:5620 ActivityProfilerController.cpp:321] Completed Stage: Post Processing 2023-03-31T06:24:33.6419454Z PASSED [ 36%] 2023-03-31T06:24:33.6419716Z test_autograd.py::TestAutograd::test_record_function_multithreaded PASSED [ 37%] 2023-03-31T06:24:33.6420053Z test_autograd.py::TestAutograd::test_reentrant_child_error PASSED [ 37%] 2023-03-31T06:24:33.6420381Z test_autograd.py::TestAutograd::test_reentrant_priority PASSED [ 37%] 2023-03-31T06:24:33.6420725Z test_autograd.py::TestAutograd::test_reentrant_with_callbacks_both_depths PASSED [ 37%] 2023-03-31T06:24:33.6421108Z test_autograd.py::TestAutograd::test_reentrant_with_callbacks_depth_0 PASSED [ 37%] 2023-03-31T06:24:33.6421486Z test_autograd.py::TestAutograd::test_reentrant_with_callbacks_depth_1 PASSED [ 37%] 2023-03-31T06:24:33.6421835Z test_autograd.py::TestAutograd::test_reentrant_with_leaf_variable_hook PASSED [ 38%] 2023-03-31T06:24:33.6422176Z test_autograd.py::TestAutograd::test_reentrant_with_non_leaf_variable_hook PASSED [ 38%] 2023-03-31T06:24:33.6422515Z test_autograd.py::TestAutograd::test_requires_grad PASSED [ 38%] 2023-03-31T06:24:33.6422909Z test_autograd.py::TestAutograd::test_requires_grad_ PASSED [ 38%] 2023-03-31T06:24:33.6423240Z test_autograd.py::TestAutograd::test_requires_grad_inplace PASSED [ 38%] 2023-03-31T06:24:33.6423545Z test_autograd.py::TestAutograd::test_retain_grad PASSED [ 39%] 2023-03-31T06:24:33.6423854Z test_autograd.py::TestAutograd::test_retain_grad_cycle PASSED [ 39%] 2023-03-31T06:24:33.6424179Z test_autograd.py::TestAutograd::test_retain_grad_inplace PASSED [ 39%] 2023-03-31T06:24:33.6424496Z test_autograd.py::TestAutograd::test_retain_grad_inplace_over_view PASSED [ 39%] 2023-03-31T06:24:33.6424857Z test_autograd.py::TestAutograd::test_retains_grad_can_always_observe_tensor_prehook PASSED [ 39%] 2023-03-31T06:24:33.6425228Z test_autograd.py::TestAutograd::test_retains_grad_inplace_multiple_outputs PASSED [ 39%] 2023-03-31T06:24:33.6425576Z test_autograd.py::TestAutograd::test_return_duplicate PASSED [ 40%] 2023-03-31T06:24:33.6425889Z test_autograd.py::TestAutograd::test_return_duplicate_inplace PASSED [ 40%] 2023-03-31T06:24:33.6426241Z test_autograd.py::TestAutograd::test_return_leaf PASSED [ 40%] 2023-03-31T06:24:33.6426554Z test_autograd.py::TestAutograd::test_return_leaf_inplace PASSED [ 40%] 2023-03-31T06:24:33.6426861Z test_autograd.py::TestAutograd::test_save_none_for_backward PASSED [ 40%] 2023-03-31T06:24:33.6427189Z test_autograd.py::TestAutograd::test_save_on_cpu_and_checkpoint PASSED [ 40%] 2023-03-31T06:24:33.6427508Z test_autograd.py::TestAutograd::test_save_output_nr PASSED [ 41%] 2023-03-31T06:24:33.6427850Z test_autograd.py::TestAutograd::test_saved_tensor_hooks_custom_error_propagaation PASSED [ 41%] 2023-03-31T06:24:33.6428253Z test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_default_hooks PASSED [ 41%] 2023-03-31T06:24:33.6428718Z test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_hooks PASSED [ 41%] 2023-03-31T06:24:33.6429142Z test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_saved_original_with_default_hooks PASSED [ 41%] 2023-03-31T06:24:33.6429556Z test_autograd.py::TestAutograd::test_saved_variable_packing_unpacking_saved_original_with_hooks PASSED [ 41%] 2023-03-31T06:24:33.6429935Z test_autograd.py::TestAutograd::test_saved_variable_saved_original_inplace_detach PASSED [ 42%] 2023-03-31T06:24:33.6430293Z test_autograd.py::TestAutograd::test_saved_variable_version_counter PASSED [ 42%] 2023-03-31T06:24:33.6430634Z test_autograd.py::TestAutograd::test_saved_variables_deprecated PASSED [ 42%] 2023-03-31T06:24:33.6430948Z test_autograd.py::TestAutograd::test_saving_variable_to_disk PASSED [ 42%] 2023-03-31T06:24:33.6431270Z test_autograd.py::TestAutograd::test_select_expanded_v PASSED [ 42%] 2023-03-31T06:24:33.6431576Z test_autograd.py::TestAutograd::test_select_sum PASSED [ 43%] 2023-03-31T06:24:33.6431894Z test_autograd.py::TestAutograd::test_set_data_preserve_pyobj PASSED [ 43%] 2023-03-31T06:24:33.6432207Z test_autograd.py::TestAutograd::test_set_data_self_requires_grad PASSED [ 43%] 2023-03-31T06:24:33.6432530Z test_autograd.py::TestAutograd::test_set_data_tensorimpl_type PASSED [ 43%] 2023-03-31T06:24:33.6432848Z test_autograd.py::TestAutograd::test_set_grad_coroutines PASSED [ 43%] 2023-03-31T06:24:33.6433206Z test_autograd.py::TestAutograd::test_set_grad_coroutines_benign_exceptions PASSED [ 43%] 2023-03-31T06:24:33.6433607Z test_autograd.py::TestAutograd::test_set_grad_coroutines_critical_exceptions PASSED [ 44%] 2023-03-31T06:24:33.6433953Z test_autograd.py::TestAutograd::test_set_grad_coroutines_exit PASSED [ 44%] 2023-03-31T06:24:33.6434271Z test_autograd.py::TestAutograd::test_set_grad_enabled PASSED [ 44%] 2023-03-31T06:24:33.6434581Z test_autograd.py::TestAutograd::test_set_grad_generator_functions PASSED [ 44%] 2023-03-31T06:24:33.6434930Z test_autograd.py::TestAutograd::test_set_grad_generator_functions_recursive PASSED [ 44%] 2023-03-31T06:24:33.6435264Z test_autograd.py::TestAutograd::test_setitem PASSED [ 44%] 2023-03-31T06:24:33.6435559Z test_autograd.py::TestAutograd::test_setitem_mask PASSED [ 45%] 2023-03-31T06:24:33.6435914Z test_autograd.py::TestAutograd::test_setting_default_saved_variable_hooks_twice_should_not_fail PASSED [ 45%] 2023-03-31T06:24:33.6436319Z test_autograd.py::TestAutograd::test_setting_default_saved_variable_hooks_twice_should_use_inner PASSED [ 45%] 2023-03-31T06:24:33.6436672Z test_autograd.py::TestAutograd::test_shape PASSED [ 45%] 2023-03-31T06:24:33.6436967Z test_autograd.py::TestAutograd::test_sharded_grad PASSED [ 45%] 2023-03-31T06:24:33.6437275Z test_autograd.py::TestAutograd::test_simple_reentrant PASSED [ 45%] 2023-03-31T06:24:33.6437587Z test_autograd.py::TestAutograd::test_slice_expanded_v PASSED [ 46%] 2023-03-31T06:24:33.6437897Z test_autograd.py::TestAutograd::test_sparse_gather_both_scalar PASSED [ 46%] 2023-03-31T06:24:33.6438248Z test_autograd.py::TestAutograd::test_sparse_gather_dim0 PASSED [ 46%] 2023-03-31T06:24:33.6438561Z test_autograd.py::TestAutograd::test_sparse_gather_dim1 PASSED [ 46%] 2023-03-31T06:24:33.6438871Z test_autograd.py::TestAutograd::test_sparse_gather_dim_neg PASSED [ 46%] 2023-03-31T06:24:33.6439178Z test_autograd.py::TestAutograd::test_sparse_gather_ind_scalar PASSED [ 47%] 2023-03-31T06:24:33.6439492Z test_autograd.py::TestAutograd::test_sparse_gather_x_scalar PASSED [ 47%] 2023-03-31T06:24:33.6439805Z test_autograd.py::TestAutograd::test_sparse_mm_backward PASSED [ 47%] 2023-03-31T06:24:33.6440107Z test_autograd.py::TestAutograd::test_tensor_grad_warnings PASSED [ 47%] 2023-03-31T06:24:33.6440427Z test_autograd.py::TestAutograd::test_tensor_hooks_inplace PASSED [ 47%] 2023-03-31T06:24:33.6440797Z test_autograd.py::TestAutograd::test_tensor_hooks_inplace_multiple_outputs PASSED [ 47%] 2023-03-31T06:24:33.6441149Z test_autograd.py::TestAutograd::test_tensor_hooks_inplace_over_view PASSED [ 48%] 2023-03-31T06:24:33.6441467Z test_autograd.py::TestAutograd::test_thread_shutdown PASSED [ 48%] 2023-03-31T06:24:33.6441787Z test_autograd.py::TestAutograd::test_to_sparse_backward PASSED [ 48%] 2023-03-31T06:24:33.6442099Z test_autograd.py::TestAutograd::test_too_many_grads PASSED [ 48%] 2023-03-31T06:24:33.6442405Z test_autograd.py::TestAutograd::test_type_conversions PASSED [ 48%] 2023-03-31T06:24:33.6442722Z test_autograd.py::TestAutograd::test_unrelated_inputs PASSED [ 48%] 2023-03-31T06:24:33.6443222Z test_autograd.py::TestAutograd::test_unsafe_set_version_counter PASSED [ 49%] 2023-03-31T06:24:33.6443672Z test_autograd.py::TestAutograd::test_unused_output PASSED [ 49%] 2023-03-31T06:24:33.6444065Z test_autograd.py::TestAutograd::test_var_mean_differentiable PASSED [ 49%] 2023-03-31T06:24:33.6444390Z test_autograd.py::TestAutograd::test_variable_traverse PASSED [ 49%] 2023-03-31T06:24:33.6444709Z test_autograd.py::TestAutograd::test_version_counter PASSED [ 49%] 2023-03-31T06:24:33.6445008Z test_autograd.py::TestAutograd::test_view_func_replay PASSED [ 50%] 2023-03-31T06:24:33.6445322Z test_autograd.py::TestAutograd::test_view_replay_enabled PASSED [ 50%] 2023-03-31T06:24:33.6445711Z test_autograd.py::TestAutograd::test_volatile_deprecated PASSED [ 50%] 2023-03-31T06:24:33.6446079Z test_autograd.py::TestAutograd::test_will_engine_execute_node PASSED [ 50%] 2023-03-31T06:24:33.6446409Z test_autograd.py::TestAutograd::test_wrapped_number_saved_variable_hooks PASSED [ 50%] 2023-03-31T06:24:33.6446816Z test_autograd.py::TestAutogradForwardModeBatchedGrad::test_inplace_on_view_not_same_layout PASSED [ 50%] 2023-03-31T06:24:33.6447255Z test_autograd.py::TestAutogradForwardModeBatchedGrad::test_inplace_on_view_same_layout PASSED [ 51%] 2023-03-31T06:24:33.6447708Z test_autograd.py::TestAutogradForwardModeBatchedGrad::test_metadata_check_for_storage_numel_skipped PASSED [ 51%] 2023-03-31T06:24:33.6448131Z test_autograd.py::TestAutogradForwardModeBatchedGrad::test_out_of_place_basic PASSED [ 51%] 2023-03-31T06:24:33.6448556Z test_autograd.py::TestAutogradForwardModeBatchedGrad::test_out_of_place_not_same_layout PASSED [ 51%] 2023-03-31T06:24:33.6448963Z test_autograd.py::TestAutogradForwardMode::test_advanced_packing_unpacking PASSED [ 51%] 2023-03-31T06:24:33.6449330Z test_autograd.py::TestAutogradForwardMode::test_backward_graph_destruction PASSED [ 51%] 2023-03-31T06:24:33.6449704Z test_autograd.py::TestAutogradForwardMode::test_basic_packing_unpacking PASSED [ 52%] 2023-03-31T06:24:33.6450091Z test_autograd.py::TestAutogradForwardMode::test_codegen_ignores_undefined_outputs PASSED [ 52%] 2023-03-31T06:24:33.6450475Z test_autograd.py::TestAutogradForwardMode::test_create_new_zeros_with_same_meta PASSED [ 52%] 2023-03-31T06:24:33.6450829Z test_autograd.py::TestAutogradForwardMode::test_default_level PASSED [ 52%] 2023-03-31T06:24:33.6451226Z test_autograd.py::TestAutogradForwardMode::test_detach_view_tracking PASSED [ 52%] 2023-03-31T06:24:33.6451593Z test_autograd.py::TestAutogradForwardMode::test_forward_level_cleanup PASSED [ 52%] 2023-03-31T06:24:33.6451937Z test_autograd.py::TestAutogradForwardMode::test_fwd_grad_enabled PASSED [ 53%] 2023-03-31T06:24:33.6452284Z test_autograd.py::TestAutogradForwardMode::test_grad_cleanup PASSED [ 53%] 2023-03-31T06:24:33.6452648Z test_autograd.py::TestAutogradForwardMode::test_make_dual_forbid_integral_dtype PASSED [ 53%] 2023-03-31T06:24:33.6453049Z test_autograd.py::TestAutogradForwardMode::test_make_dual_inference_tensor_in_inference_mode PASSED [ 53%] 2023-03-31T06:24:33.6453427Z test_autograd.py::TestAutogradForwardMode::test_make_dual_torch_dispatch PASSED [ 53%] 2023-03-31T06:24:33.6453840Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_check_conj PASSED [ 54%] 2023-03-31T06:24:33.6454231Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_checks_ignores_size_zero PASSED [ 54%] 2023-03-31T06:24:33.6454614Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_checks_storage_numel PASSED [ 54%] 2023-03-31T06:24:33.6455031Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_ignore_storage_offset_for_zero_numel_tensor PASSED [ 54%] 2023-03-31T06:24:33.6455452Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_when_primal_has_conj_bit PASSED [ 54%] 2023-03-31T06:24:33.6455850Z test_autograd.py::TestAutogradForwardMode::test_metadata_check_when_primal_has_neg_bit PASSED [ 54%] 2023-03-31T06:24:33.6456213Z test_autograd.py::TestAutogradForwardMode::test_nested_level PASSED [ 55%] 2023-03-31T06:24:33.6456568Z test_autograd.py::TestAutogradForwardMode::test_non_differentiable PASSED [ 55%] 2023-03-31T06:24:33.6456921Z test_autograd.py::TestAutogradForwardMode::test_out_variant PASSED [ 55%] 2023-03-31T06:24:33.6457263Z test_autograd.py::TestAutogradForwardMode::test_print PASSED [ 55%] 2023-03-31T06:24:33.6457628Z test_autograd.py::TestAutogradForwardMode::test_set_fw_grad_having_own_fw_grad_at_same_level PASSED [ 55%] 2023-03-31T06:24:33.6458013Z test_autograd.py::TestAutogradForwardMode::test_set_fwd_grad_enabled PASSED [ 55%] 2023-03-31T06:24:33.6458362Z test_autograd.py::TestAutogradForwardMode::test_size_check PASSED [ 56%] 2023-03-31T06:24:33.6458763Z test_autograd.py::TestAutogradForwardMode::test_view_inplace_always_creates_a_view PASSED [ 56%] 2023-03-31T06:24:33.6459188Z test_autograd.py::TestAutogradForwardMode::test_view_inplace_differentiable_views PASSED [ 56%] 2023-03-31T06:24:33.6459591Z test_autograd.py::TestAutogradForwardMode::test_view_inplace_non_differentiable_views PASSED [ 56%] 2023-03-31T06:24:33.6459979Z test_autograd.py::TestAllowMutationOnSaved::test_backward_out_of_context PASSED [ 56%] 2023-03-31T06:24:33.6460326Z test_autograd.py::TestAllowMutationOnSaved::test_basic PASSED [ 56%] 2023-03-31T06:24:33.6460681Z test_autograd.py::TestAllowMutationOnSaved::test_disallow_nesting PASSED [ 57%] 2023-03-31T06:24:33.6461040Z test_autograd.py::TestAllowMutationOnSaved::test_double_backward PASSED [ 57%] 2023-03-31T06:24:33.6461391Z test_autograd.py::TestAllowMutationOnSaved::test_save_base_and_modify_view PASSED [ 57%] 2023-03-31T06:24:33.6461764Z test_autograd.py::TestAllowMutationOnSaved::test_save_view_modify_base PASSED [ 57%] 2023-03-31T06:24:33.6462134Z test_autograd.py::TestAllowMutationOnSaved::test_saved_but_not_anymore PASSED [ 57%] 2023-03-31T06:24:33.6462528Z test_autograd.py::TestAllowMutationOnSaved::test_saved_same_tensor_different_versions PASSED [ 58%] 2023-03-31T06:24:33.6462983Z test_autograd.py::TestAllowMutationOnSaved::test_saved_same_tensor_many_times PASSED [ 58%] 2023-03-31T06:24:33.6463347Z test_autograd.py::TestAllowMutationOnSaved::test_views PASSED [ 58%] 2023-03-31T06:24:33.6463695Z test_autograd.py::TestAllowMutationOnSaved::test_with_math_views PASSED [ 58%] 2023-03-31T06:24:33.6464071Z test_autograd.py::TestAllowMutationOnSaved::test_with_out_variant PASSED [ 58%] 2023-03-31T06:24:33.6464451Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_context_manager PASSED [ 58%] 2023-03-31T06:24:33.6464841Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_decorator PASSED [ 59%] 2023-03-31T06:24:33.6465250Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_existing_autograd_session PASSED [ 59%] 2023-03-31T06:24:33.6465659Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_handle_direct_view_on_rebase PASSED [ 59%] 2023-03-31T06:24:33.6466088Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_handle_indirect_view_on_rebase PASSED [ 59%] 2023-03-31T06:24:33.6466520Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_functional_op PASSED [ 59%] 2023-03-31T06:24:33.6466990Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_inplace_op PASSED [ 59%] 2023-03-31T06:24:33.6467402Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_inf_mode_view_op PASSED [ 60%] 2023-03-31T06:24:33.6467833Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_functional_op PASSED [ 60%] 2023-03-31T06:24:33.6468282Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_inplace_op PASSED [ 60%] 2023-03-31T06:24:33.6468721Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_inf_tensor_in_normal_mode_view_op PASSED [ 60%] 2023-03-31T06:24:33.6469125Z test_autograd.py::TestAutogradInferenceMode::test_inference_mode_tensor_creation PASSED [ 60%] 2023-03-31T06:24:33.6469537Z test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_functional_op PASSED [ 60%] 2023-03-31T06:24:33.6469953Z test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_inplace_op PASSED [ 61%] 2023-03-31T06:24:33.6470356Z test_autograd.py::TestAutogradInferenceMode::test_mix_inference_and_normal_tensor_view_op PASSED [ 61%] 2023-03-31T06:24:33.6470779Z test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_inplace_output_in_inference_mode PASSED [ 61%] 2023-03-31T06:24:33.6471195Z test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_inplace_output_in_normal_mode PASSED [ 61%] 2023-03-31T06:24:33.6471649Z test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_view_output_in_inference_mode PASSED [ 61%] 2023-03-31T06:24:33.6472081Z test_autograd.py::TestAutogradInferenceMode::test_normal_tensor_view_output_in_normal_mode PASSED [ 62%] 2023-03-31T06:24:33.6472456Z test_autograd.py::TestMultithreadAutograd::test_cat_stack_r_to_c PASSED [ 62%] 2023-03-31T06:24:33.6472828Z test_autograd.py::TestMultithreadAutograd::test_dataparallel_saved_tensors_hooks PASSED [ 62%] 2023-03-31T06:24:33.6473202Z test_autograd.py::TestMultithreadAutograd::test_fork_join_in_middle PASSED [ 62%] 2023-03-31T06:24:33.6473545Z test_autograd.py::TestMultithreadAutograd::test_multi_grad_hooks PASSED [ 62%] 2023-03-31T06:24:33.6473925Z test_autograd.py::TestMultithreadAutograd::test_multithreaded_exception_propagation PASSED [ 62%] 2023-03-31T06:24:33.6474308Z test_autograd.py::TestMultithreadAutograd::test_preserve_backtrace PASSED [ 63%] 2023-03-31T06:24:33.6474658Z test_autograd.py::TestMultithreadAutograd::test_python_thread_in_middle PASSED [ 63%] 2023-03-31T06:24:33.6475019Z test_autograd.py::TestMultithreadAutograd::test_simple_backward PASSED [ 63%] 2023-03-31T06:24:33.6475385Z test_autograd.py::TestMultithreadAutograd::test_simple_backward_same_input PASSED [ 63%] 2023-03-31T06:24:33.6475764Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_early_stop_False PASSED [ 63%] 2023-03-31T06:24:33.6476134Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_early_stop_True PASSED [ 63%] 2023-03-31T06:24:33.6476526Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_kwargs_early_stop_False PASSED [ 64%] 2023-03-31T06:24:33.6476953Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_kwargs_early_stop_True PASSED [ 64%] 2023-03-31T06:24:33.6477376Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_False PASSED [ 64%] 2023-03-31T06:24:33.6477808Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_True PASSED [ 64%] 2023-03-31T06:24:33.6478248Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_reentrant_backwards_early_stop_False PASSED [ 64%] 2023-03-31T06:24:33.6478676Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_reentrant_backwards_early_stop_True PASSED [ 64%] 2023-03-31T06:24:33.6479078Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_same_graph_early_stop_False PASSED [ 65%] 2023-03-31T06:24:33.6479518Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_same_graph_early_stop_True PASSED [ 65%] 2023-03-31T06:24:33.6479906Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_set_early_stop PASSED [ 65%] 2023-03-31T06:24:33.6480307Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_set_early_stop_no_recompution_needed PASSED [ 65%] 2023-03-31T06:24:33.6480709Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_two_children_early_stop_False PASSED [ 65%] 2023-03-31T06:24:33.6481120Z test_autograd.py::TestNestedCheckpoint::test_nested_checkpoint_two_children_early_stop_True PASSED [ 66%] 2023-03-31T06:24:33.6481701Z test_autograd.py::TestAutogradComplex::test_view_func_for_complex_views <- test/autograd/test_complex.py PASSED [ 66%] 2023-03-31T06:24:33.6482219Z test_autograd.py::TestAutogradComplex::test_view_with_multi_output <- test/autograd/test_complex.py PASSED [ 66%] 2023-03-31T06:24:33.6482766Z test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_base_tensor <- test/autograd/test_functional.py PASSED [ 66%] 2023-03-31T06:24:33.6483541Z test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_cuda_base_tensor <- test/autograd/test_functional.py SKIPPED (test requires CUDA) [ 66%] 2023-03-31T06:24:33.6484204Z test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_cuda_logging_tensor <- test/autograd/test_functional.py SKIPPED (test requires CUDA) [ 66%] 2023-03-31T06:24:33.6484896Z test_autograd.py::TestAutogradFunctional::test_construct_standard_basis_for_logging_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6485526Z test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_False_base_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6486138Z test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_False_logging_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6486744Z test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_True_base_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6487349Z test_autograd.py::TestAutogradFunctional::test_hessian_create_graph_vectorize_True_logging_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6487924Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 67%] 2023-03-31T06:24:33.6488497Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6489086Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_vectorize_base_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6489695Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_strict_vectorize_logging_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6490280Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_False_base_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6490920Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_False_logging_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6491513Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_True_base_tensor <- test/autograd/test_functional.py PASSED [ 68%] 2023-03-31T06:24:33.6492112Z test_autograd.py::TestAutogradFunctional::test_hessian_err_check_vectorize_True_logging_tensor <- test/autograd/test_functional.py PASSED [ 69%] 2023-03-31T06:24:33.6492684Z test_autograd.py::TestAutogradFunctional::test_hessian_match_vhp_hvp_base_tensor <- test/autograd/test_functional.py PASSED [ 69%] 2023-03-31T06:24:33.6493237Z test_autograd.py::TestAutogradFunctional::test_hessian_match_vhp_hvp_logging_tensor <- test/autograd/test_functional.py PASSED [ 69%] 2023-03-31T06:24:33.6493788Z test_autograd.py::TestAutogradFunctional::test_hessian_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 69%] 2023-03-31T06:24:33.6494403Z test_autograd.py::TestAutogradFunctional::test_hessian_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 69%] 2023-03-31T06:24:33.6494951Z test_autograd.py::TestAutogradFunctional::test_hessian_output_base_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6495487Z test_autograd.py::TestAutogradFunctional::test_hessian_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6496055Z test_autograd.py::TestAutogradFunctional::test_hessian_output_vectorized_base_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6496639Z test_autograd.py::TestAutogradFunctional::test_hessian_output_vectorized_logging_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6497225Z test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_False_base_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6497809Z test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_False_logging_tensor <- test/autograd/test_functional.py PASSED [ 70%] 2023-03-31T06:24:33.6498397Z test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_True_base_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6498989Z test_autograd.py::TestAutogradFunctional::test_hessian_scalar_vectorize_True_logging_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6499638Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_multi_input_base_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6500270Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_multi_input_logging_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6500874Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_simple_base_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6501489Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_simple_logging_tensor <- test/autograd/test_functional.py PASSED [ 71%] 2023-03-31T06:24:33.6502111Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_unrelated_outputs_base_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6502746Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_correctness_unrelated_outputs_logging_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6503444Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_raises_no_warnings_base_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6504051Z test_autograd.py::TestAutogradFunctional::test_hessian_vectorize_raises_no_warnings_logging_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6504623Z test_autograd.py::TestAutogradFunctional::test_hvp_create_graph_base_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6505216Z test_autograd.py::TestAutogradFunctional::test_hvp_create_graph_logging_tensor <- test/autograd/test_functional.py PASSED [ 72%] 2023-03-31T06:24:33.6505768Z test_autograd.py::TestAutogradFunctional::test_hvp_err_check_base_tensor <- test/autograd/test_functional.py PASSED [ 73%] 2023-03-31T06:24:33.6506314Z test_autograd.py::TestAutogradFunctional::test_hvp_err_check_logging_tensor <- test/autograd/test_functional.py PASSED [ 73%] 2023-03-31T06:24:33.6506876Z test_autograd.py::TestAutogradFunctional::test_hvp_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 73%] 2023-03-31T06:24:33.6507428Z test_autograd.py::TestAutogradFunctional::test_hvp_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 73%] 2023-03-31T06:24:33.6507973Z test_autograd.py::TestAutogradFunctional::test_hvp_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 73%] 2023-03-31T06:24:33.6508545Z test_autograd.py::TestAutogradFunctional::test_hvp_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6509082Z test_autograd.py::TestAutogradFunctional::test_hvp_output_base_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6509606Z test_autograd.py::TestAutogradFunctional::test_hvp_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6510144Z test_autograd.py::TestAutogradFunctional::test_hvp_scalar_base_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6510679Z test_autograd.py::TestAutogradFunctional::test_hvp_scalar_logging_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6511264Z test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_False_base_tensor <- test/autograd/test_functional.py PASSED [ 74%] 2023-03-31T06:24:33.6511864Z test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_False_logging_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6512473Z test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_True_base_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6513077Z test_autograd.py::TestAutogradFunctional::test_jacobian_create_graph_vectorize_True_logging_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6513700Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6514298Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6514890Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_vectorize_base_tensor <- test/autograd/test_functional.py PASSED [ 75%] 2023-03-31T06:24:33.6515489Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_strict_vectorize_logging_tensor <- test/autograd/test_functional.py PASSED [ 76%] 2023-03-31T06:24:33.6516089Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_False_base_tensor <- test/autograd/test_functional.py PASSED [ 76%] 2023-03-31T06:24:33.6516677Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_False_logging_tensor <- test/autograd/test_functional.py PASSED [ 76%] 2023-03-31T06:24:33.6517275Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_True_base_tensor <- test/autograd/test_functional.py PASSED [ 76%] 2023-03-31T06:24:33.6517869Z test_autograd.py::TestAutogradFunctional::test_jacobian_err_check_vectorize_True_logging_tensor <- test/autograd/test_functional.py PASSED [ 76%] 2023-03-31T06:24:33.6518446Z test_autograd.py::TestAutogradFunctional::test_jacobian_match_vjp_jvp_base_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6519014Z test_autograd.py::TestAutogradFunctional::test_jacobian_match_vjp_jvp_logging_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6519750Z test_autograd.py::TestAutogradFunctional::test_jacobian_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6520312Z test_autograd.py::TestAutogradFunctional::test_jacobian_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6520862Z test_autograd.py::TestAutogradFunctional::test_jacobian_output_base_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6521418Z test_autograd.py::TestAutogradFunctional::test_jacobian_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 77%] 2023-03-31T06:24:33.6521981Z test_autograd.py::TestAutogradFunctional::test_jacobian_output_vectorized_base_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6522562Z test_autograd.py::TestAutogradFunctional::test_jacobian_output_vectorized_logging_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6523277Z test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_base_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6523829Z test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_logging_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6524388Z test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_vectorized_base_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6524972Z test_autograd.py::TestAutogradFunctional::test_jacobian_scalar_vectorized_logging_tensor <- test/autograd/test_functional.py PASSED [ 78%] 2023-03-31T06:24:33.6525629Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_devices_base_tensor <- test/autograd/test_functional.py SKIPPED (test requires CUDA) [ 79%] 2023-03-31T06:24:33.6526325Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_devices_logging_tensor <- test/autograd/test_functional.py SKIPPED (test requires CUDA) [ 79%] 2023-03-31T06:24:33.6526976Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_dtype_base_tensor <- test/autograd/test_functional.py PASSED [ 79%] 2023-03-31T06:24:33.6527614Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_different_dtype_logging_tensor <- test/autograd/test_functional.py PASSED [ 79%] 2023-03-31T06:24:33.6528304Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_base_tensor <- test/autograd/test_functional.py PASSED [ 79%] 2023-03-31T06:24:33.6528968Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_logging_tensor <- test/autograd/test_functional.py PASSED [ 79%] 2023-03-31T06:24:33.6529601Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_multi_output_base_tensor <- test/autograd/test_functional.py PASSED [ 80%] 2023-03-31T06:24:33.6530260Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_multi_input_multi_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 80%] 2023-03-31T06:24:33.6530892Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_simple_base_tensor <- test/autograd/test_functional.py PASSED [ 80%] 2023-03-31T06:24:33.6531503Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_simple_logging_tensor <- test/autograd/test_functional.py PASSED [ 80%] 2023-03-31T06:24:33.6532135Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_unrelated_outputs_base_tensor <- test/autograd/test_functional.py PASSED [ 80%] 2023-03-31T06:24:33.6532763Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_unrelated_outputs_logging_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6533390Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_zero_dim_base_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6534041Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_correctness_zero_dim_logging_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6534648Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_raises_no_warnings_base_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6535248Z test_autograd.py::TestAutogradFunctional::test_jacobian_vectorize_raises_no_warnings_logging_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6535824Z test_autograd.py::TestAutogradFunctional::test_jvp_create_graph_base_tensor <- test/autograd/test_functional.py PASSED [ 81%] 2023-03-31T06:24:33.6536380Z test_autograd.py::TestAutogradFunctional::test_jvp_create_graph_logging_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6536928Z test_autograd.py::TestAutogradFunctional::test_jvp_err_check_base_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6537507Z test_autograd.py::TestAutogradFunctional::test_jvp_err_check_logging_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6538057Z test_autograd.py::TestAutogradFunctional::test_jvp_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6538617Z test_autograd.py::TestAutogradFunctional::test_jvp_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6539163Z test_autograd.py::TestAutogradFunctional::test_jvp_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 82%] 2023-03-31T06:24:33.6539694Z test_autograd.py::TestAutogradFunctional::test_jvp_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6540233Z test_autograd.py::TestAutogradFunctional::test_jvp_output_base_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6540780Z test_autograd.py::TestAutogradFunctional::test_jvp_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6541322Z test_autograd.py::TestAutogradFunctional::test_jvp_scalar_base_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6541847Z test_autograd.py::TestAutogradFunctional::test_jvp_scalar_logging_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6542397Z test_autograd.py::TestAutogradFunctional::test_vhp_create_graph_base_tensor <- test/autograd/test_functional.py PASSED [ 83%] 2023-03-31T06:24:33.6543077Z test_autograd.py::TestAutogradFunctional::test_vhp_create_graph_logging_tensor <- test/autograd/test_functional.py PASSED [ 84%] 2023-03-31T06:24:33.6543669Z test_autograd.py::TestAutogradFunctional::test_vhp_err_check_base_tensor <- test/autograd/test_functional.py PASSED [ 84%] 2023-03-31T06:24:33.6544200Z test_autograd.py::TestAutogradFunctional::test_vhp_err_check_logging_tensor <- test/autograd/test_functional.py PASSED [ 84%] 2023-03-31T06:24:33.6544755Z test_autograd.py::TestAutogradFunctional::test_vhp_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 84%] 2023-03-31T06:24:33.6545318Z test_autograd.py::TestAutogradFunctional::test_vhp_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 84%] 2023-03-31T06:24:33.6545869Z test_autograd.py::TestAutogradFunctional::test_vhp_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6546395Z test_autograd.py::TestAutogradFunctional::test_vhp_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6546934Z test_autograd.py::TestAutogradFunctional::test_vhp_output_base_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6547474Z test_autograd.py::TestAutogradFunctional::test_vhp_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6548013Z test_autograd.py::TestAutogradFunctional::test_vhp_scalar_base_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6548580Z test_autograd.py::TestAutogradFunctional::test_vhp_scalar_logging_tensor <- test/autograd/test_functional.py PASSED [ 85%] 2023-03-31T06:24:33.6549130Z test_autograd.py::TestAutogradFunctional::test_vjp_create_graph_base_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6549692Z test_autograd.py::TestAutogradFunctional::test_vjp_create_graph_logging_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6550248Z test_autograd.py::TestAutogradFunctional::test_vjp_err_check_base_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6550783Z test_autograd.py::TestAutogradFunctional::test_vjp_err_check_logging_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6551332Z test_autograd.py::TestAutogradFunctional::test_vjp_err_check_strict_base_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6551925Z test_autograd.py::TestAutogradFunctional::test_vjp_err_check_strict_logging_tensor <- test/autograd/test_functional.py PASSED [ 86%] 2023-03-31T06:24:33.6552477Z test_autograd.py::TestAutogradFunctional::test_vjp_no_grad_base_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6553004Z test_autograd.py::TestAutogradFunctional::test_vjp_no_grad_logging_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6553535Z test_autograd.py::TestAutogradFunctional::test_vjp_output_base_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6554072Z test_autograd.py::TestAutogradFunctional::test_vjp_output_logging_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6554602Z test_autograd.py::TestAutogradFunctional::test_vjp_scalar_base_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6555137Z test_autograd.py::TestAutogradFunctional::test_vjp_scalar_logging_tensor <- test/autograd/test_functional.py PASSED [ 87%] 2023-03-31T06:24:33.6555591Z test_autograd.py::TestAutogradDeviceTypeCPU::test_advanced_indexing_backwards_large_cpu SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T06:24:33.6556036Z test_autograd.py::TestAutogradDeviceTypeCPU::test_advanced_indexing_backwards_memory_format_cpu PASSED [ 88%] 2023-03-31T06:24:33.6556462Z test_autograd.py::TestAutogradDeviceTypeCPU::test_backward_device_cpu SKIPPED (fewer than 2 devices detected) [ 88%] 2023-03-31T06:24:33.6556877Z test_autograd.py::TestAutogradDeviceTypeCPU::test_complex_scalar_backward_cpu PASSED [ 88%] 2023-03-31T06:24:33.6557290Z test_autograd.py::TestAutogradDeviceTypeCPU::test_copy__cpu PASSED [ 88%] 2023-03-31T06:24:33.6557702Z test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_forward_ad_broadcasting_cpu PASSED [ 89%] 2023-03-31T06:24:33.6558102Z test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_forward_ad_same_layout_copies_grad_cpu PASSED [ 89%] 2023-03-31T06:24:33.6558485Z test_autograd.py::TestAutogradDeviceTypeCPU::test_copy_r_to_c_cpu PASSED [ 89%] 2023-03-31T06:24:33.6558889Z test_autograd.py::TestAutogradDeviceTypeCPU::test_cross_device_reentrant_autograd_cpu SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T06:24:33.6559326Z test_autograd.py::TestAutogradDeviceTypeCPU::test_free_unneeded_tensor_cpu SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T06:24:33.6559709Z test_autograd.py::TestAutogradDeviceTypeCPU::test_grad_assignment_cpu PASSED [ 89%] 2023-03-31T06:24:33.6560134Z test_autograd.py::TestAutogradDeviceTypeCPU::test_gradcheck_input_output_different_device_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T06:24:33.6560582Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_multiple_output_view_of_view_cpu PASSED [ 90%] 2023-03-31T06:24:33.6560977Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_base_cpu PASSED [ 90%] 2023-03-31T06:24:33.6561375Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_view_cpu PASSED [ 90%] 2023-03-31T06:24:33.6561790Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_backprop_view_of_view_cpu PASSED [ 90%] 2023-03-31T06:24:33.6562233Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_gradcheck_cpu PASSED [ 90%] 2023-03-31T06:24:33.6562630Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_makes_base_require_grad_cpu PASSED [ 91%] 2023-03-31T06:24:33.6563167Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_modify_base_cpu PASSED [ 91%] 2023-03-31T06:24:33.6563576Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multi_output_safe_cpu PASSED [ 91%] 2023-03-31T06:24:33.6563991Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multi_output_unsafe_cpu PASSED [ 91%] 2023-03-31T06:24:33.6564389Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_multiple_outputs_cpu PASSED [ 91%] 2023-03-31T06:24:33.6564793Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_non_contig_cpu PASSED [ 91%] 2023-03-31T06:24:33.6565252Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_of_multiple_output_view_cpu PASSED [ 92%] 2023-03-31T06:24:33.6565643Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_of_view_cpu PASSED [ 92%] 2023-03-31T06:24:33.6566023Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_python_cpu PASSED [ 92%] 2023-03-31T06:24:33.6566416Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inplace_on_view_then_no_grad_cpu PASSED [ 92%] 2023-03-31T06:24:33.6566856Z test_autograd.py::TestAutogradDeviceTypeCPU::test_inputbuffer_add_multidevice_cpu SKIPPED (fewer than 2 devices detected) [ 92%] 2023-03-31T06:24:33.6567290Z test_autograd.py::TestAutogradDeviceTypeCPU::test_min_max_median_backprops_to_all_values_cpu PASSED [ 93%] 2023-03-31T06:24:33.6567678Z test_autograd.py::TestAutogradDeviceTypeCPU::test_mv_grad_stride_0_cpu PASSED [ 93%] 2023-03-31T06:24:33.6568058Z test_autograd.py::TestAutogradDeviceTypeCPU::test_non_differentiable_ops_cpu PASSED [ 93%] 2023-03-31T06:24:33.6568441Z test_autograd.py::TestAutogradDeviceTypeCPU::test_parameter_resize_cpu PASSED [ 93%] 2023-03-31T06:24:33.6568819Z test_autograd.py::TestAutogradDeviceTypeCPU::test_pin_memory_cpu SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T06:24:33.6569236Z test_autograd.py::TestAutogradDeviceTypeCPU::test_pow_real_negative_base_complex_exponent_cpu PASSED [ 93%] 2023-03-31T06:24:33.6569631Z test_autograd.py::TestAutogradDeviceTypeCPU::test_profiler_emit_itt_cpu PASSED [ 94%] 2023-03-31T06:24:33.6570065Z test_autograd.py::TestAutogradDeviceTypeCPU::test_profiler_emit_nvtx_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T06:24:33.6570501Z test_autograd.py::TestAutogradDeviceTypeCPU::test_pyscalar_conversions_cpu PASSED [ 94%] 2023-03-31T06:24:33.6570917Z test_autograd.py::TestAutogradDeviceTypeCPU::test_reentrant_parent_error_on_cpu_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T06:24:33.6571335Z test_autograd.py::TestAutogradDeviceTypeCPU::test_requires_grad_factory_cpu_float32 PASSED [ 94%] 2023-03-31T06:24:33.6571719Z test_autograd.py::TestAutogradDeviceTypeCPU::test_requires_grad_factory_cpu_float64 PASSED [ 94%] 2023-03-31T06:24:33.6572104Z test_autograd.py::TestAutogradDeviceTypeCPU::test_resize_version_bump_cpu PASSED [ 95%] 2023-03-31T06:24:33.6572526Z test_autograd.py::TestAutogradDeviceTypeCPU::test_rnn_backward_to_input_but_not_parameters_cpu SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T06:24:33.6572981Z test_autograd.py::TestAutogradDeviceTypeCPU::test_scatter_index_reduce_amin_amax_backprops_to_all_values_cpu PASSED [ 95%] 2023-03-31T06:24:33.6573415Z test_autograd.py::TestAutogradDeviceTypeCPU::test_scatter_index_reduce_prod_gradgrad_error_cpu PASSED [ 95%] 2023-03-31T06:24:33.6573842Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_float32 PASSED [ 95%] 2023-03-31T06:24:33.6574263Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_float64 PASSED [ 95%] 2023-03-31T06:24:33.6574682Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int16 PASSED [ 96%] 2023-03-31T06:24:33.6575085Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int32 PASSED [ 96%] 2023-03-31T06:24:33.6575542Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int64 PASSED [ 96%] 2023-03-31T06:24:33.6575955Z test_autograd.py::TestAutogradDeviceTypeCPU::test_set_requires_grad_only_for_floats_cpu_int8 PASSED [ 96%] 2023-03-31T06:24:33.6576375Z test_autograd.py::TestAutogradDeviceTypeCPU::test_simple_reentrant_cross_device_cpu SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T06:24:33.6576791Z test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_backward_cpu_complex128 PASSED [ 97%] 2023-03-31T06:24:33.6577176Z test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_backward_cpu_float64 PASSED [ 97%] 2023-03-31T06:24:33.6577577Z test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_ctor_getter_backward_cpu_complex128 PASSED [ 97%] 2023-03-31T06:24:33.6578023Z test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_ctor_getter_backward_cpu_float64 PASSED [ 97%] 2023-03-31T06:24:33.6578420Z test_autograd.py::TestAutogradDeviceTypeCPU::test_sparse_mask_autograd_cpu PASSED [ 97%] 2023-03-31T06:24:33.6578804Z test_autograd.py::TestAutogradDeviceTypeCPU::test_strided_leaf_grad_layout_cpu PASSED [ 97%] 2023-03-31T06:24:33.6579177Z test_autograd.py::TestAutogradDeviceTypeCPU::test_to_r_to_c_cpu PASSED [ 98%] 2023-03-31T06:24:33.6579573Z test_autograd.py::TestAutogradDeviceTypeCPU::test_unused_output_device_cpu SKIPPED (fewer than 2 devices detected) [ 98%] 2023-03-31T06:24:33.6579987Z test_autograd.py::TestAutogradDeviceTypeCPU::test_warning_in_backward_cpu PASSED [ 98%] 2023-03-31T06:24:33.6580362Z test_autograd.py::TestAutogradDeviceTypeCPU::test_where_functional_cpu PASSED [ 98%] 2023-03-31T06:24:33.6580720Z test_autograd.py::TestAutogradDeviceTypeCPU::test_where_scalar_cpu PASSED [ 98%] 2023-03-31T06:24:33.6581151Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_autograd_composite_implicit_and_dispatch_registration_cpu PASSED [ 98%] 2023-03-31T06:24:33.6581630Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_autograd_multiple_dispatch_registrations_cpu PASSED [ 99%] 2023-03-31T06:24:33.6582094Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_backward_single_threaded_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T06:24:33.6582531Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_backward_tls_stash_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T06:24:33.6583057Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_foward_mode_AD_cpu PASSED [ 99%] 2023-03-31T06:24:33.6583508Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_per_dispatch_key_input_saving_cpu PASSED [ 99%] 2023-03-31T06:24:33.6583914Z test_autograd.py::TestAutogradMultipleDispatchCPU::test_view_copy_cpu PASSED [100%] 2023-03-31T06:24:33.6584119Z 2023-03-31T06:24:33.6584504Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_autograd/test_autograd-e1c0d42718cf0b95.xml - 2023-03-31T06:24:33.6584891Z ================= 548 passed, 25 skipped, 1 xfailed in 17.05s ================== 2023-03-31T06:24:33.6585222Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:24:33.6585423Z 2023-03-31T06:24:33.6585812Z ##[endgroup] 2023-03-31T06:24:33.6586214Z FINISHED PRINTING LOG FILE of test_autograd (/var/lib/jenkins/workspace/test/test-reports/test_autograd_b7ivmmwj.log) 2023-03-31T06:24:33.6586443Z 2023-03-31T06:24:33.6586609Z Running test_torch ... [2023-03-31 06:24:33.621815] 2023-03-31T06:24:33.6587218Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_torch.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:24:33.622157] 2023-03-31T06:24:47.9740464Z 2023-03-31T06:24:47.9742250Z Expand the folded group to see the log file of test_torch 2023-03-31T06:24:47.9743250Z ##[group]PRINTING LOG FILE of test_torch (/var/lib/jenkins/workspace/test/test-reports/test_torch_b6ffbma0.log) 2023-03-31T06:24:47.9744537Z Test results will be stored in test-reports/python-pytest/test_torch/test_torch-e4a1d68a94426a17.xml 2023-03-31T06:24:47.9745048Z ============================= test session starts ============================== 2023-03-31T06:24:47.9745727Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:24:47.9746217Z cachedir: .pytest_cache 2023-03-31T06:24:47.9746890Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:24:47.9747452Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:24:47.9748242Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:24:47.9748727Z collecting ... collected 841 items 2023-03-31T06:24:47.9833329Z Running 841 items in this shard: test/test_torch.py::TestBasicVitalSigns::test_basic_vitals, test/test_torch.py::TestBasicVitalSigns::test_basic_vitals_read_write, test/test_torch.py::TestBasicVitalSigns::test_dataloader_vitals, test/test_torch.py::TestTorch::test_RNGState, test/test_torch.py::TestTorch::test_RNGStateAliasing, test/test_torch.py::TestTorch::test_RNG_after_pickle, test/test_torch.py::TestTorch::test_Size, test/test_torch.py::TestTorch::test_Size_iter, test/test_torch.py::TestTorch::test_Size_scalar, test/test_torch.py::TestTorch::test_add_meta_scalar, test/test_torch.py::TestTorch::test_allow_tensor_metadata_change, test/test_torch.py::TestTorch::test_apply, test/test_torch.py::TestTorch::test_as_subclass, test/test_torch.py::TestTorch::test_assert_async, test/test_torch.py::TestTorch::test_backward_hooks_traverse, test/test_torch.py::TestTorch::test_batch_norm_cpu_inference, test/test_torch.py::TestTorch::test_bmm_multithreaded, test/test_torch.py::TestTorch::test_boxMullerState, test/test_torch.py::TestTorch::test_c10_layer_norm, test/test_torch.py::TestTorch::test_cat_neg_dim, test/test_torch.py::TestTorch::test_check, test/test_torch.py::TestTorch::test_chunk_neg_dim, test/test_torch.py::TestTorch::test_conj_neg_tolist, test/test_torch.py::TestTorch::test_contains, test/test_torch.py::TestTorch::test_copy_broadcast, test/test_torch.py::TestTorch::test_copy_dtypes, test/test_torch.py::TestTorch::test_copy_float16, test/test_torch.py::TestTorch::test_copy_many_to_one, test/test_torch.py::TestTorch::test_copy_transpose, test/test_torch.py::TestTorch::test_cuda_not_built, test/test_torch.py::TestTorch::test_cummax_neg_dim, test/test_torch.py::TestTorch::test_cummin_neg_dim, test/test_torch.py::TestTorch::test_cumprod_neg_dim, test/test_torch.py::TestTorch::test_cumsum_neg_dim, test/test_torch.py::TestTorch::test_cxx_flags, test/test_torch.py::TestTorch::test_dead_weak_ref, test/test_torch.py::TestTorch::test_deepcopy_gradient, test/test_torch.py::TestTorch::test_deepcopy_parameter, test/test_torch.py::TestTorch::test_deterministic_flag, test/test_torch.py::TestTorch::test_device, test/test_torch.py::TestTorch::test_dir, test/test_torch.py::TestTorch::test_doc, test/test_torch.py::TestTorch::test_doc_template, test/test_torch.py::TestTorch::test_dot_data_use, test/test_torch.py::TestTorch::test_dtype_is_signed, test/test_torch.py::TestTorch::test_element_size, test/test_torch.py::TestTorch::test_empty_meta, test/test_torch.py::TestTorch::test_empty_storage_view, test/test_torch.py::TestTorch::test_equal, test/test_torch.py::TestTorch::test_error_msg_type_translation, test/test_torch.py::TestTorch::test_fill_diagonal, test/test_torch.py::TestTorch::test_fix_weakref_no_leak, test/test_torch.py::TestTorch::test_format_scalar_meta, test/test_torch.py::TestTorch::test_from_buffer, test/test_torch.py::TestTorch::test_from_file, test/test_torch.py::TestTorch::test_gather_neg_dim, test/test_torch.py::TestTorch::test_generator_cpu, test/test_torch.py::TestTorch::test_has_internal_overlap, test/test_torch.py::TestTorch::test_has_storage, test/test_torch.py::TestTorch::test_index_add, test/test_torch.py::TestTorch::test_index_add_all_dtypes, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_neg_dim, test/test_torch.py::TestTorch::test_index_copy_neg_dim, test/test_torch.py::TestTorch::test_index_fill_neg_dim, test/test_torch.py::TestTorch::test_index_select_neg_dim, test/test_torch.py::TestTorch::test_invalid_generator_raises, test/test_torch.py::TestTorch::test_is_nonzero, test/test_torch.py::TestTorch::test_is_same_size, test/test_torch.py::TestTorch::test_iter, test/test_torch.py::TestTorch::test_kthvalue_neg_dim, test/test_torch.py::TestTorch::test_logcumsumexp_neg_dim, test/test_torch.py::TestTorch::test_manual_seed, test/test_torch.py::TestTorch::test_map, test/test_torch.py::TestTorch::test_map2, test/test_torch.py::TestTorch::test_max_neg_dim, test/test_torch.py::TestTorch::test_mean_neg_dim, test/test_torch.py::TestTorch::test_median_neg_dim, test/test_torch.py::TestTorch::test_memory_format, test/test_torch.py::TestTorch::test_memory_format_contiguous_returns_same_tensor_if_already_satisfies, test/test_torch.py::TestTorch::test_memory_format_empty, test/test_torch.py::TestTorch::test_min_neg_dim, test/test_torch.py::TestTorch::test_mode_neg_dim, test/test_torch.py::TestTorch::test_multinomial_invalid_probs, test/test_torch.py::TestTorch::test_nanmedian_neg_dim, test/test_torch.py::TestTorch::test_narrow_neg_dim, test/test_torch.py::TestTorch::test_ndim, test/test_torch.py::TestTorch::test_new, test/test_torch.py::TestTorch::test_newaxis_numpy_comparison, test/test_torch.py::TestTorch::test_newindex, test/test_torch.py::TestTorch::test_no_cuda_monkeypatch, test/test_torch.py::TestTorch::test_norm_neg_dim, test/test_torch.py::TestTorch::test_normal_shape, test/test_torch.py::TestTorch::test_numel, test/test_torch.py::TestTorch::test_parallel_info, test/test_torch.py::TestTorch::test_parsing_double, test/test_torch.py::TestTorch::test_parsing_int64, test/test_torch.py::TestTorch::test_parsing_intlist, test/test_torch.py::TestTorch::test_permute, test/test_torch.py::TestTorch::test_pickle, test/test_torch.py::TestTorch::test_pickle_dtype, test/test_torch.py::TestTorch::test_pickle_function, test/test_torch.py::TestTorch::test_pickle_parameter, test/test_torch.py::TestTorch::test_pickle_parameter_no_requires_grad, test/test_torch.py::TestTorch::test_pickle_size, test/test_torch.py::TestTorch::test_pin_memory, test/test_torch.py::TestTorch::test_print, test/test_torch.py::TestTorch::test_prod_neg_dim, test/test_torch.py::TestTorch::test_pyobj_preserved, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_renorm_neg_dim, test/test_torch.py::TestTorch::test_resurrected_weak_ref, test/test_torch.py::TestTorch::test_reversed, test/test_torch.py::TestTorch::test_scatter_neg_dim, test/test_torch.py::TestTorch::test_select_neg_dim, test/test_torch.py::TestTorch::test_set_flush_denormal, test/test_torch.py::TestTorch::test_setting_real_imag_to_a_number, test/test_torch.py::TestTorch::test_show_config, test/test_torch.py::TestTorch::test_size_neg_dim, test/test_torch.py::TestTorch::test_sizeof, test/test_torch.py::TestTorch::test_slice, test/test_torch.py::TestTorch::test_slow_test, test/test_torch.py::TestTorch::test_sobolengine_bounds, test/test_torch.py::TestTorch::test_sobolengine_bounds_scrambled, test/test_torch.py::TestTorch::test_sobolengine_continuing, test/test_torch.py::TestTorch::test_sobolengine_continuing_scrambled, test/test_torch.py::TestTorch::test_sobolengine_distribution, test/test_torch.py::TestTorch::test_sobolengine_distribution_scrambled, test/test_torch.py::TestTorch::test_sobolengine_draw, test/test_torch.py::TestTorch::test_sobolengine_draw_base2, test/test_torch.py::TestTorch::test_sobolengine_draw_base2_scrambled, test/test_torch.py::TestTorch::test_sobolengine_draw_scrambled, test/test_torch.py::TestTorch::test_sobolengine_fast_forward, test/test_torch.py::TestTorch::test_sobolengine_fast_forward_scrambled, test/test_torch.py::TestTorch::test_sobolengine_first_point, test/test_torch.py::TestTorch::test_sobolengine_high_dim, test/test_torch.py::TestTorch::test_sobolengine_raise, test/test_torch.py::TestTorch::test_sobolengine_reset, test/test_torch.py::TestTorch::test_sobolengine_reset_scrambled, test/test_torch.py::TestTorch::test_sort_neg_dim, test/test_torch.py::TestTorch::test_split_neg_dim, test/test_torch.py::TestTorch::test_squeeze_neg_dim, test/test_torch.py::TestTorch::test_std_neg_dim, test/test_torch.py::TestTorch::test_storage_casts, test/test_torch.py::TestTorch::test_storage_error, test/test_torch.py::TestTorch::test_storage_error_no_attribute, test/test_torch.py::TestTorch::test_structseq_repr, test/test_torch.py::TestTorch::test_subclass_preserved, test/test_torch.py::TestTorch::test_subclass_tensors, test/test_torch.py::TestTorch::test_sum_neg_dim, test/test_torch.py::TestTorch::test_t_not_2d_error, test/test_torch.py::TestTorch::test_tensor_base_init, test/test_torch.py::TestTorch::test_tensor_base_new, test/test_torch.py::TestTorch::test_tensor_ctor_scalar, test/test_torch.py::TestTorch::test_tensor_cycle_via_dict, test/test_torch.py::TestTorch::test_tensor_cycle_via_slots, test/test_torch.py::TestTorch::test_tensor_dict_dealloc, test/test_torch.py::TestTorch::test_tensor_finalizer_dealloc, test/test_torch.py::TestTorch::test_tensor_set, test/test_torch.py::TestTorch::test_tensor_set_errors, test/test_torch.py::TestTorch::test_tensor_slot_dealloc, test/test_torch.py::TestTorch::test_tensor_weakref_dealloc, test/test_torch.py::TestTorch::test_tensor_where_scalar, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_to, test/test_torch.py::TestTorch::test_to_with_tensor, test/test_torch.py::TestTorch::test_topk_neg_dim, test/test_torch.py::TestTorch::test_torch_from_file, test/test_torch.py::TestTorch::test_transpose_neg_dim, test/test_torch.py::TestTorch::test_type, test/test_torch.py::TestTorch::test_type_alias, test/test_torch.py::TestTorch::test_type_conversion_via_dtype_name, test/test_torch.py::TestTorch::test_typed_storage_deprecation_warning, test/test_torch.py::TestTorch::test_typed_storage_internal_no_warning, test/test_torch.py::TestTorch::test_unbind_neg_dim, test/test_torch.py::TestTorch::test_unflatten, test/test_torch.py::TestTorch::test_unfold_neg_dim, test/test_torch.py::TestTorch::test_unsqueeze_neg_dim, test/test_torch.py::TestTorch::test_upsample_nearest1d_meta, test/test_torch.py::TestTorch::test_upsample_nearest2d_meta, test/test_torch.py::TestTorch::test_var_neg_dim, test/test_torch.py::TestTorch::test_warn_types, test/test_torch.py::TestTorch::test_wildcard_import, test/test_torch.py::TestVitalSignsCudaCPU::test_cuda_vitals_gpu_only_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_assertRaisesRegex_ignore_msg_non_native_device_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_edge_cases_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_edge_cases_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_bfloat16_float_copy_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_bool_tensor_value_change_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_add_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_addcdiv_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_addcmul_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_atan2_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_copy_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_dist_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_div_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_eq_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_fmod_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_ge_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_gt_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_le_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_lerp_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_lt_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_map2_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_map_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_fill_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_scatter_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_select_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_max_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_min_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_mul_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_ne_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_pow_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_remainder_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_sub_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_no_inf_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_no_inf_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_cuda_backward_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_euclidean_large_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_grad_p_lt_1_no_nan_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_large_batch_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_large_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_non_contiguous_batch_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_non_contiguous_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_norm_batch_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_norm_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cdist_same_inputs_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_check_tensor_all_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_check_tensor_internal_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_clone_all_dtypes_and_devices_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_clone_not_memory_dense_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_clone_zero_stride_dim_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_complex_half_experimental_warning_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_constants_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_conv_transposed_backward_agnostic_to_memory_format_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_conv_transposed_large_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex32, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_all_dtypes_and_devices_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_math_view_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_mem_overlap_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_cpp_warnings_have_python_context_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cublas_config_nondeterministic_alert_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cummax_cummin_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cummax_discontiguous_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cummin_discontiguous_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cumprod_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_cumsum_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_scalar_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_scalar_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_device_guard_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_dim_function_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_discontiguous_out_cumsum_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_dist_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_errors_index_copy_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_expected_failure_xla_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_no_zero_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_exponential_no_zero_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_gather_backward_deterministic_path_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_gather_backward_one_dim_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_gradient_type_promotion_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_hook_remove_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_add_deterministic_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_add_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_deterministic_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_put_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_put_non_accumulate_deterministic_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_invalid_shapes_grid_sampler_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_is_set_to_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_is_signed_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex32, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_large_cumprod_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_large_cumsum_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_logcumsumexp_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_bool_tensor_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bfloat16_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bfloat16_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bool_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bool_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex128_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex128_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex64_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex64_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float16_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float16_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float32_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float32_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float64_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float64_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int16_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int16_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int32_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int32_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int64_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int64_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int8_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int8_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_uint8_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_uint8_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_bool_tensor_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_large_tensor_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_discontiguous_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_clone_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_consistency_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_cpu_and_cuda_ops_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_empty_like_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_factory_like_functions_preserve_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_operators_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_preserved_after_permute_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_propagation_rules_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_to_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_type_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_type_shortcuts_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_module_share_memory_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_device_constrain_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_empty_w_replacement_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_empty_wo_replacement_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_gpu_device_constrain_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_rng_state_advance_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_narrow_copy_non_contiguous_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_narrow_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveAvgPool2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveAvgPool3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveMaxPool2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AvgPool3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_CTCLoss_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_EmbeddingBag_max_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_FractionalMaxPool2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_FractionalMaxPool3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxPool3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_NLLLoss_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad1d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad1d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_bincount_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_grid_sample_2d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_grid_sample_3d_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_histc_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_bicubic_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_bilinear_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_linear_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_trilinear_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_kthvalue_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_median_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_put_accumulate_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_put_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_nullary_op_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_pairwise_distance_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_pdist_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_pdist_norm_large_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_pickle_gradscaler_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_pin_memory_from_constructor_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_put_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_repeat_interleave_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scalar_check_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_bool_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_non_unique_index_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_one_dim_deterministic_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_to_large_input_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_bool_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_multiply_unsupported_dtypes_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_multiply_unsupported_dtypes_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_to_large_input_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_scatter_zero_size_index_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_serialization_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_shift_mem_overlap_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_skip_xla_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_all_devices_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_qint32, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_qint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_quint4x2, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_quint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_strides_propagation_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_sync_warning_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_take_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_set_errors_multigpu_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_shape_empty_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_ternary_op_mem_overlap_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_bool, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_complex128, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_complex64, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int16, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int32, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int64, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int8, test/test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_uint8, test/test_torch.py::TestTorchDeviceTypeCPU::test_unfold_all_devices_and_dtypes_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_unfold_scalars_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_bfloat16, test/test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float16, test/test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float32, test/test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float64, test/test_torch.py::TestTorchDeviceTypeCPU::test_untyped_storage_meta_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_warn_always_caught_cpu, test/test_torch.py::TestTorchDeviceTypeCPU::test_where_scalar_handcrafted_values_cpu 2023-03-31T06:24:47.9926103Z 2023-03-31T06:24:47.9926367Z test_torch.py::TestBasicVitalSigns::test_basic_vitals PASSED [ 0%] 2023-03-31T06:24:47.9926983Z test_torch.py::TestBasicVitalSigns::test_basic_vitals_read_write PASSED [ 0%] 2023-03-31T06:24:47.9927393Z test_torch.py::TestBasicVitalSigns::test_dataloader_vitals PASSED [ 0%] 2023-03-31T06:24:47.9927773Z test_torch.py::TestTorch::test_RNGState PASSED [ 0%] 2023-03-31T06:24:47.9928715Z test_torch.py::TestTorch::test_RNGStateAliasing PASSED [ 0%] 2023-03-31T06:24:47.9929045Z test_torch.py::TestTorch::test_RNG_after_pickle PASSED [ 0%] 2023-03-31T06:24:47.9939863Z test_torch.py::TestTorch::test_Size PASSED [ 0%] 2023-03-31T06:24:47.9940182Z test_torch.py::TestTorch::test_Size_iter PASSED [ 0%] 2023-03-31T06:24:47.9940487Z test_torch.py::TestTorch::test_Size_scalar PASSED [ 1%] 2023-03-31T06:24:47.9940793Z test_torch.py::TestTorch::test_add_meta_scalar PASSED [ 1%] 2023-03-31T06:24:47.9941098Z test_torch.py::TestTorch::test_allow_tensor_metadata_change PASSED [ 1%] 2023-03-31T06:24:47.9941412Z test_torch.py::TestTorch::test_apply PASSED [ 1%] 2023-03-31T06:24:47.9941714Z test_torch.py::TestTorch::test_as_subclass PASSED [ 1%] 2023-03-31T06:24:47.9942008Z test_torch.py::TestTorch::test_assert_async PASSED [ 1%] 2023-03-31T06:24:47.9942397Z test_torch.py::TestTorch::test_backward_hooks_traverse PASSED [ 1%] 2023-03-31T06:24:47.9942874Z test_torch.py::TestTorch::test_batch_norm_cpu_inference PASSED [ 1%] 2023-03-31T06:24:47.9943245Z test_torch.py::TestTorch::test_bmm_multithreaded PASSED [ 2%] 2023-03-31T06:24:47.9943542Z test_torch.py::TestTorch::test_boxMullerState PASSED [ 2%] 2023-03-31T06:24:47.9943883Z test_torch.py::TestTorch::test_c10_layer_norm SKIPPED (Pytorch is compiled without Caffe2) [ 2%] 2023-03-31T06:24:47.9944221Z test_torch.py::TestTorch::test_cat_neg_dim PASSED [ 2%] 2023-03-31T06:24:47.9944508Z test_torch.py::TestTorch::test_check PASSED [ 2%] 2023-03-31T06:24:47.9944788Z test_torch.py::TestTorch::test_chunk_neg_dim PASSED [ 2%] 2023-03-31T06:24:47.9945083Z test_torch.py::TestTorch::test_conj_neg_tolist PASSED [ 2%] 2023-03-31T06:24:47.9945377Z test_torch.py::TestTorch::test_contains PASSED [ 2%] 2023-03-31T06:24:47.9945657Z test_torch.py::TestTorch::test_copy_broadcast PASSED [ 2%] 2023-03-31T06:24:47.9945956Z test_torch.py::TestTorch::test_copy_dtypes PASSED [ 3%] 2023-03-31T06:24:47.9946253Z test_torch.py::TestTorch::test_copy_float16 PASSED [ 3%] 2023-03-31T06:24:47.9946546Z test_torch.py::TestTorch::test_copy_many_to_one PASSED [ 3%] 2023-03-31T06:24:47.9946829Z test_torch.py::TestTorch::test_copy_transpose PASSED [ 3%] 2023-03-31T06:24:47.9947126Z test_torch.py::TestTorch::test_cuda_not_built PASSED [ 3%] 2023-03-31T06:24:47.9947482Z test_torch.py::TestTorch::test_cummax_neg_dim PASSED [ 3%] 2023-03-31T06:24:47.9947764Z test_torch.py::TestTorch::test_cummin_neg_dim PASSED [ 3%] 2023-03-31T06:24:47.9948059Z test_torch.py::TestTorch::test_cumprod_neg_dim PASSED [ 3%] 2023-03-31T06:24:47.9948358Z test_torch.py::TestTorch::test_cumsum_neg_dim PASSED [ 4%] 2023-03-31T06:24:47.9948652Z test_torch.py::TestTorch::test_cxx_flags PASSED [ 4%] 2023-03-31T06:24:47.9949609Z test_torch.py::TestTorch::test_dead_weak_ref [W PyInterpreter.cpp:223] Warning: Deallocating Tensor that still has live PyObject references. This probably happened because you took out a weak reference to Tensor and didn't call _fix_weakref() after dereferencing it. Subsequent accesses to this tensor via the PyObject will now fail. (function decref) 2023-03-31T06:24:47.9950195Z PASSED [ 4%] 2023-03-31T06:24:47.9950460Z test_torch.py::TestTorch::test_deepcopy_gradient PASSED [ 4%] 2023-03-31T06:24:47.9950777Z test_torch.py::TestTorch::test_deepcopy_parameter PASSED [ 4%] 2023-03-31T06:24:47.9951075Z test_torch.py::TestTorch::test_deterministic_flag PASSED [ 4%] 2023-03-31T06:24:47.9951376Z test_torch.py::TestTorch::test_device PASSED [ 4%] 2023-03-31T06:24:47.9951667Z test_torch.py::TestTorch::test_dir PASSED [ 4%] 2023-03-31T06:24:47.9951941Z test_torch.py::TestTorch::test_doc PASSED [ 4%] 2023-03-31T06:24:47.9952231Z test_torch.py::TestTorch::test_doc_template PASSED [ 5%] 2023-03-31T06:24:47.9952529Z test_torch.py::TestTorch::test_dot_data_use PASSED [ 5%] 2023-03-31T06:24:47.9952826Z test_torch.py::TestTorch::test_dtype_is_signed PASSED [ 5%] 2023-03-31T06:24:47.9953110Z test_torch.py::TestTorch::test_element_size PASSED [ 5%] 2023-03-31T06:24:47.9953403Z test_torch.py::TestTorch::test_empty_meta PASSED [ 5%] 2023-03-31T06:24:47.9953701Z test_torch.py::TestTorch::test_empty_storage_view PASSED [ 5%] 2023-03-31T06:24:47.9953977Z test_torch.py::TestTorch::test_equal PASSED [ 5%] 2023-03-31T06:24:47.9954275Z test_torch.py::TestTorch::test_error_msg_type_translation PASSED [ 5%] 2023-03-31T06:24:47.9954643Z test_torch.py::TestTorch::test_fill_diagonal PASSED [ 6%] 2023-03-31T06:24:47.9954978Z test_torch.py::TestTorch::test_fix_weakref_no_leak PASSED [ 6%] 2023-03-31T06:24:47.9955268Z test_torch.py::TestTorch::test_format_scalar_meta PASSED [ 6%] 2023-03-31T06:24:47.9955561Z test_torch.py::TestTorch::test_from_buffer PASSED [ 6%] 2023-03-31T06:24:47.9955855Z test_torch.py::TestTorch::test_from_file PASSED [ 6%] 2023-03-31T06:24:47.9956136Z test_torch.py::TestTorch::test_gather_neg_dim PASSED [ 6%] 2023-03-31T06:24:47.9956434Z test_torch.py::TestTorch::test_generator_cpu PASSED [ 6%] 2023-03-31T06:24:47.9956736Z test_torch.py::TestTorch::test_has_internal_overlap PASSED [ 6%] 2023-03-31T06:24:47.9957036Z test_torch.py::TestTorch::test_has_storage PASSED [ 7%] 2023-03-31T06:24:47.9957316Z test_torch.py::TestTorch::test_index_add PASSED [ 7%] 2023-03-31T06:24:47.9957616Z test_torch.py::TestTorch::test_index_add_all_dtypes PASSED [ 7%] 2023-03-31T06:24:47.9958544Z test_torch.py::TestTorch::test_index_add_correctness SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/91184 for platform(s) linux, rocm, win, windows. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 7%] 2023-03-31T06:24:47.9959208Z test_torch.py::TestTorch::test_index_add_neg_dim PASSED [ 7%] 2023-03-31T06:24:47.9959503Z test_torch.py::TestTorch::test_index_copy_neg_dim PASSED [ 7%] 2023-03-31T06:24:47.9959804Z test_torch.py::TestTorch::test_index_fill_neg_dim PASSED [ 7%] 2023-03-31T06:24:47.9960108Z test_torch.py::TestTorch::test_index_select_neg_dim PASSED [ 7%] 2023-03-31T06:24:47.9960406Z test_torch.py::TestTorch::test_invalid_generator_raises PASSED [ 7%] 2023-03-31T06:24:47.9960708Z test_torch.py::TestTorch::test_is_nonzero PASSED [ 8%] 2023-03-31T06:24:47.9960995Z test_torch.py::TestTorch::test_is_same_size PASSED [ 8%] 2023-03-31T06:24:47.9961284Z test_torch.py::TestTorch::test_iter PASSED [ 8%] 2023-03-31T06:24:47.9961611Z test_torch.py::TestTorch::test_kthvalue_neg_dim PASSED [ 8%] 2023-03-31T06:24:47.9961912Z test_torch.py::TestTorch::test_logcumsumexp_neg_dim PASSED [ 8%] 2023-03-31T06:24:47.9962216Z test_torch.py::TestTorch::test_manual_seed PASSED [ 8%] 2023-03-31T06:24:47.9962488Z test_torch.py::TestTorch::test_map PASSED [ 8%] 2023-03-31T06:24:47.9962762Z test_torch.py::TestTorch::test_map2 PASSED [ 8%] 2023-03-31T06:24:47.9963243Z test_torch.py::TestTorch::test_max_neg_dim PASSED [ 9%] 2023-03-31T06:24:47.9963555Z test_torch.py::TestTorch::test_mean_neg_dim PASSED [ 9%] 2023-03-31T06:24:47.9963838Z test_torch.py::TestTorch::test_median_neg_dim PASSED [ 9%] 2023-03-31T06:24:47.9964134Z test_torch.py::TestTorch::test_memory_format PASSED [ 9%] 2023-03-31T06:24:47.9964488Z test_torch.py::TestTorch::test_memory_format_contiguous_returns_same_tensor_if_already_satisfies PASSED [ 9%] 2023-03-31T06:24:47.9964840Z test_torch.py::TestTorch::test_memory_format_empty PASSED [ 9%] 2023-03-31T06:24:47.9965145Z test_torch.py::TestTorch::test_min_neg_dim PASSED [ 9%] 2023-03-31T06:24:47.9965441Z test_torch.py::TestTorch::test_mode_neg_dim PASSED [ 9%] 2023-03-31T06:24:47.9965816Z test_torch.py::TestTorch::test_multinomial_invalid_probs SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 9%] 2023-03-31T06:24:47.9966236Z test_torch.py::TestTorch::test_nanmedian_neg_dim PASSED [ 10%] 2023-03-31T06:24:47.9966572Z test_torch.py::TestTorch::test_narrow_neg_dim PASSED [ 10%] 2023-03-31T06:24:47.9966865Z test_torch.py::TestTorch::test_ndim PASSED [ 10%] 2023-03-31T06:24:47.9967137Z test_torch.py::TestTorch::test_new PASSED [ 10%] 2023-03-31T06:24:47.9967444Z test_torch.py::TestTorch::test_newaxis_numpy_comparison PASSED [ 10%] 2023-03-31T06:24:47.9967751Z test_torch.py::TestTorch::test_newindex PASSED [ 10%] 2023-03-31T06:24:47.9968056Z test_torch.py::TestTorch::test_no_cuda_monkeypatch PASSED [ 10%] 2023-03-31T06:24:47.9968345Z test_torch.py::TestTorch::test_norm_neg_dim PASSED [ 10%] 2023-03-31T06:24:47.9968642Z test_torch.py::TestTorch::test_normal_shape PASSED [ 11%] 2023-03-31T06:24:47.9968934Z test_torch.py::TestTorch::test_numel PASSED [ 11%] 2023-03-31T06:24:47.9969213Z test_torch.py::TestTorch::test_parallel_info PASSED [ 11%] 2023-03-31T06:24:47.9969510Z test_torch.py::TestTorch::test_parsing_double PASSED [ 11%] 2023-03-31T06:24:47.9969806Z test_torch.py::TestTorch::test_parsing_int64 PASSED [ 11%] 2023-03-31T06:24:47.9970101Z test_torch.py::TestTorch::test_parsing_intlist PASSED [ 11%] 2023-03-31T06:24:47.9970385Z test_torch.py::TestTorch::test_permute PASSED [ 11%] 2023-03-31T06:24:47.9970717Z test_torch.py::TestTorch::test_pickle PASSED [ 11%] 2023-03-31T06:24:47.9971012Z test_torch.py::TestTorch::test_pickle_dtype PASSED [ 12%] 2023-03-31T06:24:47.9971295Z test_torch.py::TestTorch::test_pickle_function PASSED [ 12%] 2023-03-31T06:24:47.9971596Z test_torch.py::TestTorch::test_pickle_parameter PASSED [ 12%] 2023-03-31T06:24:47.9971914Z test_torch.py::TestTorch::test_pickle_parameter_no_requires_grad PASSED [ 12%] 2023-03-31T06:24:47.9972225Z test_torch.py::TestTorch::test_pickle_size PASSED [ 12%] 2023-03-31T06:24:47.9972503Z test_torch.py::TestTorch::test_pin_memory PASSED [ 12%] 2023-03-31T06:24:47.9972794Z test_torch.py::TestTorch::test_print PASSED [ 12%] 2023-03-31T06:24:47.9973132Z test_torch.py::TestTorch::test_prod_neg_dim PASSED [ 12%] 2023-03-31T06:24:47.9973420Z test_torch.py::TestTorch::test_pyobj_preserved PASSED [ 12%] 2023-03-31T06:24:47.9973719Z test_torch.py::TestTorch::test_qengine PASSED [ 13%] 2023-03-31T06:24:47.9974010Z test_torch.py::TestTorch::test_renorm_neg_dim PASSED [ 13%] 2023-03-31T06:24:47.9974313Z test_torch.py::TestTorch::test_resurrected_weak_ref PASSED [ 13%] 2023-03-31T06:24:47.9974599Z test_torch.py::TestTorch::test_reversed PASSED [ 13%] 2023-03-31T06:24:47.9974894Z test_torch.py::TestTorch::test_scatter_neg_dim PASSED [ 13%] 2023-03-31T06:24:47.9975191Z test_torch.py::TestTorch::test_select_neg_dim PASSED [ 13%] 2023-03-31T06:24:47.9975511Z test_torch.py::TestTorch::test_set_flush_denormal SKIPPED (flush_denormal not supported) [ 13%] 2023-03-31T06:24:47.9975856Z test_torch.py::TestTorch::test_setting_real_imag_to_a_number PASSED [ 13%] 2023-03-31T06:24:47.9976162Z test_torch.py::TestTorch::test_show_config PASSED [ 14%] 2023-03-31T06:24:47.9976454Z test_torch.py::TestTorch::test_size_neg_dim PASSED [ 14%] 2023-03-31T06:24:47.9976731Z test_torch.py::TestTorch::test_sizeof PASSED [ 14%] 2023-03-31T06:24:47.9977015Z test_torch.py::TestTorch::test_slice PASSED [ 14%] 2023-03-31T06:24:47.9977398Z test_torch.py::TestTorch::test_slow_test SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 14%] 2023-03-31T06:24:47.9977770Z test_torch.py::TestTorch::test_sobolengine_bounds PASSED [ 14%] 2023-03-31T06:24:47.9978083Z test_torch.py::TestTorch::test_sobolengine_bounds_scrambled PASSED [ 14%] 2023-03-31T06:24:47.9978397Z test_torch.py::TestTorch::test_sobolengine_continuing PASSED [ 14%] 2023-03-31T06:24:47.9978718Z test_torch.py::TestTorch::test_sobolengine_continuing_scrambled PASSED [ 14%] 2023-03-31T06:24:47.9979030Z test_torch.py::TestTorch::test_sobolengine_distribution PASSED [ 15%] 2023-03-31T06:24:47.9979352Z test_torch.py::TestTorch::test_sobolengine_distribution_scrambled PASSED [ 15%] 2023-03-31T06:24:47.9979670Z test_torch.py::TestTorch::test_sobolengine_draw PASSED [ 15%] 2023-03-31T06:24:47.9979960Z test_torch.py::TestTorch::test_sobolengine_draw_base2 PASSED [ 15%] 2023-03-31T06:24:47.9980282Z test_torch.py::TestTorch::test_sobolengine_draw_base2_scrambled PASSED [ 15%] 2023-03-31T06:24:47.9980603Z test_torch.py::TestTorch::test_sobolengine_draw_scrambled PASSED [ 15%] 2023-03-31T06:24:47.9980917Z test_torch.py::TestTorch::test_sobolengine_fast_forward PASSED [ 15%] 2023-03-31T06:24:47.9981222Z test_torch.py::TestTorch::test_sobolengine_fast_forward_scrambled PASSED [ 15%] 2023-03-31T06:24:47.9981541Z test_torch.py::TestTorch::test_sobolengine_first_point PASSED [ 16%] 2023-03-31T06:24:47.9981855Z test_torch.py::TestTorch::test_sobolengine_high_dim PASSED [ 16%] 2023-03-31T06:24:47.9982272Z test_torch.py::TestTorch::test_sobolengine_raise PASSED [ 16%] 2023-03-31T06:24:47.9982589Z test_torch.py::TestTorch::test_sobolengine_reset PASSED [ 16%] 2023-03-31T06:24:47.9982900Z test_torch.py::TestTorch::test_sobolengine_reset_scrambled PASSED [ 16%] 2023-03-31T06:24:47.9983209Z test_torch.py::TestTorch::test_sort_neg_dim PASSED [ 16%] 2023-03-31T06:24:47.9983490Z test_torch.py::TestTorch::test_split_neg_dim PASSED [ 16%] 2023-03-31T06:24:47.9983784Z test_torch.py::TestTorch::test_squeeze_neg_dim PASSED [ 16%] 2023-03-31T06:24:47.9984084Z test_torch.py::TestTorch::test_std_neg_dim PASSED [ 17%] 2023-03-31T06:24:47.9984363Z test_torch.py::TestTorch::test_storage_casts PASSED [ 17%] 2023-03-31T06:24:47.9984697Z test_torch.py::TestTorch::test_storage_error PASSED [ 17%] 2023-03-31T06:24:47.9985007Z test_torch.py::TestTorch::test_storage_error_no_attribute PASSED [ 17%] 2023-03-31T06:24:47.9985314Z test_torch.py::TestTorch::test_structseq_repr PASSED [ 17%] 2023-03-31T06:24:47.9985605Z test_torch.py::TestTorch::test_subclass_preserved PASSED [ 17%] 2023-03-31T06:24:47.9985907Z test_torch.py::TestTorch::test_subclass_tensors PASSED [ 17%] 2023-03-31T06:24:47.9986202Z test_torch.py::TestTorch::test_sum_neg_dim PASSED [ 17%] 2023-03-31T06:24:47.9986482Z test_torch.py::TestTorch::test_t_not_2d_error PASSED [ 17%] 2023-03-31T06:24:47.9986774Z test_torch.py::TestTorch::test_tensor_base_init PASSED [ 18%] 2023-03-31T06:24:47.9987069Z test_torch.py::TestTorch::test_tensor_base_new PASSED [ 18%] 2023-03-31T06:24:47.9987368Z test_torch.py::TestTorch::test_tensor_ctor_scalar PASSED [ 18%] 2023-03-31T06:24:47.9987658Z test_torch.py::TestTorch::test_tensor_cycle_via_dict PASSED [ 18%] 2023-03-31T06:24:47.9987958Z test_torch.py::TestTorch::test_tensor_cycle_via_slots PASSED [ 18%] 2023-03-31T06:24:47.9988261Z test_torch.py::TestTorch::test_tensor_dict_dealloc PASSED [ 18%] 2023-03-31T06:24:47.9988559Z test_torch.py::TestTorch::test_tensor_finalizer_dealloc PASSED [ 18%] 2023-03-31T06:24:47.9988899Z test_torch.py::TestTorch::test_tensor_set PASSED [ 18%] 2023-03-31T06:24:47.9989229Z test_torch.py::TestTorch::test_tensor_set_errors PASSED [ 19%] 2023-03-31T06:24:47.9989528Z test_torch.py::TestTorch::test_tensor_slot_dealloc PASSED [ 19%] 2023-03-31T06:24:47.9989822Z test_torch.py::TestTorch::test_tensor_weakref_dealloc PASSED [ 19%] 2023-03-31T06:24:47.9990123Z test_torch.py::TestTorch::test_tensor_where_scalar PASSED [ 19%] 2023-03-31T06:24:47.9990437Z test_torch.py::TestTorch::test_tensoriterator_output_setup PASSED [ 19%] 2023-03-31T06:24:47.9990728Z test_torch.py::TestTorch::test_to PASSED [ 19%] 2023-03-31T06:24:47.9991013Z test_torch.py::TestTorch::test_to_with_tensor PASSED [ 19%] 2023-03-31T06:24:47.9991308Z test_torch.py::TestTorch::test_topk_neg_dim PASSED [ 19%] 2023-03-31T06:24:47.9991605Z test_torch.py::TestTorch::test_torch_from_file PASSED [ 19%] 2023-03-31T06:24:47.9991891Z test_torch.py::TestTorch::test_transpose_neg_dim PASSED [ 20%] 2023-03-31T06:24:47.9992187Z test_torch.py::TestTorch::test_type PASSED [ 20%] 2023-03-31T06:24:47.9992479Z test_torch.py::TestTorch::test_type_alias PASSED [ 20%] 2023-03-31T06:24:47.9992772Z test_torch.py::TestTorch::test_type_conversion_via_dtype_name PASSED [ 20%] 2023-03-31T06:24:47.9993095Z test_torch.py::TestTorch::test_typed_storage_deprecation_warning PASSED [ 20%] 2023-03-31T06:24:47.9993428Z test_torch.py::TestTorch::test_typed_storage_internal_no_warning PASSED [ 20%] 2023-03-31T06:24:47.9993773Z test_torch.py::TestTorch::test_unbind_neg_dim PASSED [ 20%] 2023-03-31T06:24:47.9994056Z test_torch.py::TestTorch::test_unflatten PASSED [ 20%] 2023-03-31T06:24:47.9994351Z test_torch.py::TestTorch::test_unfold_neg_dim PASSED [ 21%] 2023-03-31T06:24:47.9994650Z test_torch.py::TestTorch::test_unsqueeze_neg_dim PASSED [ 21%] 2023-03-31T06:24:47.9994943Z test_torch.py::TestTorch::test_upsample_nearest1d_meta PASSED [ 21%] 2023-03-31T06:24:47.9995254Z test_torch.py::TestTorch::test_upsample_nearest2d_meta PASSED [ 21%] 2023-03-31T06:24:47.9995551Z test_torch.py::TestTorch::test_var_neg_dim PASSED [ 21%] 2023-03-31T06:24:47.9995844Z test_torch.py::TestTorch::test_warn_types PASSED [ 21%] 2023-03-31T06:24:47.9996165Z test_torch.py::TestTorch::test_wildcard_import PASSED [ 21%] 2023-03-31T06:24:47.9996520Z test_torch.py::TestVitalSignsCudaCPU::test_cuda_vitals_gpu_only_cpu SKIPPED (Only runs on cuda) [ 21%] 2023-03-31T06:24:47.9996901Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_complex128 PASSED [ 21%] 2023-03-31T06:24:47.9997241Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_complex64 PASSED [ 22%] 2023-03-31T06:24:47.9997586Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_float32 PASSED [ 22%] 2023-03-31T06:24:47.9997927Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_float64 PASSED [ 22%] 2023-03-31T06:24:47.9998266Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int16 PASSED [ 22%] 2023-03-31T06:24:47.9998590Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int32 PASSED [ 22%] 2023-03-31T06:24:47.9998925Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int64 PASSED [ 22%] 2023-03-31T06:24:47.9999262Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_int8 PASSED [ 22%] 2023-03-31T06:24:47.9999589Z test_torch.py::TestTorchDeviceTypeCPU::test_addcdiv_cpu_uint8 PASSED [ 22%] 2023-03-31T06:24:47.9999935Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_complex128 PASSED [ 23%] 2023-03-31T06:24:48.0000286Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_complex64 PASSED [ 23%] 2023-03-31T06:24:48.0000630Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_float32 PASSED [ 23%] 2023-03-31T06:24:48.0000986Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_float64 PASSED [ 23%] 2023-03-31T06:24:48.0001366Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int16 PASSED [ 23%] 2023-03-31T06:24:48.0001700Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int32 PASSED [ 23%] 2023-03-31T06:24:48.0002019Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int64 PASSED [ 23%] 2023-03-31T06:24:48.0002354Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_int8 PASSED [ 23%] 2023-03-31T06:24:48.0002690Z test_torch.py::TestTorchDeviceTypeCPU::test_addcmul_cpu_uint8 PASSED [ 24%] 2023-03-31T06:24:48.0003283Z test_torch.py::TestTorchDeviceTypeCPU::test_assertRaisesRegex_ignore_msg_non_native_device_cpu PASSED [ 24%] 2023-03-31T06:24:48.0003811Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_edge_cases_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 24%] 2023-03-31T06:24:48.0004322Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_edge_cases_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 24%] 2023-03-31T06:24:48.0004752Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_mem_overlap_cpu PASSED [ 24%] 2023-03-31T06:24:48.0005126Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:24:48.0005470Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_float32 PASSED [ 24%] 2023-03-31T06:24:48.0005831Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_p_cpu_float64 PASSED [ 24%] 2023-03-31T06:24:48.0006266Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_bool PASSED [ 24%] 2023-03-31T06:24:48.0006616Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_float32 PASSED [ 25%] 2023-03-31T06:24:48.0006979Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_float64 PASSED [ 25%] 2023-03-31T06:24:48.0007343Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int16 PASSED [ 25%] 2023-03-31T06:24:48.0007707Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int32 PASSED [ 25%] 2023-03-31T06:24:48.0008054Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int64 PASSED [ 25%] 2023-03-31T06:24:48.0008417Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_int8 PASSED [ 25%] 2023-03-31T06:24:48.0008777Z test_torch.py::TestTorchDeviceTypeCPU::test_bernoulli_self_cpu_uint8 PASSED [ 25%] 2023-03-31T06:24:48.0009159Z test_torch.py::TestTorchDeviceTypeCPU::test_bfloat16_float_copy_cpu PASSED [ 25%] 2023-03-31T06:24:48.0009518Z test_torch.py::TestTorchDeviceTypeCPU::test_bool_tensor_value_change_cpu PASSED [ 26%] 2023-03-31T06:24:48.0009873Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_add_cpu PASSED [ 26%] 2023-03-31T06:24:48.0010229Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_addcdiv_cpu PASSED [ 26%] 2023-03-31T06:24:48.0010569Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_addcmul_cpu PASSED [ 26%] 2023-03-31T06:24:48.0010928Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_atan2_cpu PASSED [ 26%] 2023-03-31T06:24:48.0011289Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_copy_cpu PASSED [ 26%] 2023-03-31T06:24:48.0011629Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_dist_cpu PASSED [ 26%] 2023-03-31T06:24:48.0011982Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_div_cpu PASSED [ 26%] 2023-03-31T06:24:48.0012332Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_eq_cpu PASSED [ 26%] 2023-03-31T06:24:48.0012680Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_fmod_cpu PASSED [ 27%] 2023-03-31T06:24:48.0013013Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_ge_cpu PASSED [ 27%] 2023-03-31T06:24:48.0013355Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_gt_cpu PASSED [ 27%] 2023-03-31T06:24:48.0013700Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_le_cpu PASSED [ 27%] 2023-03-31T06:24:48.0014074Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_lerp_cpu PASSED [ 27%] 2023-03-31T06:24:48.0014456Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_lt_cpu PASSED [ 27%] 2023-03-31T06:24:48.0014801Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_map2_cpu PASSED [ 27%] 2023-03-31T06:24:48.0015147Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_map_cpu PASSED [ 27%] 2023-03-31T06:24:48.0015490Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_fill_cpu PASSED [ 28%] 2023-03-31T06:24:48.0015863Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_scatter_cpu PASSED [ 28%] 2023-03-31T06:24:48.0016237Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_masked_select_cpu PASSED [ 28%] 2023-03-31T06:24:48.0016586Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_max_cpu PASSED [ 28%] 2023-03-31T06:24:48.0016935Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_min_cpu PASSED [ 28%] 2023-03-31T06:24:48.0017288Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_mul_cpu PASSED [ 28%] 2023-03-31T06:24:48.0017635Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_ne_cpu PASSED [ 28%] 2023-03-31T06:24:48.0017969Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_pow_cpu PASSED [ 28%] 2023-03-31T06:24:48.0018330Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_remainder_cpu PASSED [ 29%] 2023-03-31T06:24:48.0018688Z test_torch.py::TestTorchDeviceTypeCPU::test_broadcast_fn_sub_cpu PASSED [ 29%] 2023-03-31T06:24:48.0019029Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_bool PASSED [ 29%] 2023-03-31T06:24:48.0019427Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_complex128 PASSED [ 29%] 2023-03-31T06:24:48.0019796Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_complex64 PASSED [ 29%] 2023-03-31T06:24:48.0020162Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_float32 PASSED [ 29%] 2023-03-31T06:24:48.0020511Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_float64 PASSED [ 29%] 2023-03-31T06:24:48.0020875Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int16 PASSED [ 29%] 2023-03-31T06:24:48.0021228Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int32 PASSED [ 29%] 2023-03-31T06:24:48.0021566Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int64 PASSED [ 30%] 2023-03-31T06:24:48.0021921Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_int8 PASSED [ 30%] 2023-03-31T06:24:48.0022403Z test_torch.py::TestTorchDeviceTypeCPU::test_bytes_to_scalar_cpu_uint8 PASSED [ 30%] 2023-03-31T06:24:48.0022761Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:24:48.0023093Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float16 PASSED [ 30%] 2023-03-31T06:24:48.0023436Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float32 PASSED [ 30%] 2023-03-31T06:24:48.0023779Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_cpu_float64 PASSED [ 30%] 2023-03-31T06:24:48.0024122Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:24:48.0024489Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float16 PASSED [ 31%] 2023-03-31T06:24:48.0024849Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float32 PASSED [ 31%] 2023-03-31T06:24:48.0025204Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_kstest_cpu_float64 PASSED [ 31%] 2023-03-31T06:24:48.0025626Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_no_inf_cpu_bfloat16 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 31%] 2023-03-31T06:24:48.0026122Z test_torch.py::TestTorchDeviceTypeCPU::test_cauchy_no_inf_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 31%] 2023-03-31T06:24:48.0026561Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_cuda_backward_cpu SKIPPED (Only runs on cuda) [ 31%] 2023-03-31T06:24:48.0026967Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_empty_cpu PASSED [ 31%] 2023-03-31T06:24:48.0027336Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_euclidean_large_cpu PASSED [ 31%] 2023-03-31T06:24:48.0027695Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_grad_p_lt_1_no_nan_cpu PASSED [ 31%] 2023-03-31T06:24:48.0028120Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_large_batch_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 32%] 2023-03-31T06:24:48.0028519Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_large_cpu PASSED [ 32%] 2023-03-31T06:24:48.0028881Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_non_contiguous_batch_cpu PASSED [ 32%] 2023-03-31T06:24:48.0029248Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_non_contiguous_cpu PASSED [ 32%] 2023-03-31T06:24:48.0029604Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_norm_batch_cpu PASSED [ 32%] 2023-03-31T06:24:48.0029934Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_norm_cpu PASSED [ 32%] 2023-03-31T06:24:48.0030278Z test_torch.py::TestTorchDeviceTypeCPU::test_cdist_same_inputs_cpu PASSED [ 32%] 2023-03-31T06:24:48.0030626Z test_torch.py::TestTorchDeviceTypeCPU::test_check_tensor_all_cpu PASSED [ 32%] 2023-03-31T06:24:48.0030965Z test_torch.py::TestTorchDeviceTypeCPU::test_check_tensor_internal_cpu PASSED [ 33%] 2023-03-31T06:24:48.0031333Z test_torch.py::TestTorchDeviceTypeCPU::test_clone_all_dtypes_and_devices_cpu PASSED [ 33%] 2023-03-31T06:24:48.0031699Z test_torch.py::TestTorchDeviceTypeCPU::test_clone_not_memory_dense_cpu PASSED [ 33%] 2023-03-31T06:24:48.0032056Z test_torch.py::TestTorchDeviceTypeCPU::test_clone_zero_stride_dim_cpu PASSED [ 33%] 2023-03-31T06:24:48.0032455Z test_torch.py::TestTorchDeviceTypeCPU::test_complex_half_experimental_warning_cpu PASSED [ 33%] 2023-03-31T06:24:48.0032822Z test_torch.py::TestTorchDeviceTypeCPU::test_constants_cpu PASSED [ 33%] 2023-03-31T06:24:48.0033206Z test_torch.py::TestTorchDeviceTypeCPU::test_conv_transposed_backward_agnostic_to_memory_format_cpu PASSED [ 33%] 2023-03-31T06:24:48.0033617Z test_torch.py::TestTorchDeviceTypeCPU::test_conv_transposed_large_cpu SKIPPED (Only runs on cuda) [ 33%] 2023-03-31T06:24:48.0033992Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:24:48.0034331Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_bool PASSED [ 34%] 2023-03-31T06:24:48.0034672Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex128 PASSED [ 34%] 2023-03-31T06:24:48.0035045Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex32 PASSED [ 34%] 2023-03-31T06:24:48.0035389Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_complex64 PASSED [ 34%] 2023-03-31T06:24:48.0035732Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float16 PASSED [ 34%] 2023-03-31T06:24:48.0036055Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float32 PASSED [ 34%] 2023-03-31T06:24:48.0036392Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_float64 PASSED [ 34%] 2023-03-31T06:24:48.0036728Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int16 PASSED [ 34%] 2023-03-31T06:24:48.0037059Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int32 PASSED [ 35%] 2023-03-31T06:24:48.0037378Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int64 PASSED [ 35%] 2023-03-31T06:24:48.0037708Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_int8 PASSED [ 35%] 2023-03-31T06:24:48.0038042Z test_torch.py::TestTorchDeviceTypeCPU::test_copy__cpu_uint8 PASSED [ 35%] 2023-03-31T06:24:48.0038382Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_all_dtypes_and_devices_cpu PASSED [ 35%] 2023-03-31T06:24:48.0038741Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_math_view_cpu PASSED [ 35%] 2023-03-31T06:24:48.0039091Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_mem_overlap_cpu_float64 PASSED [ 35%] 2023-03-31T06:24:48.0039473Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_complex64 PASSED [ 35%] 2023-03-31T06:24:48.0039888Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_float32 PASSED [ 36%] 2023-03-31T06:24:48.0040293Z test_torch.py::TestTorchDeviceTypeCPU::test_copy_transpose_math_view_cpu_int64 PASSED [ 36%] 2023-03-31T06:24:48.0040661Z test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_complex64 PASSED [ 36%] 2023-03-31T06:24:48.0041012Z test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_float32 PASSED [ 36%] 2023-03-31T06:24:48.0041347Z test_torch.py::TestTorchDeviceTypeCPU::test_corrcoef_cpu_int32 PASSED [ 36%] 2023-03-31T06:24:48.0041691Z test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_complex64 PASSED [ 36%] 2023-03-31T06:24:48.0042035Z test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_float32 PASSED [ 36%] 2023-03-31T06:24:48.0042357Z test_torch.py::TestTorchDeviceTypeCPU::test_cov_cpu_int32 PASSED [ 36%] 2023-03-31T06:24:48.0042719Z test_torch.py::TestTorchDeviceTypeCPU::test_cpp_warnings_have_python_context_cpu PASSED [ 36%] 2023-03-31T06:24:48.0043347Z test_torch.py::TestTorchDeviceTypeCPU::test_cublas_config_nondeterministic_alert_cpu SKIPPED (Only runs on cuda) [ 37%] 2023-03-31T06:24:48.0043750Z test_torch.py::TestTorchDeviceTypeCPU::test_cummax_cummin_cpu PASSED [ 37%] 2023-03-31T06:24:48.0044091Z test_torch.py::TestTorchDeviceTypeCPU::test_cummax_discontiguous_cpu PASSED [ 37%] 2023-03-31T06:24:48.0044455Z test_torch.py::TestTorchDeviceTypeCPU::test_cummin_discontiguous_cpu PASSED [ 37%] 2023-03-31T06:24:48.0044815Z test_torch.py::TestTorchDeviceTypeCPU::test_cumprod_cpu PASSED [ 37%] 2023-03-31T06:24:48.0045195Z test_torch.py::TestTorchDeviceTypeCPU::test_cumsum_cpu PASSED [ 37%] 2023-03-31T06:24:48.0045538Z test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_cpu_complex64 PASSED [ 37%] 2023-03-31T06:24:48.0045891Z test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_cpu_float32 PASSED [ 37%] 2023-03-31T06:24:48.0046256Z test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_scalar_cpu_complex64 PASSED [ 38%] 2023-03-31T06:24:48.0046614Z test_torch.py::TestTorchDeviceTypeCPU::test_deepcopy_scalar_cpu_float32 PASSED [ 38%] 2023-03-31T06:24:48.0047008Z test_torch.py::TestTorchDeviceTypeCPU::test_device_guard_cpu SKIPPED (fewer than 2 devices detected) [ 38%] 2023-03-31T06:24:48.0047385Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_bool PASSED [ 38%] 2023-03-31T06:24:48.0047713Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_complex128 PASSED [ 38%] 2023-03-31T06:24:48.0048157Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_complex64 PASSED [ 38%] 2023-03-31T06:24:48.0048501Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float16 PASSED [ 38%] 2023-03-31T06:24:48.0048841Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float32 PASSED [ 38%] 2023-03-31T06:24:48.0049165Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_float64 PASSED [ 39%] 2023-03-31T06:24:48.0049500Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int16 PASSED [ 39%] 2023-03-31T06:24:48.0049838Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int32 PASSED [ 39%] 2023-03-31T06:24:48.0050160Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int64 PASSED [ 39%] 2023-03-31T06:24:48.0050494Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_int8 PASSED [ 39%] 2023-03-31T06:24:48.0050826Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_cpu_uint8 PASSED [ 39%] 2023-03-31T06:24:48.0051175Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_bool PASSED [ 39%] 2023-03-31T06:24:48.0051527Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_complex128 PASSED [ 39%] 2023-03-31T06:24:48.0051898Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_complex64 PASSED [ 39%] 2023-03-31T06:24:48.0052263Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float16 PASSED [ 40%] 2023-03-31T06:24:48.0052605Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float32 PASSED [ 40%] 2023-03-31T06:24:48.0053004Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_float64 PASSED [ 40%] 2023-03-31T06:24:48.0053398Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int16 PASSED [ 40%] 2023-03-31T06:24:48.0053749Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int32 PASSED [ 40%] 2023-03-31T06:24:48.0054089Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int64 PASSED [ 40%] 2023-03-31T06:24:48.0054441Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_int8 PASSED [ 40%] 2023-03-31T06:24:48.0054798Z test_torch.py::TestTorchDeviceTypeCPU::test_diff_noncontig_cpu_uint8 PASSED [ 40%] 2023-03-31T06:24:48.0055141Z test_torch.py::TestTorchDeviceTypeCPU::test_dim_function_empty_cpu PASSED [ 41%] 2023-03-31T06:24:48.0055503Z test_torch.py::TestTorchDeviceTypeCPU::test_discontiguous_out_cumsum_cpu PASSED [ 41%] 2023-03-31T06:24:48.0055859Z test_torch.py::TestTorchDeviceTypeCPU::test_dist_cpu PASSED [ 41%] 2023-03-31T06:24:48.0056205Z test_torch.py::TestTorchDeviceTypeCPU::test_errors_index_copy_cpu PASSED [ 41%] 2023-03-31T06:24:48.0056550Z test_torch.py::TestTorchDeviceTypeCPU::test_expected_failure_xla_cpu PASSED [ 41%] 2023-03-31T06:24:48.0056912Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:24:48.0057265Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float16 PASSED [ 41%] 2023-03-31T06:24:48.0057609Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float32 PASSED [ 41%] 2023-03-31T06:24:48.0057964Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_cpu_float64 PASSED [ 41%] 2023-03-31T06:24:48.0058361Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:24:48.0058735Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float16 PASSED [ 42%] 2023-03-31T06:24:48.0059095Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float32 PASSED [ 42%] 2023-03-31T06:24:48.0059465Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_kstest_cpu_float64 PASSED [ 42%] 2023-03-31T06:24:48.0059862Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_no_zero_cpu_float16 SKIPPED (Only runs on cuda) [ 42%] 2023-03-31T06:24:48.0060263Z test_torch.py::TestTorchDeviceTypeCPU::test_exponential_no_zero_cpu_float32 SKIPPED (Only runs on cuda) [ 42%] 2023-03-31T06:24:48.0060667Z test_torch.py::TestTorchDeviceTypeCPU::test_gather_backward_deterministic_path_cpu PASSED [ 42%] 2023-03-31T06:24:48.0061079Z test_torch.py::TestTorchDeviceTypeCPU::test_gather_backward_one_dim_cpu PASSED [ 42%] 2023-03-31T06:24:48.0061440Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:24:48.0061778Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float16 PASSED [ 43%] 2023-03-31T06:24:48.0062128Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float32 PASSED [ 43%] 2023-03-31T06:24:48.0062555Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_float64 PASSED [ 43%] 2023-03-31T06:24:48.0062904Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int16 PASSED [ 43%] 2023-03-31T06:24:48.0063239Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int32 PASSED [ 43%] 2023-03-31T06:24:48.0063582Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int64 PASSED [ 43%] 2023-03-31T06:24:48.0063925Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_int8 PASSED [ 43%] 2023-03-31T06:24:48.0064256Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_cpu_uint8 PASSED [ 43%] 2023-03-31T06:24:48.0064615Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:24:48.0064982Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float16 PASSED [ 44%] 2023-03-31T06:24:48.0065355Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float32 PASSED [ 44%] 2023-03-31T06:24:48.0065701Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_float64 PASSED [ 44%] 2023-03-31T06:24:48.0066094Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int16 PASSED [ 44%] 2023-03-31T06:24:48.0066486Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int32 PASSED [ 44%] 2023-03-31T06:24:48.0066829Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int64 PASSED [ 44%] 2023-03-31T06:24:48.0067184Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_int8 PASSED [ 44%] 2023-03-31T06:24:48.0067542Z test_torch.py::TestTorchDeviceTypeCPU::test_geometric_kstest_cpu_uint8 PASSED [ 45%] 2023-03-31T06:24:48.0067903Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_complex64 PASSED [ 45%] 2023-03-31T06:24:48.0068245Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_float32 PASSED [ 45%] 2023-03-31T06:24:48.0068598Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_all_cpu_int64 PASSED [ 45%] 2023-03-31T06:24:48.0068969Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_complex64 PASSED [ 45%] 2023-03-31T06:24:48.0069337Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_float32 PASSED [ 45%] 2023-03-31T06:24:48.0069711Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_extreme_cases_cpu_int64 PASSED [ 45%] 2023-03-31T06:24:48.0070077Z test_torch.py::TestTorchDeviceTypeCPU::test_gradient_type_promotion_cpu PASSED [ 45%] 2023-03-31T06:24:48.0070434Z test_torch.py::TestTorchDeviceTypeCPU::test_hook_remove_cpu PASSED [ 46%] 2023-03-31T06:24:48.0070777Z test_torch.py::TestTorchDeviceTypeCPU::test_index_add_deterministic_cpu PASSED [ 46%] 2023-03-31T06:24:48.0071168Z test_torch.py::TestTorchDeviceTypeCPU::test_index_add_mem_overlap_cpu PASSED [ 46%] 2023-03-31T06:24:48.0071525Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:24:48.0071862Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_bool PASSED [ 46%] 2023-03-31T06:24:48.0072218Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_complex128 PASSED [ 46%] 2023-03-31T06:24:48.0072582Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_complex64 PASSED [ 46%] 2023-03-31T06:24:48.0072942Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float16 PASSED [ 46%] 2023-03-31T06:24:48.0073281Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float32 PASSED [ 46%] 2023-03-31T06:24:48.0073631Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_float64 PASSED [ 47%] 2023-03-31T06:24:48.0074009Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int16 PASSED [ 47%] 2023-03-31T06:24:48.0074342Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int32 PASSED [ 47%] 2023-03-31T06:24:48.0074688Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int64 PASSED [ 47%] 2023-03-31T06:24:48.0075030Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_int8 PASSED [ 47%] 2023-03-31T06:24:48.0075380Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_cpu_uint8 PASSED [ 47%] 2023-03-31T06:24:48.0075728Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_deterministic_cpu PASSED [ 47%] 2023-03-31T06:24:48.0076095Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_mem_overlap_cpu PASSED [ 47%] 2023-03-31T06:24:48.0076465Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:24:48.0076824Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_bool PASSED [ 48%] 2023-03-31T06:24:48.0077197Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_complex128 PASSED [ 48%] 2023-03-31T06:24:48.0077583Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_complex64 PASSED [ 48%] 2023-03-31T06:24:48.0077956Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float16 PASSED [ 48%] 2023-03-31T06:24:48.0078311Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float32 PASSED [ 48%] 2023-03-31T06:24:48.0078684Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_float64 PASSED [ 48%] 2023-03-31T06:24:48.0079090Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int16 PASSED [ 48%] 2023-03-31T06:24:48.0079474Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int32 PASSED [ 48%] 2023-03-31T06:24:48.0079830Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int64 PASSED [ 49%] 2023-03-31T06:24:48.0080193Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_int8 PASSED [ 49%] 2023-03-31T06:24:48.0080561Z test_torch.py::TestTorchDeviceTypeCPU::test_index_copy_scalars_cpu_uint8 PASSED [ 49%] 2023-03-31T06:24:48.0080905Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:24:48.0081255Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_bool PASSED [ 49%] 2023-03-31T06:24:48.0081609Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_complex128 PASSED [ 49%] 2023-03-31T06:24:48.0081955Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_complex64 PASSED [ 49%] 2023-03-31T06:24:48.0082313Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float16 PASSED [ 49%] 2023-03-31T06:24:48.0082667Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float32 PASSED [ 50%] 2023-03-31T06:24:48.0083176Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_float64 PASSED [ 50%] 2023-03-31T06:24:48.0083536Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int16 PASSED [ 50%] 2023-03-31T06:24:48.0083886Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int32 PASSED [ 50%] 2023-03-31T06:24:48.0084236Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int64 PASSED [ 50%] 2023-03-31T06:24:48.0084625Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_int8 PASSED [ 50%] 2023-03-31T06:24:48.0084971Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_cpu_uint8 PASSED [ 50%] 2023-03-31T06:24:48.0085327Z test_torch.py::TestTorchDeviceTypeCPU::test_index_fill_mem_overlap_cpu PASSED [ 50%] 2023-03-31T06:24:48.0085695Z test_torch.py::TestTorchDeviceTypeCPU::test_index_put_mem_overlap_cpu PASSED [ 51%] 2023-03-31T06:24:48.0086071Z test_torch.py::TestTorchDeviceTypeCPU::test_index_put_non_accumulate_deterministic_cpu PASSED [ 51%] 2023-03-31T06:24:48.0086474Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:24:48.0086861Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float16 PASSED [ 51%] 2023-03-31T06:24:48.0087284Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float32 PASSED [ 51%] 2023-03-31T06:24:48.0087653Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_float64 PASSED [ 51%] 2023-03-31T06:24:48.0088028Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int16 PASSED [ 51%] 2023-03-31T06:24:48.0088400Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int32 PASSED [ 51%] 2023-03-31T06:24:48.0088758Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int64 PASSED [ 51%] 2023-03-31T06:24:48.0089124Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_int8 PASSED [ 52%] 2023-03-31T06:24:48.0089498Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amax_cpu_uint8 PASSED [ 52%] 2023-03-31T06:24:48.0089879Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:24:48.0090245Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float16 PASSED [ 52%] 2023-03-31T06:24:48.0090625Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float32 PASSED [ 52%] 2023-03-31T06:24:48.0091007Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_float64 PASSED [ 52%] 2023-03-31T06:24:48.0091368Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int16 PASSED [ 52%] 2023-03-31T06:24:48.0091739Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int32 PASSED [ 52%] 2023-03-31T06:24:48.0092146Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int64 PASSED [ 53%] 2023-03-31T06:24:48.0092564Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_int8 PASSED [ 53%] 2023-03-31T06:24:48.0092921Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_amin_cpu_uint8 PASSED [ 53%] 2023-03-31T06:24:48.0093295Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:24:48.0093678Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float16 PASSED [ 53%] 2023-03-31T06:24:48.0094061Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float32 PASSED [ 53%] 2023-03-31T06:24:48.0094422Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_float64 PASSED [ 53%] 2023-03-31T06:24:48.0094795Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int16 PASSED [ 53%] 2023-03-31T06:24:48.0095170Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int32 PASSED [ 53%] 2023-03-31T06:24:48.0095531Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int64 PASSED [ 54%] 2023-03-31T06:24:48.0095896Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_int8 PASSED [ 54%] 2023-03-31T06:24:48.0096270Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_mean_cpu_uint8 PASSED [ 54%] 2023-03-31T06:24:48.0096647Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:24:48.0097047Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float16 PASSED [ 54%] 2023-03-31T06:24:48.0097426Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float32 PASSED [ 54%] 2023-03-31T06:24:48.0097801Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_float64 PASSED [ 54%] 2023-03-31T06:24:48.0098176Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int16 PASSED [ 54%] 2023-03-31T06:24:48.0098537Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int32 PASSED [ 55%] 2023-03-31T06:24:48.0098905Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int64 PASSED [ 55%] 2023-03-31T06:24:48.0099274Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_int8 PASSED [ 55%] 2023-03-31T06:24:48.0099630Z test_torch.py::TestTorchDeviceTypeCPU::test_index_reduce_reduce_prod_cpu_uint8 PASSED [ 55%] 2023-03-31T06:24:48.0100023Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:24:48.0100380Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_bool PASSED [ 55%] 2023-03-31T06:24:48.0100736Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_complex128 PASSED [ 55%] 2023-03-31T06:24:48.0101083Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_complex64 PASSED [ 55%] 2023-03-31T06:24:48.0101443Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float16 PASSED [ 56%] 2023-03-31T06:24:48.0101799Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float32 PASSED [ 56%] 2023-03-31T06:24:48.0102205Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_float64 PASSED [ 56%] 2023-03-31T06:24:48.0102567Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int16 PASSED [ 56%] 2023-03-31T06:24:48.0102919Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int32 PASSED [ 56%] 2023-03-31T06:24:48.0103274Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int64 PASSED [ 56%] 2023-03-31T06:24:48.0103613Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_int8 PASSED [ 56%] 2023-03-31T06:24:48.0103966Z test_torch.py::TestTorchDeviceTypeCPU::test_index_select_cpu_uint8 PASSED [ 56%] 2023-03-31T06:24:48.0104330Z test_torch.py::TestTorchDeviceTypeCPU::test_invalid_shapes_grid_sampler_cpu PASSED [ 56%] 2023-03-31T06:24:48.0104710Z test_torch.py::TestTorchDeviceTypeCPU::test_is_set_to_cpu PASSED [ 57%] 2023-03-31T06:24:48.0105080Z test_torch.py::TestTorchDeviceTypeCPU::test_is_signed_cpu PASSED [ 57%] 2023-03-31T06:24:48.0105417Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_bfloat16 PASSED [ 57%] 2023-03-31T06:24:48.0105753Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_bool PASSED [ 57%] 2023-03-31T06:24:48.0106076Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex128 PASSED [ 57%] 2023-03-31T06:24:48.0106421Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex32 PASSED [ 57%] 2023-03-31T06:24:48.0106766Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_complex64 PASSED [ 57%] 2023-03-31T06:24:48.0107095Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float16 PASSED [ 57%] 2023-03-31T06:24:48.0107429Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float32 PASSED [ 58%] 2023-03-31T06:24:48.0107764Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_float64 PASSED [ 58%] 2023-03-31T06:24:48.0108099Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int16 PASSED [ 58%] 2023-03-31T06:24:48.0108417Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int32 PASSED [ 58%] 2023-03-31T06:24:48.0108743Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int64 PASSED [ 58%] 2023-03-31T06:24:48.0109073Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_int8 PASSED [ 58%] 2023-03-31T06:24:48.0109391Z test_torch.py::TestTorchDeviceTypeCPU::test_item_cpu_uint8 PASSED [ 58%] 2023-03-31T06:24:48.0109794Z test_torch.py::TestTorchDeviceTypeCPU::test_large_cumprod_cpu_float16 SKIPPED (Only runs on cuda) [ 58%] 2023-03-31T06:24:48.0110197Z test_torch.py::TestTorchDeviceTypeCPU::test_large_cumsum_cpu_float16 SKIPPED (Only runs on cuda) [ 58%] 2023-03-31T06:24:48.0110572Z test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_bfloat16 PASSED [ 59%] 2023-03-31T06:24:48.0110915Z test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float16 PASSED [ 59%] 2023-03-31T06:24:48.0111265Z test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float32 PASSED [ 59%] 2023-03-31T06:24:48.0111614Z test_torch.py::TestTorchDeviceTypeCPU::test_log_normal_cpu_float64 PASSED [ 59%] 2023-03-31T06:24:48.0111947Z test_torch.py::TestTorchDeviceTypeCPU::test_logcumsumexp_cpu PASSED [ 59%] 2023-03-31T06:24:48.0112305Z test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_bfloat16 PASSED [ 59%] 2023-03-31T06:24:48.0112702Z test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float16 PASSED [ 59%] 2023-03-31T06:24:48.0113067Z test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float32 PASSED [ 59%] 2023-03-31T06:24:48.0113415Z test_torch.py::TestTorchDeviceTypeCPU::test_lognormal_kstest_cpu_float64 PASSED [ 60%] 2023-03-31T06:24:48.0113774Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_bool_tensor_cpu PASSED [ 60%] 2023-03-31T06:24:48.0114135Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bfloat16_bool PASSED [ 60%] 2023-03-31T06:24:48.0114488Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bfloat16_uint8 PASSED [ 60%] 2023-03-31T06:24:48.0114847Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bool_bool PASSED [ 60%] 2023-03-31T06:24:48.0115203Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_bool_uint8 PASSED [ 60%] 2023-03-31T06:24:48.0115567Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex128_bool PASSED [ 60%] 2023-03-31T06:24:48.0115924Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex128_uint8 PASSED [ 60%] 2023-03-31T06:24:48.0116294Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex64_bool PASSED [ 60%] 2023-03-31T06:24:48.0116661Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_complex64_uint8 PASSED [ 61%] 2023-03-31T06:24:48.0117014Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float16_bool PASSED [ 61%] 2023-03-31T06:24:48.0117405Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float16_uint8 PASSED [ 61%] 2023-03-31T06:24:48.0117797Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float32_bool PASSED [ 61%] 2023-03-31T06:24:48.0118159Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float32_uint8 PASSED [ 61%] 2023-03-31T06:24:48.0118507Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float64_bool PASSED [ 61%] 2023-03-31T06:24:48.0118866Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_float64_uint8 PASSED [ 61%] 2023-03-31T06:24:48.0119229Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int16_bool PASSED [ 61%] 2023-03-31T06:24:48.0119587Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int16_uint8 PASSED [ 62%] 2023-03-31T06:24:48.0119928Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int32_bool PASSED [ 62%] 2023-03-31T06:24:48.0120281Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int32_uint8 PASSED [ 62%] 2023-03-31T06:24:48.0120639Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int64_bool PASSED [ 62%] 2023-03-31T06:24:48.0120978Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int64_uint8 PASSED [ 62%] 2023-03-31T06:24:48.0121334Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int8_bool PASSED [ 62%] 2023-03-31T06:24:48.0121688Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_int8_uint8 PASSED [ 62%] 2023-03-31T06:24:48.0122044Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_uint8_bool PASSED [ 62%] 2023-03-31T06:24:48.0122390Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_cpu_uint8_uint8 PASSED [ 63%] 2023-03-31T06:24:48.0122781Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_fill_mem_overlap_cpu PASSED [ 63%] 2023-03-31T06:24:48.0123326Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_bool_tensor_cpu PASSED [ 63%] 2023-03-31T06:24:48.0123682Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:24:48.0124050Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_bool PASSED [ 63%] 2023-03-31T06:24:48.0124417Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_complex128 PASSED [ 63%] 2023-03-31T06:24:48.0124789Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_complex64 PASSED [ 63%] 2023-03-31T06:24:48.0125138Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float16 PASSED [ 63%] 2023-03-31T06:24:48.0125495Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float32 PASSED [ 63%] 2023-03-31T06:24:48.0125912Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_float64 PASSED [ 64%] 2023-03-31T06:24:48.0126258Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int16 PASSED [ 64%] 2023-03-31T06:24:48.0126610Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int32 PASSED [ 64%] 2023-03-31T06:24:48.0126966Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int64 PASSED [ 64%] 2023-03-31T06:24:48.0127321Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_int8 PASSED [ 64%] 2023-03-31T06:24:48.0127664Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_cpu_uint8 PASSED [ 64%] 2023-03-31T06:24:48.0128051Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_large_tensor_cpu SKIPPED (Only runs on cuda) [ 64%] 2023-03-31T06:24:48.0128444Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_scatter_mem_overlap_cpu PASSED [ 64%] 2023-03-31T06:24:48.0128801Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:24:48.0129158Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_bool PASSED [ 65%] 2023-03-31T06:24:48.0129522Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_complex128 PASSED [ 65%] 2023-03-31T06:24:48.0129888Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_complex64 PASSED [ 65%] 2023-03-31T06:24:48.0130235Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float16 PASSED [ 65%] 2023-03-31T06:24:48.0130634Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float32 PASSED [ 65%] 2023-03-31T06:24:48.0131023Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_float64 PASSED [ 65%] 2023-03-31T06:24:48.0131361Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int16 PASSED [ 65%] 2023-03-31T06:24:48.0131711Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int32 PASSED [ 65%] 2023-03-31T06:24:48.0132059Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int64 PASSED [ 66%] 2023-03-31T06:24:48.0132418Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_int8 PASSED [ 66%] 2023-03-31T06:24:48.0132758Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_cpu_uint8 PASSED [ 66%] 2023-03-31T06:24:48.0133118Z test_torch.py::TestTorchDeviceTypeCPU::test_masked_select_discontiguous_cpu PASSED [ 66%] 2023-03-31T06:24:48.0133483Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_clone_cpu PASSED [ 66%] 2023-03-31T06:24:48.0133832Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_consistency_cpu PASSED [ 66%] 2023-03-31T06:24:48.0134236Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_cpu_and_cuda_ops_cpu SKIPPED (Only runs on cuda) [ 66%] 2023-03-31T06:24:48.0134632Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_empty_like_cpu PASSED [ 66%] 2023-03-31T06:24:48.0135028Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_factory_like_functions_preserve_cpu PASSED [ 67%] 2023-03-31T06:24:48.0135404Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_operators_cpu PASSED [ 67%] 2023-03-31T06:24:48.0135822Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_preserved_after_permute_cpu PASSED [ 67%] 2023-03-31T06:24:48.0136215Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_propagation_rules_cpu PASSED [ 67%] 2023-03-31T06:24:48.0136579Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_to_cpu PASSED [ 67%] 2023-03-31T06:24:48.0136917Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_type_cpu PASSED [ 67%] 2023-03-31T06:24:48.0145608Z test_torch.py::TestTorchDeviceTypeCPU::test_memory_format_type_shortcuts_cpu PASSED [ 67%] 2023-03-31T06:24:48.0151183Z test_torch.py::TestTorchDeviceTypeCPU::test_module_share_memory_cpu SKIPPED (Only runs on cuda) [ 67%] 2023-03-31T06:24:48.0151583Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:24:48.0151962Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_float32 PASSED [ 68%] 2023-03-31T06:24:48.0152397Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_cpu_float64 PASSED [ 68%] 2023-03-31T06:24:48.0152769Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_float32 PASSED [ 68%] 2023-03-31T06:24:48.0153129Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_cpu_float64 PASSED [ 68%] 2023-03-31T06:24:48.0153545Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float16 SKIPPED (Only runs on cuda) [ 68%] 2023-03-31T06:24:48.0153971Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float32 SKIPPED (Only runs on cuda) [ 68%] 2023-03-31T06:24:48.0154567Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_deterministic_cpu_float64 SKIPPED (Only runs on cuda) [ 68%] 2023-03-31T06:24:48.0155227Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_device_constrain_cpu SKIPPED (Only runs on cuda) [ 68%] 2023-03-31T06:24:48.0155866Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_empty_w_replacement_cpu PASSED [ 69%] 2023-03-31T06:24:48.0156511Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_empty_wo_replacement_cpu PASSED [ 69%] 2023-03-31T06:24:48.0157237Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_gpu_device_constrain_cpu SKIPPED (fewer than 2 devices detected) [ 69%] 2023-03-31T06:24:48.0158083Z test_torch.py::TestTorchDeviceTypeCPU::test_multinomial_rng_state_advance_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 69%] 2023-03-31T06:24:48.0158865Z test_torch.py::TestTorchDeviceTypeCPU::test_narrow_copy_non_contiguous_cpu PASSED [ 69%] 2023-03-31T06:24:48.0159433Z test_torch.py::TestTorchDeviceTypeCPU::test_narrow_empty_cpu PASSED [ 69%] 2023-03-31T06:24:48.0159837Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveAvgPool2d_cpu PASSED [ 69%] 2023-03-31T06:24:48.0160250Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveAvgPool3d_cpu PASSED [ 69%] 2023-03-31T06:24:48.0160675Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AdaptiveMaxPool2d_cpu PASSED [ 70%] 2023-03-31T06:24:48.0161085Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_AvgPool3d_cpu PASSED [ 70%] 2023-03-31T06:24:48.0161476Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_CTCLoss_cpu PASSED [ 70%] 2023-03-31T06:24:48.0161867Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_EmbeddingBag_max_cpu PASSED [ 70%] 2023-03-31T06:24:48.0162288Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_FractionalMaxPool2d_cpu PASSED [ 70%] 2023-03-31T06:24:48.0162717Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_FractionalMaxPool3d_cpu PASSED [ 70%] 2023-03-31T06:24:48.0163375Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxPool3d_cpu PASSED [ 70%] 2023-03-31T06:24:48.0163837Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float16 SKIPPED (float16 not implemented on CPU) [ 70%] 2023-03-31T06:24:48.0164294Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float32 PASSED [ 70%] 2023-03-31T06:24:48.0164839Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool1d_cpu_float64 PASSED [ 71%] 2023-03-31T06:24:48.0165282Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float16 SKIPPED (float16 not implemented on CPU) [ 71%] 2023-03-31T06:24:48.0165731Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float32 PASSED [ 71%] 2023-03-31T06:24:48.0166206Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool2d_cpu_float64 PASSED [ 71%] 2023-03-31T06:24:48.0166691Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float16 SKIPPED (float16 not implemented on CPU) [ 71%] 2023-03-31T06:24:48.0167124Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float32 PASSED [ 71%] 2023-03-31T06:24:48.0167583Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_MaxUnpool3d_cpu_float64 PASSED [ 71%] 2023-03-31T06:24:48.0167983Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_NLLLoss_cpu PASSED [ 71%] 2023-03-31T06:24:48.0168386Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad1d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0168780Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad2d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0169187Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReflectionPad3d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0169602Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad1d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0170003Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad2d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0170414Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_ReplicationPad3d_cpu PASSED [ 72%] 2023-03-31T06:24:48.0170822Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_bincount_cpu PASSED [ 72%] 2023-03-31T06:24:48.0171221Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_bool PASSED [ 72%] 2023-03-31T06:24:48.0171615Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_complex128 PASSED [ 73%] 2023-03-31T06:24:48.0172020Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_complex64 PASSED [ 73%] 2023-03-31T06:24:48.0172420Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_float32 PASSED [ 73%] 2023-03-31T06:24:48.0172818Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_float64 PASSED [ 73%] 2023-03-31T06:24:48.0173206Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int16 PASSED [ 73%] 2023-03-31T06:24:48.0173596Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int32 PASSED [ 73%] 2023-03-31T06:24:48.0173992Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int64 PASSED [ 73%] 2023-03-31T06:24:48.0174373Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_int8 PASSED [ 73%] 2023-03-31T06:24:48.0174764Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_cumsum_cpu_uint8 PASSED [ 73%] 2023-03-31T06:24:48.0175162Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_grid_sample_2d_cpu PASSED [ 74%] 2023-03-31T06:24:48.0175560Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_grid_sample_3d_cpu PASSED [ 74%] 2023-03-31T06:24:48.0175942Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_histc_cpu PASSED [ 74%] 2023-03-31T06:24:48.0176347Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_bicubic_cpu PASSED [ 74%] 2023-03-31T06:24:48.0176767Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_bilinear_cpu PASSED [ 74%] 2023-03-31T06:24:48.0177187Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_linear_cpu PASSED [ 74%] 2023-03-31T06:24:48.0177624Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_interpolate_trilinear_cpu PASSED [ 74%] 2023-03-31T06:24:48.0178036Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_kthvalue_cpu_float64 PASSED [ 74%] 2023-03-31T06:24:48.0178438Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_median_cpu_float64 PASSED [ 75%] 2023-03-31T06:24:48.0178819Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_put_accumulate_cpu PASSED [ 75%] 2023-03-31T06:24:48.0179248Z test_torch.py::TestTorchDeviceTypeCPU::test_nondeterministic_alert_put_cpu PASSED [ 75%] 2023-03-31T06:24:48.0179649Z test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float16 PASSED [ 75%] 2023-03-31T06:24:48.0180006Z test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float32 PASSED [ 75%] 2023-03-31T06:24:48.0180376Z test_torch.py::TestTorchDeviceTypeCPU::test_normal_kstest_cpu_float64 PASSED [ 75%] 2023-03-31T06:24:48.0180733Z test_torch.py::TestTorchDeviceTypeCPU::test_nullary_op_mem_overlap_cpu PASSED [ 75%] 2023-03-31T06:24:48.0181094Z test_torch.py::TestTorchDeviceTypeCPU::test_pairwise_distance_empty_cpu PASSED [ 75%] 2023-03-31T06:24:48.0181453Z test_torch.py::TestTorchDeviceTypeCPU::test_pdist_empty_cpu PASSED [ 75%] 2023-03-31T06:24:48.0181812Z test_torch.py::TestTorchDeviceTypeCPU::test_pdist_norm_large_cpu SKIPPED (Only runs on cuda) [ 76%] 2023-03-31T06:24:48.0182277Z test_torch.py::TestTorchDeviceTypeCPU::test_pickle_gradscaler_cpu PASSED [ 76%] 2023-03-31T06:24:48.0182673Z test_torch.py::TestTorchDeviceTypeCPU::test_pin_memory_from_constructor_cpu SKIPPED (Only runs on cuda) [ 76%] 2023-03-31T06:24:48.0183053Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:24:48.0183423Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_complex128 PASSED [ 76%] 2023-03-31T06:24:48.0183799Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_complex64 PASSED [ 76%] 2023-03-31T06:24:48.0184170Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float16 PASSED [ 76%] 2023-03-31T06:24:48.0184519Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float32 PASSED [ 76%] 2023-03-31T06:24:48.0184878Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_float64 PASSED [ 77%] 2023-03-31T06:24:48.0185241Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int16 PASSED [ 77%] 2023-03-31T06:24:48.0185587Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int32 PASSED [ 77%] 2023-03-31T06:24:48.0185944Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int64 PASSED [ 77%] 2023-03-31T06:24:48.0186301Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_int8 PASSED [ 77%] 2023-03-31T06:24:48.0186657Z test_torch.py::TestTorchDeviceTypeCPU::test_put_accumulate_cpu_uint8 PASSED [ 77%] 2023-03-31T06:24:48.0186995Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_bfloat16 PASSED [ 77%] 2023-03-31T06:24:48.0187341Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_complex128 PASSED [ 77%] 2023-03-31T06:24:48.0187686Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_complex64 PASSED [ 78%] 2023-03-31T06:24:48.0188012Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float16 PASSED [ 78%] 2023-03-31T06:24:48.0188346Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float32 PASSED [ 78%] 2023-03-31T06:24:48.0188680Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_float64 PASSED [ 78%] 2023-03-31T06:24:48.0189016Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int16 PASSED [ 78%] 2023-03-31T06:24:48.0189335Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int32 PASSED [ 78%] 2023-03-31T06:24:48.0189662Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int64 PASSED [ 78%] 2023-03-31T06:24:48.0189994Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_int8 PASSED [ 78%] 2023-03-31T06:24:48.0190355Z test_torch.py::TestTorchDeviceTypeCPU::test_put_cpu_uint8 PASSED [ 78%] 2023-03-31T06:24:48.0190685Z test_torch.py::TestTorchDeviceTypeCPU::test_put_empty_cpu PASSED [ 79%] 2023-03-31T06:24:48.0191025Z test_torch.py::TestTorchDeviceTypeCPU::test_put_mem_overlap_cpu PASSED [ 79%] 2023-03-31T06:24:48.0191370Z test_torch.py::TestTorchDeviceTypeCPU::test_repeat_interleave_cpu PASSED [ 79%] 2023-03-31T06:24:48.0191698Z test_torch.py::TestTorchDeviceTypeCPU::test_scalar_check_cpu PASSED [ 79%] 2023-03-31T06:24:48.0192067Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_bool_cpu PASSED [ 79%] 2023-03-31T06:24:48.0192456Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_non_unique_index_cpu PASSED [ 79%] 2023-03-31T06:24:48.0192824Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_one_dim_deterministic_cpu PASSED [ 79%] 2023-03-31T06:24:48.0193237Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_add_to_large_input_cpu PASSED [ 79%] 2023-03-31T06:24:48.0193595Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_bool_cpu PASSED [ 80%] 2023-03-31T06:24:48.0193943Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_mem_overlap_cpu PASSED [ 80%] 2023-03-31T06:24:48.0194353Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_multiply_unsupported_dtypes_cpu_complex128 SKIPPED (Only runs on cuda) [ 80%] 2023-03-31T06:24:48.0194832Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_multiply_unsupported_dtypes_cpu_complex64 SKIPPED (Only runs on cuda) [ 80%] 2023-03-31T06:24:48.0195271Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:24:48.0195669Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_bool PASSED [ 80%] 2023-03-31T06:24:48.0196053Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_complex128 PASSED [ 80%] 2023-03-31T06:24:48.0196464Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_complex64 PASSED [ 80%] 2023-03-31T06:24:48.0196864Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float16 PASSED [ 80%] 2023-03-31T06:24:48.0197243Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float32 PASSED [ 81%] 2023-03-31T06:24:48.0197626Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_float64 PASSED [ 81%] 2023-03-31T06:24:48.0198017Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int16 PASSED [ 81%] 2023-03-31T06:24:48.0198403Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int32 PASSED [ 81%] 2023-03-31T06:24:48.0198772Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int64 PASSED [ 81%] 2023-03-31T06:24:48.0199163Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_int8 PASSED [ 81%] 2023-03-31T06:24:48.0199555Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_non_unique_index_cpu_uint8 PASSED [ 81%] 2023-03-31T06:24:48.0199965Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:24:48.0200370Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_bool PASSED [ 82%] 2023-03-31T06:24:48.0200796Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_complex128 PASSED [ 82%] 2023-03-31T06:24:48.0201231Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_complex64 PASSED [ 82%] 2023-03-31T06:24:48.0201661Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float16 PASSED [ 82%] 2023-03-31T06:24:48.0202069Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float32 PASSED [ 82%] 2023-03-31T06:24:48.0202488Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_float64 PASSED [ 82%] 2023-03-31T06:24:48.0202962Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int16 PASSED [ 82%] 2023-03-31T06:24:48.0203601Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int32 PASSED [ 82%] 2023-03-31T06:24:48.0204013Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int64 PASSED [ 82%] 2023-03-31T06:24:48.0204428Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_int8 PASSED [ 83%] 2023-03-31T06:24:48.0204909Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_operations_to_large_input_cpu_uint8 PASSED [ 83%] 2023-03-31T06:24:48.0205333Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:24:48.0205712Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_bool PASSED [ 83%] 2023-03-31T06:24:48.0206139Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_complex128 PASSED [ 83%] 2023-03-31T06:24:48.0206527Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_complex64 PASSED [ 83%] 2023-03-31T06:24:48.0206891Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float16 PASSED [ 83%] 2023-03-31T06:24:48.0207265Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float32 PASSED [ 83%] 2023-03-31T06:24:48.0207634Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_float64 PASSED [ 84%] 2023-03-31T06:24:48.0207989Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int16 PASSED [ 84%] 2023-03-31T06:24:48.0208356Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int32 PASSED [ 84%] 2023-03-31T06:24:48.0208720Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int64 PASSED [ 84%] 2023-03-31T06:24:48.0209093Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_int8 PASSED [ 84%] 2023-03-31T06:24:48.0209450Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_reduce_scalar_cpu_uint8 PASSED [ 84%] 2023-03-31T06:24:48.0209812Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_to_large_input_cpu PASSED [ 84%] 2023-03-31T06:24:48.0210175Z test_torch.py::TestTorchDeviceTypeCPU::test_scatter_zero_size_index_cpu PASSED [ 84%] 2023-03-31T06:24:48.0210539Z test_torch.py::TestTorchDeviceTypeCPU::test_serialization_cpu SKIPPED (Only runs on cuda) [ 85%] 2023-03-31T06:24:48.0210908Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:24:48.0211267Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_bool PASSED [ 85%] 2023-03-31T06:24:48.0211627Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_complex128 PASSED [ 85%] 2023-03-31T06:24:48.0211974Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_complex64 PASSED [ 85%] 2023-03-31T06:24:48.0212334Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float16 PASSED [ 85%] 2023-03-31T06:24:48.0212691Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float32 PASSED [ 85%] 2023-03-31T06:24:48.0213026Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_float64 PASSED [ 85%] 2023-03-31T06:24:48.0213374Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int16 PASSED [ 85%] 2023-03-31T06:24:48.0213721Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int32 PASSED [ 86%] 2023-03-31T06:24:48.0214064Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int64 PASSED [ 86%] 2023-03-31T06:24:48.0214396Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_int8 PASSED [ 86%] 2023-03-31T06:24:48.0214736Z test_torch.py::TestTorchDeviceTypeCPU::test_set_storage_cpu_uint8 PASSED [ 86%] 2023-03-31T06:24:48.0215081Z test_torch.py::TestTorchDeviceTypeCPU::test_shift_mem_overlap_cpu PASSED [ 86%] 2023-03-31T06:24:48.0215407Z test_torch.py::TestTorchDeviceTypeCPU::test_skip_xla_cpu PASSED [ 86%] 2023-03-31T06:24:48.0215817Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_all_devices_cpu SKIPPED (Only runs on cuda) [ 86%] 2023-03-31T06:24:48.0216184Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_bool PASSED [ 86%] 2023-03-31T06:24:48.0216533Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_complex128 PASSED [ 87%] 2023-03-31T06:24:48.0216874Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_complex64 PASSED [ 87%] 2023-03-31T06:24:48.0217216Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_float32 PASSED [ 87%] 2023-03-31T06:24:48.0217593Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_float64 PASSED [ 87%] 2023-03-31T06:24:48.0217949Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int16 PASSED [ 87%] 2023-03-31T06:24:48.0218288Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int32 PASSED [ 87%] 2023-03-31T06:24:48.0218623Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int64 PASSED [ 87%] 2023-03-31T06:24:48.0218999Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_int8 PASSED [ 87%] 2023-03-31T06:24:48.0219326Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_cpu_uint8 PASSED [ 87%] 2023-03-31T06:24:48.0219678Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:24:48.0220049Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_bool PASSED [ 88%] 2023-03-31T06:24:48.0220408Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_complex128 PASSED [ 88%] 2023-03-31T06:24:48.0220787Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_complex64 PASSED [ 88%] 2023-03-31T06:24:48.0221162Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float16 PASSED [ 88%] 2023-03-31T06:24:48.0221527Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float32 PASSED [ 88%] 2023-03-31T06:24:48.0221882Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_float64 PASSED [ 88%] 2023-03-31T06:24:48.0222333Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int16 PASSED [ 88%] 2023-03-31T06:24:48.0222702Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int32 PASSED [ 89%] 2023-03-31T06:24:48.0223068Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int64 PASSED [ 89%] 2023-03-31T06:24:48.0223417Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_int8 PASSED [ 89%] 2023-03-31T06:24:48.0223779Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_errors_cpu_uint8 PASSED [ 89%] 2023-03-31T06:24:48.0224159Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:24:48.0224526Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_bool PASSED [ 89%] 2023-03-31T06:24:48.0224911Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_complex128 PASSED [ 89%] 2023-03-31T06:24:48.0225307Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_complex64 PASSED [ 89%] 2023-03-31T06:24:48.0225696Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float16 PASSED [ 90%] 2023-03-31T06:24:48.0226062Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float32 PASSED [ 90%] 2023-03-31T06:24:48.0226434Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_float64 PASSED [ 90%] 2023-03-31T06:24:48.0226812Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int16 PASSED [ 90%] 2023-03-31T06:24:48.0227173Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int32 PASSED [ 90%] 2023-03-31T06:24:48.0227542Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int64 PASSED [ 90%] 2023-03-31T06:24:48.0227915Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_int8 PASSED [ 90%] 2023-03-31T06:24:48.0228286Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_meta_from_tensor_cpu_uint8 PASSED [ 90%] 2023-03-31T06:24:48.0228682Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_bool PASSED [ 90%] 2023-03-31T06:24:48.0229046Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_complex128 PASSED [ 91%] 2023-03-31T06:24:48.0229413Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_complex64 PASSED [ 91%] 2023-03-31T06:24:48.0229778Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_float32 PASSED [ 91%] 2023-03-31T06:24:48.0230123Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_float64 PASSED [ 91%] 2023-03-31T06:24:48.0230515Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int16 PASSED [ 91%] 2023-03-31T06:24:48.0230904Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int32 PASSED [ 91%] 2023-03-31T06:24:48.0231248Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int64 PASSED [ 91%] 2023-03-31T06:24:48.0231636Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_int8 PASSED [ 91%] 2023-03-31T06:24:48.0231993Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_qint32 PASSED [ 92%] 2023-03-31T06:24:48.0232348Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_qint8 PASSED [ 92%] 2023-03-31T06:24:48.0232693Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_quint4x2 PASSED [ 92%] 2023-03-31T06:24:48.0233052Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_quint8 PASSED [ 92%] 2023-03-31T06:24:48.0233405Z test_torch.py::TestTorchDeviceTypeCPU::test_storage_setitem_cpu_uint8 PASSED [ 92%] 2023-03-31T06:24:48.0233754Z test_torch.py::TestTorchDeviceTypeCPU::test_strides_propagation_cpu PASSED [ 92%] 2023-03-31T06:24:48.0234128Z test_torch.py::TestTorchDeviceTypeCPU::test_sync_warning_cpu SKIPPED (Only runs on cuda) [ 92%] 2023-03-31T06:24:48.0234487Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:24:48.0234832Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_bool PASSED [ 92%] 2023-03-31T06:24:48.0235163Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_complex128 PASSED [ 93%] 2023-03-31T06:24:48.0235504Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_complex64 PASSED [ 93%] 2023-03-31T06:24:48.0235844Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float16 PASSED [ 93%] 2023-03-31T06:24:48.0236169Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float32 PASSED [ 93%] 2023-03-31T06:24:48.0236504Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_float64 PASSED [ 93%] 2023-03-31T06:24:48.0236840Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int16 PASSED [ 93%] 2023-03-31T06:24:48.0237172Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int32 PASSED [ 93%] 2023-03-31T06:24:48.0237492Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int64 PASSED [ 93%] 2023-03-31T06:24:48.0237824Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_int8 PASSED [ 94%] 2023-03-31T06:24:48.0238152Z test_torch.py::TestTorchDeviceTypeCPU::test_take_cpu_uint8 PASSED [ 94%] 2023-03-31T06:24:48.0238473Z test_torch.py::TestTorchDeviceTypeCPU::test_take_empty_cpu PASSED [ 94%] 2023-03-31T06:24:48.0238829Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_bfloat16 PASSED [ 94%] 2023-03-31T06:24:48.0239195Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_bool PASSED [ 94%] 2023-03-31T06:24:48.0239569Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_complex128 PASSED [ 94%] 2023-03-31T06:24:48.0239935Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_complex64 PASSED [ 94%] 2023-03-31T06:24:48.0240309Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float16 PASSED [ 94%] 2023-03-31T06:24:48.0240671Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float32 PASSED [ 95%] 2023-03-31T06:24:48.0241024Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_float64 PASSED [ 95%] 2023-03-31T06:24:48.0241423Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int16 PASSED [ 95%] 2023-03-31T06:24:48.0241785Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int32 PASSED [ 95%] 2023-03-31T06:24:48.0242146Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int64 PASSED [ 95%] 2023-03-31T06:24:48.0242491Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_int8 PASSED [ 95%] 2023-03-31T06:24:48.0242850Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_from_storage_cpu_uint8 PASSED [ 95%] 2023-03-31T06:24:48.0243543Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_set_errors_multigpu_cpu SKIPPED (fewer than 2 devices detected) [ 95%] 2023-03-31T06:24:48.0243994Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_shape_empty_cpu PASSED [ 95%] 2023-03-31T06:24:48.0244350Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:24:48.0244765Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_bool PASSED [ 96%] 2023-03-31T06:24:48.0245143Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_complex128 PASSED [ 96%] 2023-03-31T06:24:48.0245513Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_complex64 PASSED [ 96%] 2023-03-31T06:24:48.0245888Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float16 PASSED [ 96%] 2023-03-31T06:24:48.0246260Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float32 PASSED [ 96%] 2023-03-31T06:24:48.0246626Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_float64 PASSED [ 96%] 2023-03-31T06:24:48.0246980Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int16 PASSED [ 96%] 2023-03-31T06:24:48.0247340Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int32 PASSED [ 97%] 2023-03-31T06:24:48.0247704Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int64 PASSED [ 97%] 2023-03-31T06:24:48.0248055Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_int8 PASSED [ 97%] 2023-03-31T06:24:48.0248419Z test_torch.py::TestTorchDeviceTypeCPU::test_tensor_storage_type_cpu_uint8 PASSED [ 97%] 2023-03-31T06:24:48.0248790Z test_torch.py::TestTorchDeviceTypeCPU::test_ternary_op_mem_overlap_cpu_float64 PASSED [ 97%] 2023-03-31T06:24:48.0249165Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:24:48.0249514Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_bool PASSED [ 97%] 2023-03-31T06:24:48.0249884Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_complex128 PASSED [ 97%] 2023-03-31T06:24:48.0250260Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_complex64 PASSED [ 97%] 2023-03-31T06:24:48.0250616Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float16 PASSED [ 98%] 2023-03-31T06:24:48.0250985Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float32 PASSED [ 98%] 2023-03-31T06:24:48.0251350Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_float64 PASSED [ 98%] 2023-03-31T06:24:48.0251714Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int16 PASSED [ 98%] 2023-03-31T06:24:48.0252061Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int32 PASSED [ 98%] 2023-03-31T06:24:48.0252421Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int64 PASSED [ 98%] 2023-03-31T06:24:48.0252780Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_int8 PASSED [ 98%] 2023-03-31T06:24:48.0253128Z test_torch.py::TestTorchDeviceTypeCPU::test_typed_storage_meta_cpu_uint8 PASSED [ 98%] 2023-03-31T06:24:48.0253498Z test_torch.py::TestTorchDeviceTypeCPU::test_unfold_all_devices_and_dtypes_cpu PASSED [ 99%] 2023-03-31T06:24:48.0253858Z test_torch.py::TestTorchDeviceTypeCPU::test_unfold_scalars_cpu PASSED [ 99%] 2023-03-31T06:24:48.0254218Z test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:24:48.0254609Z test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float16 PASSED [ 99%] 2023-03-31T06:24:48.0254967Z test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float32 PASSED [ 99%] 2023-03-31T06:24:48.0255323Z test_torch.py::TestTorchDeviceTypeCPU::test_uniform_kstest_cpu_float64 PASSED [ 99%] 2023-03-31T06:24:48.0255681Z test_torch.py::TestTorchDeviceTypeCPU::test_untyped_storage_meta_cpu PASSED [ 99%] 2023-03-31T06:24:48.0256029Z test_torch.py::TestTorchDeviceTypeCPU::test_warn_always_caught_cpu PASSED [ 99%] 2023-03-31T06:24:48.0256437Z test_torch.py::TestTorchDeviceTypeCPU::test_where_scalar_handcrafted_values_cpu PASSED [100%] 2023-03-31T06:24:48.0256679Z 2023-03-31T06:24:48.0257122Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_torch/test_torch-e4a1d68a94426a17.xml - 2023-03-31T06:24:48.0257510Z ======================= 802 passed, 39 skipped in 12.03s ======================= 2023-03-31T06:24:48.0257837Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:24:48.0258143Z [TORCH_VITAL] Dataloader.enabled True 2023-03-31T06:24:48.0258419Z [TORCH_VITAL] Dataloader.basic_unit_test TEST_VALUE_STRING 2023-03-31T06:24:48.0258659Z [TORCH_VITAL] CUDA.used False 2023-03-31T06:24:48.0258791Z 2023-03-31T06:24:48.0259167Z ##[endgroup] 2023-03-31T06:24:48.0259569Z FINISHED PRINTING LOG FILE of test_torch (/var/lib/jenkins/workspace/test/test-reports/test_torch_b6ffbma0.log) 2023-03-31T06:24:48.0259776Z 2023-03-31T06:24:48.0259947Z Running test_decomp ... [2023-03-31 06:24:47.975839] 2023-03-31T06:24:49.6408483Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:24:49.6435925Z Ignoring disabled issues: [] 2023-03-31T06:24:49.6559448Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:24:49.6587503Z Ignoring disabled issues: [] 2023-03-31T06:24:49.6620876Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_decomp.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=0', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:24:49.661743] 2023-03-31T06:24:49.6776886Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_decomp.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=1', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:24:49.677294] 2023-03-31T06:51:18.7517047Z 2023-03-31T06:51:18.7517578Z Expand the folded group to see the log file of test_decomp 2023-03-31T06:51:18.7518637Z ##[group]PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_geq5cc7g.log) 2023-03-31T06:51:18.7526154Z Test results will be stored in test-reports/python-pytest/test_decomp/test_decomp-d4f07a676c877163.xml 2023-03-31T06:51:18.7526766Z ============================= test session starts ============================== 2023-03-31T06:51:18.7528246Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:51:18.7529052Z cachedir: .pytest_cache 2023-03-31T06:51:18.7529986Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:51:18.7530667Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:51:18.7531447Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:51:18.7532064Z collecting ... collected 7785 items / 4 deselected / 7781 selected 2023-03-31T06:51:18.7925528Z Running 3928 items in this shard: test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdist_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdist_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_grid_sampler_2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_histogram_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_istft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanquantile_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_ctc_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_grid_sample_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_area_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_grad_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_grad_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_grad_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_grad_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_one_hot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pdist_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_complex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_pca_lowrank_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polar_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_quantile_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_blackman_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_cosine_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_exponential_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_exponential_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_gaussian_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_cosine_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_cosine_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_hamming_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hann_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_nuttall_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_lowrank_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_lowrank_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_indices_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_grid_sampler_2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_tril_indices_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_triu_indices_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_triu_indices_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_eval_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_eval_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_train_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_train_mode_cpu_float32, test/test_decomp.py::DecompContiguousTestsCPU::test_contiguous_softmax_cpu, test/test_decomp.py::HasDecompTest::test_has_decomposition 2023-03-31T06:51:18.8268894Z 2023-03-31T06:51:18.8271879Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8272315Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float64 PASSED [ 0%] 2023-03-31T06:51:18.8272650Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int16 PASSED [ 0%] 2023-03-31T06:51:18.8273044Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int8 PASSED [ 0%] 2023-03-31T06:51:18.8273372Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:18.8273704Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_bool PASSED [ 0%] 2023-03-31T06:51:18.8274045Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex32 PASSED [ 0%] 2023-03-31T06:51:18.8274374Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex64 PASSED [ 0%] 2023-03-31T06:51:18.8274708Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8275039Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int64 PASSED [ 0%] 2023-03-31T06:51:18.8275368Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int8 PASSED [ 0%] 2023-03-31T06:51:18.8275681Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:18.8276029Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex128 PASSED [ 0%] 2023-03-31T06:51:18.8276395Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float16 PASSED [ 0%] 2023-03-31T06:51:18.8276744Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8277096Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_float64 PASSED [ 0%] 2023-03-31T06:51:18.8277450Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int32 PASSED [ 0%] 2023-03-31T06:51:18.8277798Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:18.8278133Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_bool PASSED [ 0%] 2023-03-31T06:51:18.8278498Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_complex64 PASSED [ 0%] 2023-03-31T06:51:18.8278834Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int32 PASSED [ 0%] 2023-03-31T06:51:18.8279177Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int64 PASSED [ 0%] 2023-03-31T06:51:18.8279519Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:18.8279850Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_bool PASSED [ 0%] 2023-03-31T06:51:18.8280190Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:18.8280530Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int64 PASSED [ 0%] 2023-03-31T06:51:18.8280867Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:18.8281201Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:18.8281551Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_bool PASSED [ 0%] 2023-03-31T06:51:18.8281898Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float16 PASSED [ 0%] 2023-03-31T06:51:18.8282234Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8282582Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int64 PASSED [ 0%] 2023-03-31T06:51:18.8282981Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:18.8283508Z test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float16 PASSED [ 0%] 2023-03-31T06:51:18.8283843Z test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8284190Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float32 PASSED [ 0%] 2023-03-31T06:51:18.8284575Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float64 PASSED [ 0%] 2023-03-31T06:51:18.8284945Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:18.8285288Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int32 PASSED [ 0%] 2023-03-31T06:51:18.8285630Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_bool PASSED [ 1%] 2023-03-31T06:51:18.8286006Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int16 PASSED [ 1%] 2023-03-31T06:51:18.8286342Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8286696Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_complex128 PASSED [ 1%] 2023-03-31T06:51:18.8287051Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_complex64 PASSED [ 1%] 2023-03-31T06:51:18.8287391Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float16 PASSED [ 1%] 2023-03-31T06:51:18.8287739Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float32 PASSED [ 1%] 2023-03-31T06:51:18.8288085Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int16 PASSED [ 1%] 2023-03-31T06:51:18.8288431Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:18.8288760Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:18.8289105Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:18.8289452Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8289791Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_complex64 PASSED [ 1%] 2023-03-31T06:51:18.8290139Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float16 PASSED [ 1%] 2023-03-31T06:51:18.8290481Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int16 PASSED [ 1%] 2023-03-31T06:51:18.8290825Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:18.8291155Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:18.8291493Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_bool PASSED [ 1%] 2023-03-31T06:51:18.8291829Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int16 PASSED [ 1%] 2023-03-31T06:51:18.8292159Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int32 PASSED [ 1%] 2023-03-31T06:51:18.8292503Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:18.8292839Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:18.8293202Z test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8293577Z test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_float32 PASSED [ 1%] 2023-03-31T06:51:18.8293972Z test_decomp.py::TestDecompCPU::test_comprehensive__native_batch_norm_legit_cpu_float64 PASSED [ 1%] 2023-03-31T06:51:18.8294366Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8294760Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float16 PASSED [ 1%] 2023-03-31T06:51:18.8295137Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float32 PASSED [ 1%] 2023-03-31T06:51:18.8295577Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8295969Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float32 PASSED [ 1%] 2023-03-31T06:51:18.8296342Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float64 PASSED [ 1%] 2023-03-31T06:51:18.8296709Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:18.8297090Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex128 PASSED [ 1%] 2023-03-31T06:51:18.8297467Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex32 PASSED [ 1%] 2023-03-31T06:51:18.8297798Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_complex64 PASSED [ 1%] 2023-03-31T06:51:18.8298138Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float32 PASSED [ 1%] 2023-03-31T06:51:18.8298509Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int64 PASSED [ 1%] 2023-03-31T06:51:18.8298832Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int8 PASSED [ 1%] 2023-03-31T06:51:18.8299163Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:18.8299502Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:18.8299838Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_bool PASSED [ 2%] 2023-03-31T06:51:18.8300166Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:18.8300514Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:18.8300856Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:18.8301180Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:18.8301519Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:18.8301859Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:18.8302205Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:18.8302532Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:18.8302865Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:18.8303201Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:18.8303632Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:18.8303966Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:18.8304293Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_bool PASSED [ 2%] 2023-03-31T06:51:18.8304638Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:18.8304974Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:18.8305317Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:18.8305655Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:18.8305971Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:18.8306304Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:18.8306646Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:18.8306998Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:18.8307330Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:18.8307671Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:18.8308028Z test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:18.8308426Z test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:18.8308787Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:18.8309140Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:18.8309484Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:18.8309854Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:18.8310230Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:18.8310571Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:18.8310897Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:18.8311280Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:18.8311640Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:18.8312016Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:18.8312364Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:18.8312718Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:18.8313068Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8313393Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:18.8313741Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:18.8314085Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_bool PASSED [ 3%] 2023-03-31T06:51:18.8314432Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:18.8314771Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:18.8315112Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:18.8315447Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:18.8315766Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:18.8316108Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:18.8316450Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:18.8316788Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_bool PASSED [ 3%] 2023-03-31T06:51:18.8317116Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:18.8317462Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:18.8317799Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:18.8318117Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8318444Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:18.8318772Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:18.8319101Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:18.8319437Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:18.8319800Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_complex128 PASSED [ 3%] 2023-03-31T06:51:18.8320164Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:18.8320504Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_bool PASSED [ 3%] 2023-03-31T06:51:18.8320845Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float16 PASSED [ 3%] 2023-03-31T06:51:18.8321220Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:18.8321562Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8321882Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:18.8322211Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:18.8322585Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:18.8322945Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:18.8323589Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:18.8323931Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8324343Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:18.8324674Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_bool PASSED [ 3%] 2023-03-31T06:51:18.8325020Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8325369Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:18.8325702Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:18.8326058Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_complex128 PASSED [ 3%] 2023-03-31T06:51:18.8326415Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:18.8326766Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:18.8327088Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:18.8327423Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:18.8327756Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_bool PASSED [ 4%] 2023-03-31T06:51:18.8328087Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_complex128 PASSED [ 4%] 2023-03-31T06:51:18.8328435Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_complex64 PASSED [ 4%] 2023-03-31T06:51:18.8328778Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8329118Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:18.8329440Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:18.8329767Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:18.8330105Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:18.8330441Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:18.8330790Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:18.8331130Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:18.8331468Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:18.8331800Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8332142Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:18.8332494Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:18.8332824Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:18.8333163Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:18.8333504Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8333850Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:18.8334229Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:18.8334571Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_bool PASSED [ 4%] 2023-03-31T06:51:18.8334916Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8335257Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:18.8335644Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:18.8336028Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:18.8336370Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:18.8336707Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8337089Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:18.8337439Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:18.8337774Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:18.8338124Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:18.8338481Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_bool PASSED [ 4%] 2023-03-31T06:51:18.8338846Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex128 PASSED [ 4%] 2023-03-31T06:51:18.8339201Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex32 PASSED [ 4%] 2023-03-31T06:51:18.8339565Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_complex64 PASSED [ 4%] 2023-03-31T06:51:18.8339923Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:18.8340281Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:18.8340623Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:18.8340973Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:18.8341325Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:18.8341691Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex32 PASSED [ 5%] 2023-03-31T06:51:18.8342088Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float16 PASSED [ 5%] 2023-03-31T06:51:18.8342480Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:18.8342869Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:18.8343244Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:18.8343719Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:18.8344108Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:18.8344533Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_bool SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:18.8345028Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex64 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:18.8345514Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int16 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:18.8345925Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_complex128 PASSED [ 5%] 2023-03-31T06:51:18.8346272Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_float32 PASSED [ 5%] 2023-03-31T06:51:18.8346653Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:18.8346995Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:18.8347332Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:18.8347669Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_complex128 PASSED [ 5%] 2023-03-31T06:51:18.8348024Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_complex64 PASSED [ 5%] 2023-03-31T06:51:18.8348401Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_float32 PASSED [ 5%] 2023-03-31T06:51:18.8348764Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:18.8349110Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:18.8368468Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:18.8368849Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:18.8369211Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:18.8369562Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_bool PASSED [ 5%] 2023-03-31T06:51:18.8369910Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:18.8370251Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:18.8370581Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:18.8370932Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:18.8371268Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_complex64 PASSED [ 5%] 2023-03-31T06:51:18.8371617Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_float32 PASSED [ 5%] 2023-03-31T06:51:18.8371969Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:18.8372301Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:18.8372637Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:18.8372975Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_bool PASSED [ 5%] 2023-03-31T06:51:18.8373321Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:18.8373648Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:18.8373984Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:18.8374329Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:18.8374664Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_bool PASSED [ 6%] 2023-03-31T06:51:18.8375026Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:18.8375386Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:18.8375740Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_bool PASSED [ 6%] 2023-03-31T06:51:18.8376089Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex128 PASSED [ 6%] 2023-03-31T06:51:18.8376452Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex32 PASSED [ 6%] 2023-03-31T06:51:18.8376815Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:18.8377165Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:18.8377525Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:18.8377876Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:18.8378233Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:18.8378808Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:18.8379163Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:18.8379514Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:18.8379850Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_bool PASSED [ 6%] 2023-03-31T06:51:18.8380206Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex32 PASSED [ 6%] 2023-03-31T06:51:18.8380684Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:18.8381080Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:18.8381413Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:18.8381814Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:18.8382174Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:18.8382513Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:18.8382864Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:18.8383210Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:18.8383684Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:18.8384033Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:18.8384380Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:18.8384735Z test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:18.8385081Z test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:18.8385444Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex128 PASSED [ 6%] 2023-03-31T06:51:18.8385808Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex32 PASSED [ 6%] 2023-03-31T06:51:18.8386163Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:18.8386498Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:18.8386846Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:18.8387193Z test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:18.8387533Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_bool PASSED [ 6%] 2023-03-31T06:51:18.8387882Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:18.8388234Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:18.8388585Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:18.8388923Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:18.8389286Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:18.8389659Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:18.8390022Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_bool PASSED [ 7%] 2023-03-31T06:51:18.8390363Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:18.8390713Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:18.8391061Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_bool PASSED [ 7%] 2023-03-31T06:51:18.8391416Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:18.8391824Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:18.8392196Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:18.8392556Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_bool PASSED [ 7%] 2023-03-31T06:51:18.8392892Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:18.8393268Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:18.8393652Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_bfloat16 PASSED [ 7%] 2023-03-31T06:51:18.8393993Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_bool PASSED [ 7%] 2023-03-31T06:51:18.8394385Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex128 PASSED [ 7%] 2023-03-31T06:51:18.8394753Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex32 PASSED [ 7%] 2023-03-31T06:51:18.8395110Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float16 PASSED [ 7%] 2023-03-31T06:51:18.8395452Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float32 PASSED [ 7%] 2023-03-31T06:51:18.8395804Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:18.8396151Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:18.8396489Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:18.8396835Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:18.8397181Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_bfloat16 PASSED [ 7%] 2023-03-31T06:51:18.8397534Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_complex128 PASSED [ 7%] 2023-03-31T06:51:18.8397869Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_complex64 PASSED [ 7%] 2023-03-31T06:51:18.8398211Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:18.8398546Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:18.8398871Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_bfloat16 PASSED [ 7%] 2023-03-31T06:51:18.8399210Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_bool PASSED [ 7%] 2023-03-31T06:51:18.8399559Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex128 PASSED [ 7%] 2023-03-31T06:51:18.8399917Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex32 PASSED [ 7%] 2023-03-31T06:51:18.8400253Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_complex64 PASSED [ 7%] 2023-03-31T06:51:18.8400594Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float16 PASSED [ 7%] 2023-03-31T06:51:18.8400939Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:18.8401265Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:18.8401601Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:18.8401932Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:18.8402269Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:18.8402624Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8403009Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_complex64 PASSED [ 8%] 2023-03-31T06:51:18.8403699Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:18.8404060Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:18.8404442Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:18.8404895Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:18.8405263Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:18.8405609Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:18.8405975Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:18.8406377Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8406766Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_bool PASSED [ 8%] 2023-03-31T06:51:18.8407116Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_complex64 PASSED [ 8%] 2023-03-31T06:51:18.8407517Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:18.8407882Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:18.8408227Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:18.8408579Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:18.8408937Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:18.8409297Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8409643Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:18.8409999Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:18.8410348Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:18.8410687Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:18.8411034Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:18.8411385Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8411729Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_bool PASSED [ 8%] 2023-03-31T06:51:18.8412055Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:18.8412395Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float64 PASSED [ 8%] 2023-03-31T06:51:18.8412737Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:18.8413061Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:18.8413415Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8413782Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_bool PASSED [ 8%] 2023-03-31T06:51:18.8414156Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_complex128 PASSED [ 8%] 2023-03-31T06:51:18.8414515Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:18.8414881Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float64 PASSED [ 8%] 2023-03-31T06:51:18.8415242Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:18.8415592Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:18.8415958Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:18.8416310Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:18.8416658Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8416987Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int16 PASSED [ 9%] 2023-03-31T06:51:18.8417361Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int8 PASSED [ 9%] 2023-03-31T06:51:18.8417705Z test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8418035Z test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:18.8418383Z test_decomp.py::TestDecompCPU::test_comprehensive_cdist_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8418727Z test_decomp.py::TestDecompCPU::test_comprehensive_cdist_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:18.8419102Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_bool PASSED [ 9%] 2023-03-31T06:51:18.8419469Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8419830Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:18.8420226Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8420565Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_uint8 PASSED [ 9%] 2023-03-31T06:51:18.8420910Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8421252Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:18.8421590Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int64 PASSED [ 9%] 2023-03-31T06:51:18.8421914Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_uint8 PASSED [ 9%] 2023-03-31T06:51:18.8422261Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:51:18.8422610Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_bool PASSED [ 9%] 2023-03-31T06:51:18.8422947Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8423308Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:18.8423769Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float16 PASSED [ 9%] 2023-03-31T06:51:18.8424114Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int16 PASSED [ 9%] 2023-03-31T06:51:18.8424441Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int32 PASSED [ 9%] 2023-03-31T06:51:18.8424779Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int64 PASSED [ 9%] 2023-03-31T06:51:18.8425126Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:51:18.8425467Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8425818Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:18.8426162Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int16 PASSED [ 9%] 2023-03-31T06:51:18.8426510Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int8 PASSED [ 9%] 2023-03-31T06:51:18.8426837Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_uint8 PASSED [ 9%] 2023-03-31T06:51:18.8427179Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:51:18.8427528Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:18.8427857Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8428195Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int8 PASSED [ 9%] 2023-03-31T06:51:18.8428548Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8428924Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8429287Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:18.8429663Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:18.8430079Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:18.8430452Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_complex64 PASSED [ 10%] 2023-03-31T06:51:18.8430807Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:18.8431163Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_bool PASSED [ 10%] 2023-03-31T06:51:18.8431513Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex32 PASSED [ 10%] 2023-03-31T06:51:18.8431876Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:18.8432245Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:18.8432589Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:18.8432957Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:18.8433282Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:18.8433614Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:18.8433952Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:18.8434273Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:18.8434605Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:18.8434958Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:18.8435316Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:18.8435657Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:18.8436006Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:18.8436350Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:18.8436693Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:18.8437043Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:18.8437391Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:18.8437732Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:18.8438067Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:18.8438413Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex32 PASSED [ 10%] 2023-03-31T06:51:18.8438761Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:18.8439089Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:18.8439431Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:18.8439774Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:18.8440112Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:18.8440446Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_bool PASSED [ 10%] 2023-03-31T06:51:18.8440808Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex128 PASSED [ 10%] 2023-03-31T06:51:18.8441177Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex64 PASSED [ 10%] 2023-03-31T06:51:18.8441528Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:18.8441885Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:18.8442240Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:18.8442603Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:18.8442986Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:18.8443638Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:18.8443991Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:18.8444324Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_bool PASSED [ 10%] 2023-03-31T06:51:18.8444737Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:18.8445128Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:18.8445474Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float16 PASSED [ 11%] 2023-03-31T06:51:18.8445802Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:18.8446178Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:18.8446523Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int32 PASSED [ 11%] 2023-03-31T06:51:18.8446845Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:18.8447199Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:18.8447567Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex32 PASSED [ 11%] 2023-03-31T06:51:18.8447937Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:18.8448291Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:18.8448646Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:18.8449008Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:18.8449375Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:18.8449739Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:18.8450107Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:18.8450469Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int32 PASSED [ 11%] 2023-03-31T06:51:18.8450816Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:18.8451175Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:18.8451538Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_bool PASSED [ 11%] 2023-03-31T06:51:18.8451896Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:18.8452244Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float16 PASSED [ 11%] 2023-03-31T06:51:18.8452600Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:18.8452959Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int32 PASSED [ 11%] 2023-03-31T06:51:18.8453297Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:18.8453651Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:18.8454003Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:18.8454353Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float16 PASSED [ 11%] 2023-03-31T06:51:18.8454691Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:18.8455040Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:18.8455387Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:18.8455773Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:18.8456138Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:18.8456498Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:18.8456852Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:18.8457190Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:18.8457559Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:18.8457938Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:18.8458269Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:18.8458641Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:18.8458977Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:18.8459315Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8459634Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:18.8459971Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:18.8460314Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_bool PASSED [ 12%] 2023-03-31T06:51:18.8460645Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:18.8460994Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:18.8461332Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:18.8461668Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8461986Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:18.8462319Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:18.8462670Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:18.8463022Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_bool PASSED [ 12%] 2023-03-31T06:51:18.8463479Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_complex64 PASSED [ 12%] 2023-03-31T06:51:18.8463856Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:18.8464218Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:18.8464563Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8464926Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:18.8465277Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:18.8465619Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:18.8465969Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:18.8466308Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8466643Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:18.8466966Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:18.8467297Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:18.8467640Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_complex64 PASSED [ 12%] 2023-03-31T06:51:18.8467974Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:18.8468320Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:18.8468701Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:18.8469037Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8469363Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:18.8469696Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:18.8470097Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:18.8470469Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:18.8470817Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:18.8471165Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:18.8471545Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:18.8471879Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:18.8472223Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_uint8 PASSED [ 13%] 2023-03-31T06:51:18.8472574Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:18.8472914Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:18.8473264Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_uint8 PASSED [ 13%] 2023-03-31T06:51:18.8473617Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:18.8473965Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:18.8474297Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:18.8474642Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:18.8475007Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:18.8475385Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_complex128 PASSED [ 13%] 2023-03-31T06:51:18.8475780Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_complex64 PASSED [ 13%] 2023-03-31T06:51:18.8476168Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:18.8476551Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:18.8476917Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:18.8477290Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_int8 PASSED [ 13%] 2023-03-31T06:51:18.8477657Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:18.8478012Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_bool PASSED [ 13%] 2023-03-31T06:51:18.8478350Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:18.8478698Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float64 PASSED [ 13%] 2023-03-31T06:51:18.8479045Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:18.8479375Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:18.8479720Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int8 PASSED [ 13%] 2023-03-31T06:51:18.8480125Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_complex64 SKIPPED (diag in torch.complex64 not supported) [ 13%] 2023-03-31T06:51:18.8480574Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float64 SKIPPED (diag in torch.float64 not supported) [ 13%] 2023-03-31T06:51:18.8480988Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int16 SKIPPED (diag in torch.int16 not supported) [ 13%] 2023-03-31T06:51:18.8481484Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int32 SKIPPED (diag in torch.int32 not supported) [ 13%] 2023-03-31T06:51:18.8481903Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int64 SKIPPED (diag in torch.int64 not supported) [ 13%] 2023-03-31T06:51:18.8482313Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_uint8 SKIPPED (diag in torch.uint8 not supported) [ 13%] 2023-03-31T06:51:18.8482712Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_bool PASSED [ 13%] 2023-03-31T06:51:18.8483286Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex128 PASSED [ 13%] 2023-03-31T06:51:18.8483658Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex32 PASSED [ 13%] 2023-03-31T06:51:18.8484008Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_complex64 PASSED [ 13%] 2023-03-31T06:51:18.8484428Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float64 PASSED [ 13%] 2023-03-31T06:51:18.8484786Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:18.8485145Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_complex128 PASSED [ 13%] 2023-03-31T06:51:18.8485491Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float16 PASSED [ 13%] 2023-03-31T06:51:18.8485843Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:18.8486201Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_bool PASSED [ 13%] 2023-03-31T06:51:18.8486560Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex128 PASSED [ 13%] 2023-03-31T06:51:18.8486934Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex64 PASSED [ 14%] 2023-03-31T06:51:18.8487303Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:18.8487669Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:18.8488016Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:18.8488375Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex128 PASSED [ 14%] 2023-03-31T06:51:18.8488728Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:18.8489082Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:18.8489421Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:18.8489778Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_bool PASSED [ 14%] 2023-03-31T06:51:18.8490149Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:18.8490508Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:18.8490879Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:18.8491241Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:18.8491601Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:18.8491945Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:18.8492292Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_complex128 PASSED [ 14%] 2023-03-31T06:51:18.8492645Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:18.8492978Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int32 PASSED [ 14%] 2023-03-31T06:51:18.8493321Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:18.8493670Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:18.8494057Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_bool PASSED [ 14%] 2023-03-31T06:51:18.8494395Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:18.8494746Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:18.8495092Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:18.8495421Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:18.8495796Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:18.8496178Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_complex128 PASSED [ 14%] 2023-03-31T06:51:18.8496524Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:18.8496899Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:18.8497279Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:18.8497657Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:18.8498016Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int32 PASSED [ 14%] 2023-03-31T06:51:18.8498384Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:18.8498763Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:18.8499145Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_bool PASSED [ 14%] 2023-03-31T06:51:18.8499510Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_complex64 PASSED [ 14%] 2023-03-31T06:51:18.8499887Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int32 PASSED [ 14%] 2023-03-31T06:51:18.8500267Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:18.8500642Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:18.8501000Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:18.8501356Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:18.8501703Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:18.8502042Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:18.8502386Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:18.8502724Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:18.8503060Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int64 PASSED [ 15%] 2023-03-31T06:51:18.8503467Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:18.8503822Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:18.8504177Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:18.8504520Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex32 PASSED [ 15%] 2023-03-31T06:51:18.8504877Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_complex64 PASSED [ 15%] 2023-03-31T06:51:18.8505236Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float16 PASSED [ 15%] 2023-03-31T06:51:18.8505585Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:18.8505923Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:18.8506268Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:18.8506615Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:18.8506991Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex32 PASSED [ 15%] 2023-03-31T06:51:18.8507344Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float16 PASSED [ 15%] 2023-03-31T06:51:18.8507690Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:18.8508040Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:18.8508400Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int16 PASSED [ 15%] 2023-03-31T06:51:18.8508774Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int64 PASSED [ 15%] 2023-03-31T06:51:18.8509121Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:18.8509457Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_bool PASSED [ 15%] 2023-03-31T06:51:18.8509847Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:18.8510208Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex64 PASSED [ 15%] 2023-03-31T06:51:18.8510560Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:18.8510891Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int16 PASSED [ 15%] 2023-03-31T06:51:18.8511232Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int64 PASSED [ 15%] 2023-03-31T06:51:18.8511574Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:18.8511911Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:18.8512267Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_complex64 PASSED [ 15%] 2023-03-31T06:51:18.8512616Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:18.8512962Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:18.8513299Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int16 PASSED [ 15%] 2023-03-31T06:51:18.8513639Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:18.8513983Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_uint8 PASSED [ 15%] 2023-03-31T06:51:18.8514377Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_bfloat16 SKIPPED (empty in torch.bfloat16 not supported) [ 15%] 2023-03-31T06:51:18.8514819Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_bool SKIPPED (empty in torch.bool not supported) [ 15%] 2023-03-31T06:51:18.8515262Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex128 SKIPPED (empty in torch.complex128 not supported) [ 16%] 2023-03-31T06:51:18.8515717Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex32 SKIPPED (empty in torch.complex32 not supported) [ 16%] 2023-03-31T06:51:18.8516151Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float32 SKIPPED (empty in torch.float32 not supported) [ 16%] 2023-03-31T06:51:18.8516577Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int16 SKIPPED (empty in torch.int16 not supported) [ 16%] 2023-03-31T06:51:18.8516998Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int32 SKIPPED (empty in torch.int32 not supported) [ 16%] 2023-03-31T06:51:18.8517411Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_uint8 SKIPPED (empty in torch.uint8 not supported) [ 16%] 2023-03-31T06:51:18.8517839Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_bfloat16 SKIPPED (empty_like in torch.bfloat16 not supported) [ 16%] 2023-03-31T06:51:18.8518294Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_bool SKIPPED (empty_like in torch.bool not supported) [ 16%] 2023-03-31T06:51:18.8518748Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex128 SKIPPED (empty_like in torch.complex128 not supported) [ 16%] 2023-03-31T06:51:18.8519247Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex32 SKIPPED (empty_like in torch.complex32 not supported) [ 16%] 2023-03-31T06:51:18.8519694Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int16 SKIPPED (empty_like in torch.int16 not supported) [ 16%] 2023-03-31T06:51:18.8520131Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int32 SKIPPED (empty_like in torch.int32 not supported) [ 16%] 2023-03-31T06:51:18.8520596Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int8 SKIPPED (empty_like in torch.int8 not supported) [ 16%] 2023-03-31T06:51:18.8521069Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_uint8 SKIPPED (empty_like in torch.uint8 not supported) [ 16%] 2023-03-31T06:51:18.8521461Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:18.8521864Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex32 PASSED [ 16%] 2023-03-31T06:51:18.8522244Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex64 PASSED [ 16%] 2023-03-31T06:51:18.8522609Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:18.8522960Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:18.8523612Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:18.8523982Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int8 PASSED [ 16%] 2023-03-31T06:51:18.8524326Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_bool PASSED [ 16%] 2023-03-31T06:51:18.8524672Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex64 PASSED [ 16%] 2023-03-31T06:51:18.8525017Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float16 PASSED [ 16%] 2023-03-31T06:51:18.8525362Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float64 PASSED [ 16%] 2023-03-31T06:51:18.8525686Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:18.8526017Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:18.8526353Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_bool PASSED [ 16%] 2023-03-31T06:51:18.8526693Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_complex128 PASSED [ 16%] 2023-03-31T06:51:18.8527043Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float16 PASSED [ 16%] 2023-03-31T06:51:18.8527391Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:18.8527734Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int8 PASSED [ 16%] 2023-03-31T06:51:18.8528061Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_float64 PASSED [ 16%] 2023-03-31T06:51:18.8528400Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:18.8528736Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:18.8529062Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:18.8529399Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:18.8529733Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:18.8530064Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int8 PASSED [ 16%] 2023-03-31T06:51:18.8530396Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_bfloat16 PASSED [ 17%] 2023-03-31T06:51:18.8530742Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_bool PASSED [ 17%] 2023-03-31T06:51:18.8531088Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:18.8531421Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:18.8531768Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:18.8532180Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:18.8532525Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:18.8532853Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:18.8533195Z test_decomp.py::TestDecompCPU::test_comprehensive_erfinv_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:18.8533580Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_bfloat16 PASSED [ 17%] 2023-03-31T06:51:18.8533940Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_bool PASSED [ 17%] 2023-03-31T06:51:18.8534282Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_complex128 PASSED [ 17%] 2023-03-31T06:51:18.8534637Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:18.8535019Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float16 PASSED [ 17%] 2023-03-31T06:51:18.8535347Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:18.8535683Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:18.8536028Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:18.8536357Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:18.8536691Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:18.8537032Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:18.8537384Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_bfloat16 PASSED [ 17%] 2023-03-31T06:51:18.8537733Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_complex128 PASSED [ 17%] 2023-03-31T06:51:18.8538100Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:18.8538459Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:18.8538801Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:18.8539148Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:18.8539498Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:18.8539842Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_bool PASSED [ 17%] 2023-03-31T06:51:18.8540181Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_complex128 PASSED [ 17%] 2023-03-31T06:51:18.8540541Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:18.8540894Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:18.8541227Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:18.8541573Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:18.8541911Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:18.8542252Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:18.8542578Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:18.8542918Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:18.8543262Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:18.8543688Z test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_bfloat16 PASSED [ 17%] 2023-03-31T06:51:18.8544055Z test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float16 PASSED [ 17%] 2023-03-31T06:51:18.8544419Z test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:18.8544805Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8545132Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float16 PASSED [ 18%] 2023-03-31T06:51:18.8545465Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:18.8545803Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:18.8546123Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_uint8 PASSED [ 18%] 2023-03-31T06:51:18.8546491Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:18.8546871Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:18.8547216Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8547585Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:18.8547942Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:18.8548293Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:18.8548622Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:18.8548966Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8549317Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:18.8549677Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:18.8550020Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:18.8550362Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:18.8550715Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_bfloat16 PASSED [ 18%] 2023-03-31T06:51:18.8551064Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8551430Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:18.8551797Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:18.8552158Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:18.8552504Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:18.8552863Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:18.8553222Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:18.8553564Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:18.8553921Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:18.8554288Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:18.8554644Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:18.8554978Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:18.8555324Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:18.8555670Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8556022Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:18.8556365Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:18.8556715Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:18.8557067Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:18.8557430Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_bool PASSED [ 18%] 2023-03-31T06:51:18.8557782Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:18.8558135Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:18.8558484Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:18.8558825Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_complex64 PASSED [ 19%] 2023-03-31T06:51:18.8559218Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:18.8559591Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:18.8559926Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_complex64 PASSED [ 19%] 2023-03-31T06:51:18.8560303Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:18.8560646Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8560993Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_bool PASSED [ 19%] 2023-03-31T06:51:18.8561333Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_complex128 PASSED [ 19%] 2023-03-31T06:51:18.8561695Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:18.8562046Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:18.8562379Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:18.8562735Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8563307Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_bfloat16 PASSED [ 19%] 2023-03-31T06:51:18.8563675Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_bool PASSED [ 19%] 2023-03-31T06:51:18.8564032Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex32 PASSED [ 19%] 2023-03-31T06:51:18.8564403Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:18.8564769Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8565111Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_bool PASSED [ 19%] 2023-03-31T06:51:18.8565462Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:18.8565820Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:18.8566175Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:18.8566513Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8566865Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_bool PASSED [ 19%] 2023-03-31T06:51:18.8567212Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:18.8567545Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:18.8567899Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:18.8568250Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:18.8568596Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8568939Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_complex64 PASSED [ 19%] 2023-03-31T06:51:18.8569298Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:18.8569650Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:18.8569989Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:18.8570403Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:18.8570757Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:18.8571107Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:18.8571444Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:18.8571794Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:18.8572180Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:18.8572553Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8572910Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:18.8573343Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:18.8573700Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:18.8574038Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:18.8574391Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:18.8574738Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8575086Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:18.8575415Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:18.8575759Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:18.8576107Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:18.8576444Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8576794Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:18.8577137Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8577495Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:18.8577833Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:18.8578178Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:18.8578528Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:18.8578863Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_bfloat16 PASSED [ 20%] 2023-03-31T06:51:18.8579209Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:18.8579558Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex32 PASSED [ 20%] 2023-03-31T06:51:18.8579907Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:18.8580235Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:18.8580574Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:18.8580917Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8581258Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:18.8581623Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex32 PASSED [ 20%] 2023-03-31T06:51:18.8581976Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:18.8582323Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:18.8582656Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_bfloat16 PASSED [ 20%] 2023-03-31T06:51:18.8583040Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_bool PASSED [ 20%] 2023-03-31T06:51:18.8583468Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:18.8583804Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:18.8584152Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_bfloat16 PASSED [ 20%] 2023-03-31T06:51:18.8584508Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_complex64 PASSED [ 20%] 2023-03-31T06:51:18.8584896Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float16 PASSED [ 20%] 2023-03-31T06:51:18.8585256Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:18.8585603Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:18.8585972Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:18.8586302Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8586647Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_bool PASSED [ 21%] 2023-03-31T06:51:18.8586996Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_complex64 PASSED [ 21%] 2023-03-31T06:51:18.8587349Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:18.8587680Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:18.8588026Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:18.8588370Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8588706Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8589048Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_bool PASSED [ 21%] 2023-03-31T06:51:18.8589396Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex32 PASSED [ 21%] 2023-03-31T06:51:18.8589746Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex64 PASSED [ 21%] 2023-03-31T06:51:18.8590080Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:18.8590424Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:18.8590766Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:18.8591093Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8591444Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8591808Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_complex128 PASSED [ 21%] 2023-03-31T06:51:18.8592176Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_complex64 PASSED [ 21%] 2023-03-31T06:51:18.8592528Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:18.8592886Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:18.8593238Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8593576Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:18.8593927Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8594277Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:18.8594619Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:18.8594943Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:18.8595282Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:18.8595635Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8596019Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:18.8596382Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:18.8596733Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8597083Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8597450Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_bool PASSED [ 21%] 2023-03-31T06:51:18.8597818Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:18.8598158Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:18.8598481Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:18.8598860Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:18.8599194Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:18.8599538Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:18.8599865Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_bool PASSED [ 22%] 2023-03-31T06:51:18.8600205Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:18.8600543Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:18.8600866Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int64 PASSED [ 22%] 2023-03-31T06:51:18.8601202Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8601536Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:18.8601877Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:18.8602201Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int16 PASSED [ 22%] 2023-03-31T06:51:18.8602536Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:18.8602868Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8603473Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:18.8603818Z test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float16 PASSED [ 22%] 2023-03-31T06:51:18.8604167Z test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:18.8604512Z test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:18.8604854Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_bool PASSED [ 22%] 2023-03-31T06:51:18.8605200Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex32 PASSED [ 22%] 2023-03-31T06:51:18.8605536Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8605873Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:18.8606217Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_bool PASSED [ 22%] 2023-03-31T06:51:18.8606555Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float16 PASSED [ 22%] 2023-03-31T06:51:18.8606906Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:18.8607256Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:18.8607602Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8607934Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:18.8608278Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_bool PASSED [ 22%] 2023-03-31T06:51:18.8608629Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_complex128 PASSED [ 22%] 2023-03-31T06:51:18.8609045Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_complex64 PASSED [ 22%] 2023-03-31T06:51:18.8609399Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:18.8609743Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int16 PASSED [ 22%] 2023-03-31T06:51:18.8610085Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int64 PASSED [ 22%] 2023-03-31T06:51:18.8610448Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8610825Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:18.8611169Z test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:18.8611487Z test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int64 PASSED [ 22%] 2023-03-31T06:51:18.8611854Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_bool PASSED [ 22%] 2023-03-31T06:51:18.8612190Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:18.8612525Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int16 PASSED [ 22%] 2023-03-31T06:51:18.8612843Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:18.8613171Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:18.8613516Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:18.8613863Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:18.8614221Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:18.8614566Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:18.8614917Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:18.8615254Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:18.8615603Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:18.8615951Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_int8 PASSED [ 23%] 2023-03-31T06:51:18.8616287Z test_decomp.py::TestDecompCPU::test_comprehensive_geometric_cpu_uint8 PASSED [ 23%] 2023-03-31T06:51:18.8616636Z test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_complex128 PASSED [ 23%] 2023-03-31T06:51:18.8616989Z test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_complex64 PASSED [ 23%] 2023-03-31T06:51:18.8617342Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:18.8617689Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_complex128 PASSED [ 23%] 2023-03-31T06:51:18.8618052Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_complex64 PASSED [ 23%] 2023-03-31T06:51:18.8618407Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:18.8618744Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:18.8619095Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:18.8619439Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:18.8619784Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:18.8620130Z test_decomp.py::TestDecompCPU::test_comprehensive_grid_sampler_2d_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:18.8620486Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_bool PASSED [ 23%] 2023-03-31T06:51:18.8620823Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:18.8621148Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:18.8621518Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:18.8621849Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:18.8622180Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:18.8622507Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_complex128 PASSED [ 23%] 2023-03-31T06:51:18.8622851Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:18.8623218Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:18.8623657Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:18.8623991Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:18.8624366Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_bool PASSED [ 23%] 2023-03-31T06:51:18.8624719Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:18.8625062Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:18.8625407Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:18.8625755Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:18.8626088Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_int8 PASSED [ 23%] 2023-03-31T06:51:18.8626438Z test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:18.8626790Z test_decomp.py::TestDecompCPU::test_comprehensive_histogram_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:18.8627140Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:18.8627471Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_bool PASSED [ 24%] 2023-03-31T06:51:18.8627824Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:18.8628183Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:18.8628519Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:18.8628863Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:18.8629207Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:18.8629559Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:18.8629903Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex32 PASSED [ 24%] 2023-03-31T06:51:18.8630251Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:18.8630601Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:18.8630935Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int32 PASSED [ 24%] 2023-03-31T06:51:18.8631278Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:18.8631622Z test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:18.8631963Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:18.8632290Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:18.8632619Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_uint8 PASSED [ 24%] 2023-03-31T06:51:18.8632965Z test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:18.8633306Z test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:18.8633654Z test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:18.8634008Z test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:18.8634399Z test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex32 PASSED [ 24%] 2023-03-31T06:51:18.8634735Z test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:18.8635088Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:18.8635441Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:18.8635775Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int16 PASSED [ 24%] 2023-03-31T06:51:18.8636152Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:18.8636534Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:18.8636883Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_bool PASSED [ 24%] 2023-03-31T06:51:18.8637261Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex32 PASSED [ 24%] 2023-03-31T06:51:18.8637624Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:18.8637991Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float32 PASSED [ 24%] 2023-03-31T06:51:18.8638348Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:18.8638688Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int32 PASSED [ 24%] 2023-03-31T06:51:18.8639039Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_uint8 PASSED [ 24%] 2023-03-31T06:51:18.8639395Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_bool PASSED [ 24%] 2023-03-31T06:51:18.8639741Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex32 PASSED [ 24%] 2023-03-31T06:51:18.8640101Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:18.8640462Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float32 PASSED [ 24%] 2023-03-31T06:51:18.8640816Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:18.8641156Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8641512Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex32 PASSED [ 25%] 2023-03-31T06:51:18.8641874Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float16 PASSED [ 25%] 2023-03-31T06:51:18.8642217Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:18.8642576Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8642938Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:18.8643483Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:18.8643830Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:18.8644189Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8644549Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_bool PASSED [ 25%] 2023-03-31T06:51:18.8644896Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex32 PASSED [ 25%] 2023-03-31T06:51:18.8645264Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:18.8645627Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:18.8645986Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:18.8646325Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:18.8646681Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:18.8647072Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int8 PASSED [ 25%] 2023-03-31T06:51:18.8647403Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8647740Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_bool PASSED [ 25%] 2023-03-31T06:51:18.8648084Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_complex128 PASSED [ 25%] 2023-03-31T06:51:18.8648428Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:18.8648799Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:18.8649163Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:18.8649508Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8649843Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_bool PASSED [ 25%] 2023-03-31T06:51:18.8650235Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:18.8650589Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:18.8650936Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:18.8651274Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:18.8651630Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float16 PASSED [ 25%] 2023-03-31T06:51:18.8651986Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int8 PASSED [ 25%] 2023-03-31T06:51:18.8652322Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:18.8652483Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:18.8652640Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int8 PASSED [ 25%] 2023-03-31T06:51:18.8652806Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:18.8652968Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_bool PASSED [ 25%] 2023-03-31T06:51:18.8653129Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:18.8653287Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:18.8653445Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:18.8653590Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int64 PASSED [ 26%] 2023-03-31T06:51:18.8653747Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_int8 PASSED [ 26%] 2023-03-31T06:51:18.8653904Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_uint8 PASSED [ 26%] 2023-03-31T06:51:18.8654060Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_bool PASSED [ 26%] 2023-03-31T06:51:18.8654222Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float64 PASSED [ 26%] 2023-03-31T06:51:18.8654378Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int16 PASSED [ 26%] 2023-03-31T06:51:18.8654535Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int64 PASSED [ 26%] 2023-03-31T06:51:18.8654691Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int8 PASSED [ 26%] 2023-03-31T06:51:18.8654860Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_bfloat16 PASSED [ 26%] 2023-03-31T06:51:18.8655011Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_bool PASSED [ 26%] 2023-03-31T06:51:18.8655175Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int16 PASSED [ 26%] 2023-03-31T06:51:18.8655338Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int32 PASSED [ 26%] 2023-03-31T06:51:18.8655498Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int64 PASSED [ 26%] 2023-03-31T06:51:18.8655664Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_uint8 PASSED [ 26%] 2023-03-31T06:51:18.8655861Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_bool PASSED [ 26%] 2023-03-31T06:51:18.8656031Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float16 PASSED [ 26%] 2023-03-31T06:51:18.8656195Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float32 PASSED [ 26%] 2023-03-31T06:51:18.8656346Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int32 PASSED [ 26%] 2023-03-31T06:51:18.8656506Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int8 PASSED [ 26%] 2023-03-31T06:51:18.8656693Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_uint8 PASSED [ 26%] 2023-03-31T06:51:18.8656883Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_bfloat16 PASSED [ 26%] 2023-03-31T06:51:18.8657052Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex128 PASSED [ 26%] 2023-03-31T06:51:18.8657256Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex32 PASSED [ 26%] 2023-03-31T06:51:18.8657423Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_complex64 PASSED [ 26%] 2023-03-31T06:51:18.8657583Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int16 PASSED [ 26%] 2023-03-31T06:51:18.8657731Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_uint8 PASSED [ 26%] 2023-03-31T06:51:18.8657897Z test_decomp.py::TestDecompCPU::test_comprehensive_istft_cpu_complex128 PASSED [ 26%] 2023-03-31T06:51:18.8658112Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8658328Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8658538Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8658742Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int32 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8658950Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8659158Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_uint8 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8659377Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8659600Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_complex128 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8659811Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_complex64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8660025Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8660240Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8660455Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:18.8660669Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_uint8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8660866Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_bool SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8661071Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_complex128 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8661274Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_complex64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8661468Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8661685Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8661883Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8662077Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_uint8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8662291Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8662526Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_bool SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8662811Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_complex128 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8663022Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8663272Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8663575Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_float64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8663787Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8663984Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8664195Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8664400Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_uint8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8664599Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8664794Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_bool SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8664993Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_complex64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8665189Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8665391Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int64 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8665586Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8665775Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_uint8 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:18.8665930Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:18.8666091Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_bool PASSED [ 27%] 2023-03-31T06:51:18.8666256Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float32 PASSED [ 27%] 2023-03-31T06:51:18.8666416Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:18.8666571Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int64 PASSED [ 27%] 2023-03-31T06:51:18.8666728Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int8 PASSED [ 27%] 2023-03-31T06:51:18.8666901Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:18.8667068Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int64 PASSED [ 27%] 2023-03-31T06:51:18.8667217Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int8 PASSED [ 27%] 2023-03-31T06:51:18.8667370Z test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int32 PASSED [ 27%] 2023-03-31T06:51:18.8667527Z test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_uint8 PASSED [ 27%] 2023-03-31T06:51:18.8667731Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:18.8667893Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_bool PASSED [ 27%] 2023-03-31T06:51:18.8668062Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_complex128 PASSED [ 27%] 2023-03-31T06:51:18.8668221Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int32 PASSED [ 28%] 2023-03-31T06:51:18.8668401Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int64 PASSED [ 28%] 2023-03-31T06:51:18.8668569Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_uint8 PASSED [ 28%] 2023-03-31T06:51:18.8668724Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8668879Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int32 PASSED [ 28%] 2023-03-31T06:51:18.8669066Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int64 PASSED [ 28%] 2023-03-31T06:51:18.8669223Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int8 PASSED [ 28%] 2023-03-31T06:51:18.8669386Z test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:18.8669547Z test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8669703Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_bool PASSED [ 28%] 2023-03-31T06:51:18.8669850Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8670012Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8670176Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_uint8 PASSED [ 28%] 2023-03-31T06:51:18.8670348Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:18.8670521Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_bfloat16 PASSED [ 28%] 2023-03-31T06:51:18.8670696Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:18.8670867Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8671035Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int16 PASSED [ 28%] 2023-03-31T06:51:18.8671199Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int32 PASSED [ 28%] 2023-03-31T06:51:18.8671354Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_uint8 PASSED [ 28%] 2023-03-31T06:51:18.8671528Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:18.8671696Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8671879Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:18.8672061Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8672235Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_bool PASSED [ 28%] 2023-03-31T06:51:18.8672414Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:18.8672592Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:18.8672754Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float16 PASSED [ 28%] 2023-03-31T06:51:18.8672930Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8673102Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int16 PASSED [ 28%] 2023-03-31T06:51:18.8673273Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int32 PASSED [ 28%] 2023-03-31T06:51:18.8673443Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int64 PASSED [ 28%] 2023-03-31T06:51:18.8673648Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_int8 PASSED [ 28%] 2023-03-31T06:51:18.8673820Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:18.8673986Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8674141Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8674317Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:18.8674514Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8674713Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:18.8674887Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:18.8675092Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8675273Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8675448Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8675645Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8675804Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8675972Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8676148Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8676320Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8676504Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8676686Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8676867Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8677044Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8677217Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8677377Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8677575Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8677744Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8677908Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8678087Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8678273Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8678450Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8678627Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8678793Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8678969Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8679150Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8679332Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8679510Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8679724Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8679908Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8680089Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:18.8680264Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8680472Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8680688Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:18.8680880Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8681084Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_bfloat16 PASSED [ 29%] 2023-03-31T06:51:18.8681268Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8681445Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:18.8681615Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int32 PASSED [ 29%] 2023-03-31T06:51:18.8681784Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int64 PASSED [ 29%] 2023-03-31T06:51:18.8681959Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:18.8682120Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8682288Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float16 PASSED [ 30%] 2023-03-31T06:51:18.8682460Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8682661Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:51:18.8682863Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8683244Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8683420Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8683593Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8683772Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8683956Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:18.8684215Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 30%] 2023-03-31T06:51:18.8684466Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 30%] 2023-03-31T06:51:18.8684715Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 30%] 2023-03-31T06:51:18.8684897Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8685077Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8685253Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:18.8685432Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8685605Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8685821Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8686000Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8686172Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8686362Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8686587Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:18.8686808Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8686980Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8687185Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8687352Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8687532Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8687705Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:18.8687878Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:18.8688066Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8688246Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:18.8688425Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:18.8688596Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:18.8688768Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int16 PASSED [ 30%] 2023-03-31T06:51:18.8688925Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int8 PASSED [ 30%] 2023-03-31T06:51:18.8689095Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_uint8 PASSED [ 30%] 2023-03-31T06:51:18.8689267Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:51:18.8689445Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:51:18.8689615Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:18.8689789Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:18.8689958Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:18.8690125Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:18.8690279Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:18.8690440Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:18.8690599Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:18.8690758Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:18.8690912Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:18.8691072Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_bool PASSED [ 31%] 2023-03-31T06:51:18.8691241Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:18.8691408Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:18.8691557Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:18.8691715Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:18.8691905Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:18.8692068Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_bool PASSED [ 31%] 2023-03-31T06:51:18.8692224Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int8 PASSED [ 31%] 2023-03-31T06:51:18.8692382Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:18.8692581Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:18.8692777Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:18.8692944Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_complex64 PASSED [ 31%] 2023-03-31T06:51:18.8693092Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:18.8693274Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:18.8693433Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:18.8693597Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:18.8693757Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:18.8693912Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:18.8694067Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int8 PASSED [ 31%] 2023-03-31T06:51:18.8694225Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:18.8694385Z test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:18.8694555Z test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:18.8694733Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:18.8694904Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:18.8695083Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_bool PASSED [ 31%] 2023-03-31T06:51:18.8695275Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex32 PASSED [ 31%] 2023-03-31T06:51:18.8695458Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex64 PASSED [ 31%] 2023-03-31T06:51:18.8695644Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:18.8695813Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:18.8695994Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:18.8696173Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:18.8696346Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:18.8696517Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:18.8696685Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:18.8711664Z test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:18.8711986Z test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8712162Z test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:18.8712339Z test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:18.8712507Z test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:18.8712673Z test_decomp.py::TestDecompCPU::test_comprehensive_logdet_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8712985Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_bool PASSED [ 32%] 2023-03-31T06:51:18.8713164Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:18.8713338Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:18.8713510Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float16 PASSED [ 32%] 2023-03-31T06:51:18.8713681Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8713870Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:18.8714064Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:18.8714227Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:18.8714425Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_bool PASSED [ 32%] 2023-03-31T06:51:18.8714602Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:18.8714773Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float16 PASSED [ 32%] 2023-03-31T06:51:18.8714940Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:18.8715107Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8715277Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:18.8715450Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:18.8715617Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:18.8715770Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int32 PASSED [ 32%] 2023-03-31T06:51:18.8715937Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:18.8716112Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:18.8716278Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float16 PASSED [ 32%] 2023-03-31T06:51:18.8716449Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8716615Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int32 PASSED [ 32%] 2023-03-31T06:51:18.8716781Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:18.8716945Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_bool PASSED [ 32%] 2023-03-31T06:51:18.8717098Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:18.8717259Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:18.8717421Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int32 PASSED [ 32%] 2023-03-31T06:51:18.8717579Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:18.8717739Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:18.8717911Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:18.8718082Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:18.8718257Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:18.8718412Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int32 PASSED [ 33%] 2023-03-31T06:51:18.8718573Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:18.8718743Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:18.8718910Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:18.8719111Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int32 PASSED [ 33%] 2023-03-31T06:51:18.8719275Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:18.8719438Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:18.8719599Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float16 PASSED [ 33%] 2023-03-31T06:51:18.8719775Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:18.8719966Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int16 PASSED [ 33%] 2023-03-31T06:51:18.8720124Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int32 PASSED [ 33%] 2023-03-31T06:51:18.8720284Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int8 PASSED [ 33%] 2023-03-31T06:51:18.8720471Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_uint8 PASSED [ 33%] 2023-03-31T06:51:18.8720629Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_bool PASSED [ 33%] 2023-03-31T06:51:18.8720786Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float16 PASSED [ 33%] 2023-03-31T06:51:18.8720942Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:18.8721084Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int16 PASSED [ 33%] 2023-03-31T06:51:18.8721239Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int8 PASSED [ 33%] 2023-03-31T06:51:18.8721394Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_uint8 PASSED [ 33%] 2023-03-31T06:51:18.8721559Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:18.8721715Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:18.8721887Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:18.8722055Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:18.8722224Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:18.8722382Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:18.8722532Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:18.8722686Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float16 PASSED [ 33%] 2023-03-31T06:51:18.8722845Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:18.8723001Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int32 PASSED [ 33%] 2023-03-31T06:51:18.8723476Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:18.8723632Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int8 PASSED [ 33%] 2023-03-31T06:51:18.8723789Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_uint8 PASSED [ 33%] 2023-03-31T06:51:18.8723946Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:18.8724086Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_bool PASSED [ 33%] 2023-03-31T06:51:18.8724250Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:18.8724407Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float16 PASSED [ 33%] 2023-03-31T06:51:18.8724562Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int16 PASSED [ 33%] 2023-03-31T06:51:18.8724716Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:18.8724868Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:18.8725041Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:18.8725215Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:18.8725464Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:18.8725637Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:18.8725804Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:18.8725965Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:18.8726202Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:18.8726401Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:18.8726577Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:18.8726748Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:18.8726949Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:18.8727122Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:18.8727292Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int64 PASSED [ 34%] 2023-03-31T06:51:18.8727459Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:18.8727627Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:18.8727800Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:18.8727975Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:18.8728144Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:18.8728312Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:18.8728470Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:18.8728648Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:18.8728822Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:18.8728991Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:18.8729158Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:18.8729328Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int64 PASSED [ 34%] 2023-03-31T06:51:18.8729498Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:18.8729667Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:18.8729824Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_bool PASSED [ 34%] 2023-03-31T06:51:18.8729998Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex32 PASSED [ 34%] 2023-03-31T06:51:18.8730169Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex64 PASSED [ 34%] 2023-03-31T06:51:18.8730336Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:18.8730504Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:18.8730689Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:18.8730870Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:18.8731070Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_bfloat16 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:18.8731264Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_float32 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:18.8731484Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int64 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:18.8731679Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_uint8 SKIPPED (Skipped!) [ 35%] 2023-03-31T06:51:18.8731852Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8732018Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_bool PASSED [ 35%] 2023-03-31T06:51:18.8732187Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8732385Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:18.8732574Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:18.8732746Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8732932Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8733100Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:18.8733278Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8733458Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:18.8733636Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float16 PASSED [ 35%] 2023-03-31T06:51:18.8733814Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:18.8733988Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8734158Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_bool PASSED [ 35%] 2023-03-31T06:51:18.8734332Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:18.8734490Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8734657Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int32 PASSED [ 35%] 2023-03-31T06:51:18.8734819Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:18.8734998Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:18.8735168Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:18.8735340Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int32 PASSED [ 35%] 2023-03-31T06:51:18.8735512Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:18.8735679Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_int8 PASSED [ 35%] 2023-03-31T06:51:18.8735842Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8736015Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8736179Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:18.8736346Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int32 PASSED [ 35%] 2023-03-31T06:51:18.8736511Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:18.8736681Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_int8 PASSED [ 35%] 2023-03-31T06:51:18.8736858Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:18.8737035Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8737193Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8737371Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_complex128 PASSED [ 35%] 2023-03-31T06:51:18.8737571Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:18.8737740Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:18.8737906Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:18.8738072Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_bool PASSED [ 35%] 2023-03-31T06:51:18.8738272Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_complex128 PASSED [ 35%] 2023-03-31T06:51:18.8738463Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:18.8738625Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:18.8738772Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:18.8738966Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:18.8739143Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_complex128 PASSED [ 36%] 2023-03-31T06:51:18.8739312Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_complex64 PASSED [ 36%] 2023-03-31T06:51:18.8739480Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:18.8739644Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:18.8739814Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_complex128 PASSED [ 36%] 2023-03-31T06:51:18.8739979Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:18.8740126Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:18.8740287Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int16 PASSED [ 36%] 2023-03-31T06:51:18.8740450Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:18.8740621Z test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:18.8740787Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_bool PASSED [ 36%] 2023-03-31T06:51:18.8740953Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:18.8741120Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:18.8741322Z test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:18.8741522Z test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:18.8741690Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float16 PASSED [ 36%] 2023-03-31T06:51:18.8741871Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int16 PASSED [ 36%] 2023-03-31T06:51:18.8742050Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:18.8742228Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:18.8742407Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:18.8742585Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:18.8742754Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:18.8742921Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_bool PASSED [ 36%] 2023-03-31T06:51:18.8743076Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float16 PASSED [ 36%] 2023-03-31T06:51:18.8743236Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:18.8743497Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:18.8743701Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int16 PASSED [ 36%] 2023-03-31T06:51:18.8743872Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:18.8744037Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:18.8744201Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_complex128 PASSED [ 36%] 2023-03-31T06:51:18.8744395Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_complex64 PASSED [ 36%] 2023-03-31T06:51:18.8744567Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:18.8744735Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:18.8744896Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:18.8745087Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:18.8745251Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:18.8745411Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:18.8745650Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float16 SKIPPED (meshgrid in torch.float16 not supported) [ 37%] 2023-03-31T06:51:18.8745887Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float32 SKIPPED (meshgrid in torch.float32 not supported) [ 37%] 2023-03-31T06:51:18.8746124Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int16 SKIPPED (meshgrid in torch.int16 not supported) [ 37%] 2023-03-31T06:51:18.8746347Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_bool SKIPPED (meshgrid in torch.bool not supported) [ 37%] 2023-03-31T06:51:18.8746601Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_complex128 SKIPPED (meshgrid in torch.complex128 not supported) [ 37%] 2023-03-31T06:51:18.8746840Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float16 SKIPPED (meshgrid in torch.float16 not supported) [ 37%] 2023-03-31T06:51:18.8747078Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float64 SKIPPED (meshgrid in torch.float64 not supported) [ 37%] 2023-03-31T06:51:18.8747314Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int16 SKIPPED (meshgrid in torch.int16 not supported) [ 37%] 2023-03-31T06:51:18.8747545Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int32 SKIPPED (meshgrid in torch.int32 not supported) [ 37%] 2023-03-31T06:51:18.8747774Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_uint8 SKIPPED (meshgrid in torch.uint8 not supported) [ 37%] 2023-03-31T06:51:18.8747946Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:18.8748115Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:18.8748278Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int32 PASSED [ 37%] 2023-03-31T06:51:18.8748445Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float64 PASSED [ 37%] 2023-03-31T06:51:18.8748622Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:18.8748802Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:18.8748981Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:18.8749160Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_bool PASSED [ 37%] 2023-03-31T06:51:18.8749344Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float16 PASSED [ 37%] 2023-03-31T06:51:18.8749555Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:18.8749735Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:18.8749915Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:18.8750069Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:18.8750234Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float16 PASSED [ 37%] 2023-03-31T06:51:18.8750428Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:18.8750616Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_float64 PASSED [ 37%] 2023-03-31T06:51:18.8750778Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:18.8750964Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:18.8751124Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:18.8751282Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:18.8751429Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:18.8751590Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:18.8751748Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:18.8751904Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:18.8752061Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:18.8752221Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:18.8752391Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:18.8752565Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex128 PASSED [ 37%] 2023-03-31T06:51:18.8752722Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex32 PASSED [ 38%] 2023-03-31T06:51:18.8752886Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:18.8753048Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8753209Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int8 PASSED [ 38%] 2023-03-31T06:51:18.8753373Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float16 PASSED [ 38%] 2023-03-31T06:51:18.8753536Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8753696Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:18.8753858Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8754007Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:18.8754175Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex128 PASSED [ 38%] 2023-03-31T06:51:18.8754333Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex32 PASSED [ 38%] 2023-03-31T06:51:18.8754491Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float16 PASSED [ 38%] 2023-03-31T06:51:18.8754648Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8754806Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:18.8754964Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8755126Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_complex128 PASSED [ 38%] 2023-03-31T06:51:18.8755289Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_complex64 PASSED [ 38%] 2023-03-31T06:51:18.8755435Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8755618Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:18.8755774Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8755927Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:18.8756081Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int8 PASSED [ 38%] 2023-03-31T06:51:18.8756234Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:18.8756446Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8756661Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:18.8756829Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:18.8757040Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8757219Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:18.8757397Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_int8 PASSED [ 38%] 2023-03-31T06:51:18.8757577Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8757751Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:18.8757937Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:18.8758119Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:18.8758297Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:18.8758464Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:18.8758631Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_bool PASSED [ 38%] 2023-03-31T06:51:18.8758798Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:18.8758962Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:18.8759128Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float16 PASSED [ 38%] 2023-03-31T06:51:18.8759294Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:18.8759454Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:18.8759626Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8759781Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:18.8759950Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:18.8760118Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:18.8760277Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:18.8760448Z test_decomp.py::TestDecompCPU::test_comprehensive_nanquantile_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:18.8760611Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8760775Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_bool PASSED [ 39%] 2023-03-31T06:51:18.8760942Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:18.8761091Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:18.8761253Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:18.8761428Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8761632Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_bool PASSED [ 39%] 2023-03-31T06:51:18.8761809Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex128 PASSED [ 39%] 2023-03-31T06:51:18.8761980Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex64 PASSED [ 39%] 2023-03-31T06:51:18.8762149Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:18.8762343Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:18.8762520Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:18.8762686Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8762848Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_bool PASSED [ 39%] 2023-03-31T06:51:18.8763225Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex32 PASSED [ 39%] 2023-03-31T06:51:18.8763395Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex64 PASSED [ 39%] 2023-03-31T06:51:18.8763563Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:18.8763726Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:18.8763892Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:18.8764055Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:18.8764204Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:18.8764452Z test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_bfloat16 SKIPPED (native_batch_norm in torch.bfloat16 not supported) [ 39%] 2023-03-31T06:51:18.8764691Z test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_float64 SKIPPED (native_batch_norm in torch.float64 not supported) [ 39%] 2023-03-31T06:51:18.8764882Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:18.8765064Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:18.8765246Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:18.8765432Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:18.8765613Z test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8765776Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:18.8765920Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_bool PASSED [ 39%] 2023-03-31T06:51:18.8766088Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_complex128 PASSED [ 39%] 2023-03-31T06:51:18.8766250Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_complex64 PASSED [ 39%] 2023-03-31T06:51:18.8766407Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float16 PASSED [ 40%] 2023-03-31T06:51:18.8766564Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float32 PASSED [ 40%] 2023-03-31T06:51:18.8766720Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int16 PASSED [ 40%] 2023-03-31T06:51:18.8766874Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int32 PASSED [ 40%] 2023-03-31T06:51:18.8767032Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_uint8 PASSED [ 40%] 2023-03-31T06:51:18.8767185Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_bfloat16 PASSED [ 40%] 2023-03-31T06:51:18.8767352Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex128 PASSED [ 40%] 2023-03-31T06:51:18.8767516Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex32 PASSED [ 40%] 2023-03-31T06:51:18.8767679Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_complex64 PASSED [ 40%] 2023-03-31T06:51:18.8767910Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float16 PASSED [ 40%] 2023-03-31T06:51:18.8768072Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float64 PASSED [ 40%] 2023-03-31T06:51:18.8768230Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int16 PASSED [ 40%] 2023-03-31T06:51:18.8768388Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_uint8 PASSED [ 40%] 2023-03-31T06:51:18.8768644Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float32 SKIPPED (new_empty in torch.float32 not supported) [ 40%] 2023-03-31T06:51:18.8768905Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float64 SKIPPED (new_empty in torch.float64 not supported) [ 40%] 2023-03-31T06:51:18.8769121Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int32 SKIPPED (new_empty in torch.int32 not supported) [ 40%] 2023-03-31T06:51:18.8769381Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int8 SKIPPED (new_empty in torch.int8 not supported) [ 40%] 2023-03-31T06:51:18.8769592Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_uint8 SKIPPED (new_empty in torch.uint8 not supported) [ 40%] 2023-03-31T06:51:18.8769826Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_bfloat16 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8770063Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_bool SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8770308Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex128 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8770547Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex32 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8770784Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int16 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8771002Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int32 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8771226Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_uint8 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:18.8771399Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex128 PASSED [ 40%] 2023-03-31T06:51:18.8771575Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex64 PASSED [ 40%] 2023-03-31T06:51:18.8771744Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float16 PASSED [ 40%] 2023-03-31T06:51:18.8771908Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int16 PASSED [ 40%] 2023-03-31T06:51:18.8772070Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int32 PASSED [ 40%] 2023-03-31T06:51:18.8772229Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int64 PASSED [ 40%] 2023-03-31T06:51:18.8772395Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_int8 PASSED [ 40%] 2023-03-31T06:51:18.8772545Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_bool PASSED [ 40%] 2023-03-31T06:51:18.8772717Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex32 PASSED [ 40%] 2023-03-31T06:51:18.8772883Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float16 PASSED [ 40%] 2023-03-31T06:51:18.8773049Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float32 PASSED [ 40%] 2023-03-31T06:51:18.8773217Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int32 PASSED [ 40%] 2023-03-31T06:51:18.8773375Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int64 PASSED [ 40%] 2023-03-31T06:51:18.8773537Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int8 PASSED [ 40%] 2023-03-31T06:51:18.8773771Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_uint8 PASSED [ 41%] 2023-03-31T06:51:18.8773927Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8774097Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int16 PASSED [ 41%] 2023-03-31T06:51:18.8774259Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:18.8774421Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int8 PASSED [ 41%] 2023-03-31T06:51:18.8774606Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_uint8 PASSED [ 41%] 2023-03-31T06:51:18.8774802Z test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8775003Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8775235Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8775436Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float16 PASSED [ 41%] 2023-03-31T06:51:18.8775617Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8775811Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool3d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8776003Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool3d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8776201Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8776393Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8776577Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:18.8776765Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8776956Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8777136Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8777320Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8777534Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8777719Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8777898Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8778079Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_celu_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8778264Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8778450Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_complex64 PASSED [ 41%] 2023-03-31T06:51:18.8778631Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8778815Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_complex128 PASSED [ 41%] 2023-03-31T06:51:18.8778986Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_complex64 PASSED [ 41%] 2023-03-31T06:51:18.8779171Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8779368Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8779569Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_complex128 PASSED [ 41%] 2023-03-31T06:51:18.8779806Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_complex64 PASSED [ 41%] 2023-03-31T06:51:18.8780003Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:18.8780195Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8780390Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:18.8780645Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_complex64 PASSED [ 41%] 2023-03-31T06:51:18.8780853Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:18.8781044Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:18.8781269Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8781464Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:18.8781664Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8781860Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_bool PASSED [ 42%] 2023-03-31T06:51:18.8782059Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8782254Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int16 PASSED [ 42%] 2023-03-31T06:51:18.8782447Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int32 PASSED [ 42%] 2023-03-31T06:51:18.8782628Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_uint8 PASSED [ 42%] 2023-03-31T06:51:18.8782827Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8783019Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8783210Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8783500Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8783693Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_ctc_loss_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8783887Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8784072Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8784250Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8784417Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8784610Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8784797Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float16 PASSED [ 42%] 2023-03-31T06:51:18.8784979Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8785200Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8785419Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8785637Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8785892Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int16 PASSED [ 42%] 2023-03-31T06:51:18.8786109Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:18.8786307Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int8 PASSED [ 42%] 2023-03-31T06:51:18.8786517Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_uint8 PASSED [ 42%] 2023-03-31T06:51:18.8786745Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool2d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8786974Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool2d_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8787199Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8787392Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8787585Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gaussian_nll_loss_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8787765Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8787942Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8788109Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_gelu_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8788299Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_grid_sample_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8788486Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8788671Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:18.8788853Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:18.8789035Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:18.8789212Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardswish_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8789388Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8789575Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8789746Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int32 PASSED [ 43%] 2023-03-31T06:51:18.8789929Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int64 PASSED [ 43%] 2023-03-31T06:51:18.8790114Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int8 PASSED [ 43%] 2023-03-31T06:51:18.8790315Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8790512Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8790697Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8790881Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float16 PASSED [ 43%] 2023-03-31T06:51:18.8791065Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8791245Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_huber_loss_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8791424Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8791621Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_area_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8791847Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8792046Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8792246Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8792478Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8792698Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_uint8 PASSED [ 43%] 2023-03-31T06:51:18.8792900Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8793126Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8793316Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_trilinear_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8793501Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_complex64 PASSED [ 43%] 2023-03-31T06:51:18.8793679Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float16 PASSED [ 43%] 2023-03-31T06:51:18.8793865Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8794054Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8794241Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8794429Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_complex64 PASSED [ 43%] 2023-03-31T06:51:18.8794610Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8794795Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8794960Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int32 PASSED [ 43%] 2023-03-31T06:51:18.8795136Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int64 PASSED [ 43%] 2023-03-31T06:51:18.8795334Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8795530Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8795726Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_int64 PASSED [ 43%] 2023-03-31T06:51:18.8795910Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8796094Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:18.8796293Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:18.8796487Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:18.8796668Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int16 PASSED [ 44%] 2023-03-31T06:51:18.8796859Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int32 PASSED [ 44%] 2023-03-31T06:51:18.8797054Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int8 PASSED [ 44%] 2023-03-31T06:51:18.8797240Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8797422Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8797634Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool3d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8797822Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8798012Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_grad_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8798201Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8798406Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_grad_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8798625Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_grad_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8798814Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_grad_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8799027Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:18.8799209Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float16 PASSED [ 44%] 2023-03-31T06:51:18.8799387Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8799592Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:18.8799802Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8800005Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8800180Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:18.8800367Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:18.8800561Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_complex128 PASSED [ 44%] 2023-03-31T06:51:18.8800752Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_complex64 PASSED [ 44%] 2023-03-31T06:51:18.8800935Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8801111Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_one_hot_cpu_int64 PASSED [ 44%] 2023-03-31T06:51:18.8801363Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_bfloat16 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:18.8801611Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float16 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:18.8801852Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int32 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:18.8802096Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int8 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:18.8802275Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:18.8802468Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_complex128 PASSED [ 44%] 2023-03-31T06:51:18.8802659Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_complex64 PASSED [ 44%] 2023-03-31T06:51:18.8802852Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8803260Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int16 PASSED [ 44%] 2023-03-31T06:51:18.8803531Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int8 PASSED [ 44%] 2023-03-31T06:51:18.8803730Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_complex128 PASSED [ 44%] 2023-03-31T06:51:18.8803995Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:18.8804190Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_complex64 PASSED [ 44%] 2023-03-31T06:51:18.8804368Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:18.8804598Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float16 PASSED [ 44%] 2023-03-31T06:51:18.8804827Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:18.8805017Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int16 PASSED [ 45%] 2023-03-31T06:51:18.8805241Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int32 PASSED [ 45%] 2023-03-31T06:51:18.8805431Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:18.8805625Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_int8 PASSED [ 45%] 2023-03-31T06:51:18.8805803Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pdist_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8805991Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_bool PASSED [ 45%] 2023-03-31T06:51:18.8806173Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_complex64 PASSED [ 45%] 2023-03-31T06:51:18.8806365Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8806554Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:18.8806744Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int16 PASSED [ 45%] 2023-03-31T06:51:18.8806933Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int8 PASSED [ 45%] 2023-03-31T06:51:18.8807116Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_uint8 PASSED [ 45%] 2023-03-31T06:51:18.8807311Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:18.8807505Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_complex64 PASSED [ 45%] 2023-03-31T06:51:18.8807696Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8807877Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:18.8808065Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int16 PASSED [ 45%] 2023-03-31T06:51:18.8808257Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int8 PASSED [ 45%] 2023-03-31T06:51:18.8808449Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:18.8808641Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8808829Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int32 PASSED [ 45%] 2023-03-31T06:51:18.8809012Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:18.8809259Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_float64 SKIPPED (nn.functional.relu6 in torch.float64 not supported) [ 45%] 2023-03-31T06:51:18.8809495Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int16 SKIPPED (nn.functional.relu6 in torch.int16 not supported) [ 45%] 2023-03-31T06:51:18.8809753Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int32 SKIPPED (nn.functional.relu6 in torch.int32 not supported) [ 45%] 2023-03-31T06:51:18.8809930Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:18.8810109Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8810328Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_float32 SKIPPED (Skipped) [ 45%] 2023-03-31T06:51:18.8810534Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8810739Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:18.8810928Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_complex_cpu_complex128 PASSED [ 45%] 2023-03-31T06:51:18.8811141Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8811328Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:18.8811506Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float16 PASSED [ 45%] 2023-03-31T06:51:18.8811698Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8811886Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:18.8812084Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_complex64 PASSED [ 45%] 2023-03-31T06:51:18.8812285Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:18.8812477Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int64 PASSED [ 46%] 2023-03-31T06:51:18.8812672Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8812861Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:18.8813046Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:18.8813220Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_complex128 PASSED [ 46%] 2023-03-31T06:51:18.8813407Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_complex64 PASSED [ 46%] 2023-03-31T06:51:18.8813590Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float16 PASSED [ 46%] 2023-03-31T06:51:18.8813772Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:18.8813955Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:18.8814137Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:18.8814322Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8814513Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_complex128 PASSED [ 46%] 2023-03-31T06:51:18.8814702Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_complex64 PASSED [ 46%] 2023-03-31T06:51:18.8814877Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:18.8815060Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:18.8815245Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int32 PASSED [ 46%] 2023-03-31T06:51:18.8815427Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:18.8815610Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:18.8815826Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:18.8816010Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8816207Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:18.8816404Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:18.8816610Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:18.8816830Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int32 PASSED [ 46%] 2023-03-31T06:51:18.8817089Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:18.8817305Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:18.8817515Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:18.8817723Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int32 PASSED [ 46%] 2023-03-31T06:51:18.8817933Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8818120Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:18.8818307Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:18.8818492Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:18.8818695Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:18.8818889Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8819083Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:18.8819275Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:18.8819448Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex128 PASSED [ 46%] 2023-03-31T06:51:18.8819622Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex64 PASSED [ 46%] 2023-03-31T06:51:18.8819788Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:18.8819953Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:18.8820153Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 47%] 2023-03-31T06:51:18.8820371Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 47%] 2023-03-31T06:51:18.8820587Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 47%] 2023-03-31T06:51:18.8820791Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 47%] 2023-03-31T06:51:18.8820998Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 47%] 2023-03-31T06:51:18.8821204Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 47%] 2023-03-31T06:51:18.8821422Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 47%] 2023-03-31T06:51:18.8821673Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 47%] 2023-03-31T06:51:18.8821892Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 47%] 2023-03-31T06:51:18.8822101Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 47%] 2023-03-31T06:51:18.8822302Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 47%] 2023-03-31T06:51:18.8822551Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 47%] 2023-03-31T06:51:18.8822781Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 47%] 2023-03-31T06:51:18.8822990Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_bfloat16 SKIPPED (Skipped!) [ 47%] 2023-03-31T06:51:18.8823173Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float16 SKIPPED (Skipped!) [ 47%] 2023-03-31T06:51:18.8823433Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float32 SKIPPED (Skipped!) [ 47%] 2023-03-31T06:51:18.8823614Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_cpu_float64 SKIPPED (Skipped!) [ 47%] 2023-03-31T06:51:18.8823796Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:18.8823978Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:18.8824140Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:18.8824301Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_bool PASSED [ 47%] 2023-03-31T06:51:18.8824468Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex128 PASSED [ 47%] 2023-03-31T06:51:18.8824636Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex32 PASSED [ 47%] 2023-03-31T06:51:18.8824801Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:18.8824964Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:18.8825132Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:18.8825294Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:18.8825447Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int16 PASSED [ 47%] 2023-03-31T06:51:18.8825610Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int32 PASSED [ 47%] 2023-03-31T06:51:18.8825774Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int8 PASSED [ 47%] 2023-03-31T06:51:18.8825941Z test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_complex128 PASSED [ 47%] 2023-03-31T06:51:18.8826106Z test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:18.8826269Z test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:18.8826435Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:18.8826597Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:18.8826757Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:18.8826907Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:18.8827067Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int16 PASSED [ 47%] 2023-03-31T06:51:18.8827224Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int8 PASSED [ 47%] 2023-03-31T06:51:18.8827381Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8827553Z test_decomp.py::TestDecompCPU::test_comprehensive_pca_lowrank_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:18.8827763Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex32 PASSED [ 48%] 2023-03-31T06:51:18.8827930Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex64 PASSED [ 48%] 2023-03-31T06:51:18.8828094Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float16 PASSED [ 48%] 2023-03-31T06:51:18.8828240Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:18.8828401Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:18.8828592Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int8 PASSED [ 48%] 2023-03-31T06:51:18.8828780Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8828948Z test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:18.8829138Z test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:18.8829305Z test_decomp.py::TestDecompCPU::test_comprehensive_polar_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:18.8829491Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_bool PASSED [ 48%] 2023-03-31T06:51:18.8829663Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:18.8829850Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:18.8830032Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_bool PASSED [ 48%] 2023-03-31T06:51:18.8830216Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int16 PASSED [ 48%] 2023-03-31T06:51:18.8830397Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:18.8830582Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8830764Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_bool PASSED [ 48%] 2023-03-31T06:51:18.8830949Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:18.8831127Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:18.8831294Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8831482Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:18.8831665Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:18.8831842Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int16 PASSED [ 48%] 2023-03-31T06:51:18.8832024Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int32 PASSED [ 48%] 2023-03-31T06:51:18.8832204Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:18.8832382Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8832562Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_bool PASSED [ 48%] 2023-03-31T06:51:18.8832749Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:18.8832916Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int16 PASSED [ 48%] 2023-03-31T06:51:18.8833095Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int32 PASSED [ 48%] 2023-03-31T06:51:18.8833273Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:18.8833449Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex128 PASSED [ 48%] 2023-03-31T06:51:18.8833660Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex64 PASSED [ 48%] 2023-03-31T06:51:18.8833828Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int16 PASSED [ 48%] 2023-03-31T06:51:18.8833988Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int32 PASSED [ 48%] 2023-03-31T06:51:18.8834146Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:18.8834295Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:18.8834499Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:18.8834685Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_complex64 PASSED [ 49%] 2023-03-31T06:51:18.8834846Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:18.8835034Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:18.8835195Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int64 PASSED [ 49%] 2023-03-31T06:51:18.8835355Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:18.8835511Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_uint8 PASSED [ 49%] 2023-03-31T06:51:18.8835660Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:18.8835821Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:18.8835978Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:18.8836135Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_uint8 PASSED [ 49%] 2023-03-31T06:51:18.8836298Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_complex64 PASSED [ 49%] 2023-03-31T06:51:18.8836455Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float16 PASSED [ 49%] 2023-03-31T06:51:18.8836613Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:18.8836769Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:18.8836911Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int64 PASSED [ 49%] 2023-03-31T06:51:18.8837067Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:18.8837230Z test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_complex128 PASSED [ 49%] 2023-03-31T06:51:18.8837392Z test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_complex64 PASSED [ 49%] 2023-03-31T06:51:18.8837550Z test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:18.8837705Z test_decomp.py::TestDecompCPU::test_comprehensive_qr_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:18.8837873Z test_decomp.py::TestDecompCPU::test_comprehensive_quantile_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:18.8838038Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:18.8838192Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float16 PASSED [ 49%] 2023-03-31T06:51:18.8838351Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:18.8838512Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:18.8838672Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int64 PASSED [ 49%] 2023-03-31T06:51:18.8838831Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:18.8839002Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex128 PASSED [ 49%] 2023-03-31T06:51:18.8839175Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex64 PASSED [ 49%] 2023-03-31T06:51:18.8839341Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:18.8839507Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float16 PASSED [ 49%] 2023-03-31T06:51:18.8839682Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:18.8839843Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:18.8840015Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:18.8840183Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:18.8840353Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:18.8840555Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_uint8 PASSED [ 49%] 2023-03-31T06:51:18.8840749Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:18.8840913Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex32 PASSED [ 50%] 2023-03-31T06:51:18.8841093Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_complex64 PASSED [ 50%] 2023-03-31T06:51:18.8841258Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8841418Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:18.8841578Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:18.8841747Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_bfloat16 PASSED [ 50%] 2023-03-31T06:51:18.8841917Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:18.8842090Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex32 PASSED [ 50%] 2023-03-31T06:51:18.8842261Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_complex64 PASSED [ 50%] 2023-03-31T06:51:18.8842415Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8842581Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:18.8842748Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_like_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:18.8842906Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8843287Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:18.8843449Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:18.8843617Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex32 PASSED [ 50%] 2023-03-31T06:51:18.8843781Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:18.8843943Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int16 PASSED [ 50%] 2023-03-31T06:51:18.8844088Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:18.8844248Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int8 PASSED [ 50%] 2023-03-31T06:51:18.8844424Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:18.8844598Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_complex64 PASSED [ 50%] 2023-03-31T06:51:18.8844768Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8844938Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:18.8845108Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:18.8845278Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int8 PASSED [ 50%] 2023-03-31T06:51:18.8845435Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8845599Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:18.8845765Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int16 PASSED [ 50%] 2023-03-31T06:51:18.8845987Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int8 PASSED [ 50%] 2023-03-31T06:51:18.8846151Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_uint8 PASSED [ 50%] 2023-03-31T06:51:18.8846319Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:18.8846518Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float16 SKIPPED (Inconsistent accuracy) [ 50%] 2023-03-31T06:51:18.8846680Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_bool PASSED [ 50%] 2023-03-31T06:51:18.8846877Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:18.8847080Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_complex64 PASSED [ 50%] 2023-03-31T06:51:18.8847245Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:18.8847446Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:18.8847613Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:18.8847773Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int8 PASSED [ 51%] 2023-03-31T06:51:18.8847933Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_uint8 PASSED [ 51%] 2023-03-31T06:51:18.8848120Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex32 PASSED [ 51%] 2023-03-31T06:51:18.8848300Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:18.8848461Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:18.8848632Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_bool PASSED [ 51%] 2023-03-31T06:51:18.8848805Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex128 PASSED [ 51%] 2023-03-31T06:51:18.8848980Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex32 PASSED [ 51%] 2023-03-31T06:51:18.8849151Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:18.8849320Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:18.8849486Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int64 PASSED [ 51%] 2023-03-31T06:51:18.8849647Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_uint8 PASSED [ 51%] 2023-03-31T06:51:18.8849796Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:18.8849969Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex32 PASSED [ 51%] 2023-03-31T06:51:18.8850134Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:18.8850299Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float16 PASSED [ 51%] 2023-03-31T06:51:18.8850463Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:18.8850627Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int16 PASSED [ 51%] 2023-03-31T06:51:18.8850791Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int32 PASSED [ 51%] 2023-03-31T06:51:18.8850959Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:18.8851112Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_complex128 PASSED [ 51%] 2023-03-31T06:51:18.8851278Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:18.8851444Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float16 PASSED [ 51%] 2023-03-31T06:51:18.8851608Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float32 PASSED [ 51%] 2023-03-31T06:51:18.8851768Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int16 PASSED [ 51%] 2023-03-31T06:51:18.8851930Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int32 PASSED [ 51%] 2023-03-31T06:51:18.8852137Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_complex128 PASSED [ 51%] 2023-03-31T06:51:18.8852309Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:18.8852462Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float16 PASSED [ 51%] 2023-03-31T06:51:18.8852629Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float32 PASSED [ 51%] 2023-03-31T06:51:18.8852822Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int32 PASSED [ 51%] 2023-03-31T06:51:18.8853016Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:18.8853183Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int16 PASSED [ 51%] 2023-03-31T06:51:18.8853345Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int64 PASSED [ 51%] 2023-03-31T06:51:18.8853550Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int8 PASSED [ 51%] 2023-03-31T06:51:18.8853719Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:18.8853891Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex32 PASSED [ 51%] 2023-03-31T06:51:18.8854050Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:18.8854221Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float16 PASSED [ 51%] 2023-03-31T06:51:18.8854388Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int32 PASSED [ 52%] 2023-03-31T06:51:18.8854550Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:18.8854711Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8854874Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:18.8855036Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float16 PASSED [ 52%] 2023-03-31T06:51:18.8855199Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:18.8855345Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int32 PASSED [ 52%] 2023-03-31T06:51:18.8855503Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:18.8855666Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8855833Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:18.8855998Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_complex64 PASSED [ 52%] 2023-03-31T06:51:18.8856159Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float16 PASSED [ 52%] 2023-03-31T06:51:18.8856320Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:18.8856486Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:18.8856632Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int16 PASSED [ 52%] 2023-03-31T06:51:18.8856791Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:18.8856955Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8857114Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:18.8857273Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int16 PASSED [ 52%] 2023-03-31T06:51:18.8857431Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int32 PASSED [ 52%] 2023-03-31T06:51:18.8857591Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:18.8857750Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:18.8857914Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:18.8858137Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8858320Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:18.8858483Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8858641Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_bool PASSED [ 52%] 2023-03-31T06:51:18.8858808Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:18.8858997Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:18.8859179Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:18.8859344Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:18.8859524Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_complex64 PASSED [ 52%] 2023-03-31T06:51:18.8859686Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:18.8859846Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:18.8860001Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:18.8860156Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int8 PASSED [ 52%] 2023-03-31T06:51:18.8860315Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:18.8860491Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:18.8860663Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_bool PASSED [ 52%] 2023-03-31T06:51:18.8860822Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8860995Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8861167Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8861340Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_complex128 PASSED [ 53%] 2023-03-31T06:51:18.8861511Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8861679Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8861847Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:18.8862018Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:18.8862169Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8862337Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_complex128 PASSED [ 53%] 2023-03-31T06:51:18.8862509Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_complex64 PASSED [ 53%] 2023-03-31T06:51:18.8862676Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8862838Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8862996Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8863174Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8863455Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8863648Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8863812Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_bool PASSED [ 53%] 2023-03-31T06:51:18.8863992Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8864209Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:18.8864384Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:18.8864562Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8864742Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8864922Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8865125Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:18.8865353Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:18.8865537Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8865748Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8865925Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:18.8866102Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:18.8866279Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:18.8866456Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8866637Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8866817Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:18.8866979Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:18.8867157Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int64 PASSED [ 53%] 2023-03-31T06:51:18.8867337Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:18.8867508Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:18.8867682Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:18.8867853Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:18.8868022Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:18.8868184Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_bool PASSED [ 54%] 2023-03-31T06:51:18.8868340Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex128 PASSED [ 54%] 2023-03-31T06:51:18.8868504Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:18.8868667Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:18.8868829Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:18.8869002Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:18.8869170Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_bool PASSED [ 54%] 2023-03-31T06:51:18.8869343Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8869506Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:18.8869666Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_bool PASSED [ 54%] 2023-03-31T06:51:18.8869812Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8869971Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:18.8870128Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:18.8870316Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:18.8870471Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:18.8870634Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:18.8870793Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_bool PASSED [ 54%] 2023-03-31T06:51:18.8870960Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_complex64 PASSED [ 54%] 2023-03-31T06:51:18.8871138Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float16 PASSED [ 54%] 2023-03-31T06:51:18.8871318Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:18.8871473Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:18.8871667Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:18.8871840Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_complex128 PASSED [ 54%] 2023-03-31T06:51:18.8872008Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_complex64 PASSED [ 54%] 2023-03-31T06:51:18.8872171Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8872335Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:18.8872484Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:18.8872650Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int32 PASSED [ 54%] 2023-03-31T06:51:18.8872815Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:18.8872973Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:18.8873132Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int32 PASSED [ 54%] 2023-03-31T06:51:18.8873290Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:18.8873445Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:18.8873631Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_blackman_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:18.8873801Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_cosine_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8873992Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_exponential_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8874183Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_exponential_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:18.8874372Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_gaussian_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:18.8874564Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_cosine_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:18.8874762Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_cosine_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:18.8874957Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_hamming_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:18.8875135Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hann_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:18.8875317Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_nuttall_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:18.8875475Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:18.8875642Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_bool PASSED [ 55%] 2023-03-31T06:51:18.8875803Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:18.8875965Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int8 PASSED [ 55%] 2023-03-31T06:51:18.8876127Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:18.8876326Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_bool PASSED [ 55%] 2023-03-31T06:51:18.8876496Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_complex128 PASSED [ 55%] 2023-03-31T06:51:18.8876658Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_complex64 PASSED [ 55%] 2023-03-31T06:51:18.8876803Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:18.8876959Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:18.8877155Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:18.8877334Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:18.8877488Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:18.8877678Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:18.8877835Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_bool PASSED [ 55%] 2023-03-31T06:51:18.8878002Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_complex128 PASSED [ 55%] 2023-03-31T06:51:18.8878166Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_complex64 PASSED [ 55%] 2023-03-31T06:51:18.8878313Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:18.8878467Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:18.8878626Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:18.8878781Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:18.8878936Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:18.8879100Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:18.8879256Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_bool PASSED [ 55%] 2023-03-31T06:51:18.8879422Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_complex64 PASSED [ 55%] 2023-03-31T06:51:18.8879569Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:18.8879722Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:18.8879877Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:18.8880034Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:18.8880191Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_int8 PASSED [ 55%] 2023-03-31T06:51:18.8880344Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:18.8880509Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:18.8880669Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_bool PASSED [ 55%] 2023-03-31T06:51:18.8880824Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex32 PASSED [ 55%] 2023-03-31T06:51:18.8880987Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex64 PASSED [ 56%] 2023-03-31T06:51:18.8881148Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float16 PASSED [ 56%] 2023-03-31T06:51:18.8881311Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8881467Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int16 PASSED [ 56%] 2023-03-31T06:51:18.8881641Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_bfloat16 PASSED [ 56%] 2023-03-31T06:51:18.8881815Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8881985Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int16 PASSED [ 56%] 2023-03-31T06:51:18.8882142Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:18.8882352Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8882533Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_bfloat16 PASSED [ 56%] 2023-03-31T06:51:18.8882710Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_bool PASSED [ 56%] 2023-03-31T06:51:18.8882893Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_complex64 PASSED [ 56%] 2023-03-31T06:51:18.8883379Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float16 PASSED [ 56%] 2023-03-31T06:51:18.8883631Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8883809Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:18.8884014Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_bfloat16 PASSED [ 56%] 2023-03-31T06:51:18.8884165Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8884324Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8884482Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int16 PASSED [ 56%] 2023-03-31T06:51:18.8884643Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_uint8 PASSED [ 56%] 2023-03-31T06:51:18.8884818Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8884996Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8885171Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8885342Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8885502Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int16 PASSED [ 56%] 2023-03-31T06:51:18.8885675Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int32 PASSED [ 56%] 2023-03-31T06:51:18.8885845Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_uint8 PASSED [ 56%] 2023-03-31T06:51:18.8886019Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_bool PASSED [ 56%] 2023-03-31T06:51:18.8886194Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8886366Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int16 PASSED [ 56%] 2023-03-31T06:51:18.8886538Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int32 PASSED [ 56%] 2023-03-31T06:51:18.8886709Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:18.8886871Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8887044Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:18.8887215Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_bool PASSED [ 56%] 2023-03-31T06:51:18.8887391Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:18.8887562Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:18.8887731Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int32 PASSED [ 56%] 2023-03-31T06:51:18.8887904Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:18.8888077Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:18.8888245Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:18.8888407Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8888614Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_float64 PASSED [ 57%] 2023-03-31T06:51:18.8888785Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int16 PASSED [ 57%] 2023-03-31T06:51:18.8888954Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:18.8889121Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:18.8889346Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8889567Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:18.8889763Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:18.8889980Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8890161Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:18.8890353Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:18.8890546Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:18.8890740Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:18.8891272Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8891664Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8892052Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8892435Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8892820Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8893196Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8893577Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8893943Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 57%] 2023-03-31T06:51:18.8894118Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_bfloat16 PASSED [ 57%] 2023-03-31T06:51:18.8894289Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_bool PASSED [ 57%] 2023-03-31T06:51:18.8894461Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8894629Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int16 PASSED [ 57%] 2023-03-31T06:51:18.8894801Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:18.8894971Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:18.8895140Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_bool PASSED [ 57%] 2023-03-31T06:51:18.8895300Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8895503Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:18.8895672Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:18.8895838Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:18.8896029Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_bool PASSED [ 57%] 2023-03-31T06:51:18.8896222Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:18.8896443Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_float64 PASSED [ 57%] 2023-03-31T06:51:18.8896657Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:18.8896847Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:18.8897067Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_bool PASSED [ 57%] 2023-03-31T06:51:18.8897261Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int32 PASSED [ 58%] 2023-03-31T06:51:18.8897431Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_bfloat16 PASSED [ 58%] 2023-03-31T06:51:18.8897599Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_float32 PASSED [ 58%] 2023-03-31T06:51:18.8897767Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:18.8897938Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int32 PASSED [ 58%] 2023-03-31T06:51:18.8898104Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:18.8898266Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:18.8898435Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_bool PASSED [ 58%] 2023-03-31T06:51:18.8898592Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:18.8898752Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:18.8898917Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_bool PASSED [ 58%] 2023-03-31T06:51:18.8899084Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:18.8899252Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:18.8899416Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:18.8899611Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:18.8899801Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int32 PASSED [ 58%] 2023-03-31T06:51:18.8899979Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:18.8900174Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:18.8900563Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8900946Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8901327Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8901695Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8902103Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8902474Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:18.8902648Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_bool PASSED [ 58%] 2023-03-31T06:51:18.8902824Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:18.8903024Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:18.8903214Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:18.8903485Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_bool PASSED [ 58%] 2023-03-31T06:51:18.8903717Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:18.8903909Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:18.8904095Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:18.8904279Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_bool PASSED [ 58%] 2023-03-31T06:51:18.8904461Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:18.8904641Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int32 PASSED [ 58%] 2023-03-31T06:51:18.8904806Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:18.8904993Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_float32 PASSED [ 58%] 2023-03-31T06:51:18.8905184Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:18.8905370Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:18.8905552Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:18.8905785Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_bfloat16 SKIPPED (special.ndtr in torch.bfloat16 not supported) [ 59%] 2023-03-31T06:51:18.8906012Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_bool SKIPPED (special.ndtr in torch.bool not supported) [ 59%] 2023-03-31T06:51:18.8906242Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_float32 SKIPPED (special.ndtr in torch.float32 not supported) [ 59%] 2023-03-31T06:51:18.8906464Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int16 SKIPPED (special.ndtr in torch.int16 not supported) [ 59%] 2023-03-31T06:51:18.8906685Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int32 SKIPPED (special.ndtr in torch.int32 not supported) [ 59%] 2023-03-31T06:51:18.8906889Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int64 SKIPPED (special.ndtr in torch.int64 not supported) [ 59%] 2023-03-31T06:51:18.8907102Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_uint8 SKIPPED (special.ndtr in torch.uint8 not supported) [ 59%] 2023-03-31T06:51:18.8907273Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_bool PASSED [ 59%] 2023-03-31T06:51:18.8907448Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:18.8907622Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:18.8907790Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:18.8907997Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_bool PASSED [ 59%] 2023-03-31T06:51:18.8908231Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:18.8908433Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:18.8908618Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:18.8908851Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:18.8909079Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:18.8909279Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:18.8909505Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_uint8 PASSED [ 59%] 2023-03-31T06:51:18.8909703Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_bool PASSED [ 59%] 2023-03-31T06:51:18.8909900Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:18.8910096Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:18.8910286Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:18.8910468Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:18.8910662Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:18.8911063Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8911469Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8911862Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8912253Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8912650Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8913037Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8913437Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8913827Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8914220Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8914614Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8915004Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8915416Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:18.8915611Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:18.8915801Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:18.8916051Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8916266Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int8 PASSED [ 60%] 2023-03-31T06:51:18.8916445Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_bfloat16 PASSED [ 60%] 2023-03-31T06:51:18.8916649Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_bool PASSED [ 60%] 2023-03-31T06:51:18.8916829Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:18.8917005Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float64 PASSED [ 60%] 2023-03-31T06:51:18.8917163Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:18.8917337Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:18.8917505Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8917679Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_bool PASSED [ 60%] 2023-03-31T06:51:18.8917852Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_float64 PASSED [ 60%] 2023-03-31T06:51:18.8918021Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:18.8918193Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8918363Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:18.8918515Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_bfloat16 PASSED [ 60%] 2023-03-31T06:51:18.8918682Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex32 PASSED [ 60%] 2023-03-31T06:51:18.8918844Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float16 PASSED [ 60%] 2023-03-31T06:51:18.8919003Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:18.8919167Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8919327Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int8 PASSED [ 60%] 2023-03-31T06:51:18.8919488Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:18.8919671Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_complex64 PASSED [ 60%] 2023-03-31T06:51:18.8919835Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:18.8920004Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:18.8920171Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int8 PASSED [ 60%] 2023-03-31T06:51:18.8920343Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_bool PASSED [ 60%] 2023-03-31T06:51:18.8920523Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex32 PASSED [ 60%] 2023-03-31T06:51:18.8920699Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:18.8920873Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:18.8921043Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8921212Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:18.8921402Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_complex128 PASSED [ 60%] 2023-03-31T06:51:18.8921569Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_complex64 PASSED [ 60%] 2023-03-31T06:51:18.8921729Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:18.8921891Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_float64 PASSED [ 60%] 2023-03-31T06:51:18.8922086Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:18.8922272Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:18.8922429Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int64 PASSED [ 60%] 2023-03-31T06:51:18.8922586Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_int8 PASSED [ 60%] 2023-03-31T06:51:18.8922756Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:18.8922917Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_bool PASSED [ 61%] 2023-03-31T06:51:18.8923298Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:18.8923468Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:18.8923633Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:18.8923793Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:18.8923957Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int64 PASSED [ 61%] 2023-03-31T06:51:18.8924121Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:18.8924275Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:18.8924440Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_bool PASSED [ 61%] 2023-03-31T06:51:18.8924612Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:18.8924780Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex32 PASSED [ 61%] 2023-03-31T06:51:18.8924945Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:18.8925111Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:18.8925276Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:18.8925442Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:18.8925591Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int64 PASSED [ 61%] 2023-03-31T06:51:18.8925753Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:18.8925918Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:18.8926096Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_bool PASSED [ 61%] 2023-03-31T06:51:18.8926278Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:18.8926456Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:18.8926632Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:18.8926805Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:18.8926980Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:18.8927142Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:18.8927311Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int64 PASSED [ 61%] 2023-03-31T06:51:18.8927483Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:18.8927713Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_bool PASSED [ 61%] 2023-03-31T06:51:18.8927883Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:18.8928050Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex32 PASSED [ 61%] 2023-03-31T06:51:18.8928215Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:18.8928410Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:18.8928592Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:18.8928749Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:18.8928909Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:18.8929117Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:18.8929281Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:18.8929439Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:18.8929609Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:18.8929777Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:18.8929930Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:18.8930093Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8930277Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:18.8930457Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:18.8930635Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:18.8930807Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:18.8930978Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:18.8931139Z test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8931300Z test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:18.8931447Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8931609Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int64 PASSED [ 62%] 2023-03-31T06:51:18.8931766Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_uint8 PASSED [ 62%] 2023-03-31T06:51:18.8931925Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:18.8932086Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_bool PASSED [ 62%] 2023-03-31T06:51:18.8932253Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:18.8932416Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:18.8932573Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8932716Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int16 PASSED [ 62%] 2023-03-31T06:51:18.8932869Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int64 PASSED [ 62%] 2023-03-31T06:51:18.8933044Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:18.8933218Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:18.8933386Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:18.8933558Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8933757Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:18.8933923Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int8 PASSED [ 62%] 2023-03-31T06:51:18.8934075Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:18.8934242Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_lowrank_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8934435Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_lowrank_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:18.8934619Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:18.8934776Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_bool PASSED [ 62%] 2023-03-31T06:51:18.8934934Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:18.8935119Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:18.8935274Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:18.8935436Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:18.8935604Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_bool PASSED [ 62%] 2023-03-31T06:51:18.8935775Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:18.8935945Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int16 PASSED [ 62%] 2023-03-31T06:51:18.8936112Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:18.8936272Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:18.8936433Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_bool PASSED [ 63%] 2023-03-31T06:51:18.8936600Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:18.8936762Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float16 PASSED [ 63%] 2023-03-31T06:51:18.8936911Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:18.8937070Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:18.8937226Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:18.8937383Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_bool PASSED [ 63%] 2023-03-31T06:51:18.8937553Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:18.8937715Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:18.8937870Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:18.8938033Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:18.8938179Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:18.8938353Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:18.8938527Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:18.8938700Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:18.8938867Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:18.8939038Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:18.8939207Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:18.8939375Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:18.8939533Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:18.8939736Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:18.8939904Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:18.8940072Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:18.8940237Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:18.8940400Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:18.8940586Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:18.8940773Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:18.8940918Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:18.8941110Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:18.8941270Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:18.8941427Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:18.8941580Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:18.8941733Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:18.8941888Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:18.8942044Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:18.8942200Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:18.8942355Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:18.8942521Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_bool PASSED [ 64%] 2023-03-31T06:51:18.8942693Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:18.8942862Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_complex64 PASSED [ 64%] 2023-03-31T06:51:18.8943024Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:18.8943184Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:18.8943438Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:18.8943609Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8943756Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:18.8943916Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:18.8944088Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:18.8944251Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8944412Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:18.8944570Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:18.8944725Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:18.8944895Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:18.8945056Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:18.8945229Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8945395Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:18.8945558Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:18.8945727Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:18.8945933Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float16 PASSED [ 64%] 2023-03-31T06:51:18.8946098Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:18.8946257Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8946407Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:18.8946620Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:18.8946803Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:18.8946965Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:18.8947122Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:18.8947298Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:18.8947468Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:18.8947626Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_complex64 PASSED [ 64%] 2023-03-31T06:51:18.8947786Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8947944Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:18.8948103Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:18.8948262Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:18.8948444Z test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:18.8948612Z test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_complex64 PASSED [ 64%] 2023-03-31T06:51:18.8948792Z test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:18.8948969Z test_decomp.py::TestDecompCPU::test_comprehensive_triangular_solve_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:18.8949130Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:18.8949296Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:18.8949459Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:18.8949623Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:18.8949787Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:18.8949931Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:18.8950089Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8950247Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int8 PASSED [ 65%] 2023-03-31T06:51:18.8950417Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_indices_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8950573Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_bool PASSED [ 65%] 2023-03-31T06:51:18.8950732Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:18.8950894Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:18.8951050Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int16 PASSED [ 65%] 2023-03-31T06:51:18.8951193Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:18.8951350Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8951504Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_int8 PASSED [ 65%] 2023-03-31T06:51:18.8951680Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:18.8951878Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_bool PASSED [ 65%] 2023-03-31T06:51:18.8952053Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:18.8952226Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:18.8952396Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:18.8952592Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:18.8952773Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int16 PASSED [ 65%] 2023-03-31T06:51:18.8952936Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:18.8953095Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8953288Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_int8 PASSED [ 65%] 2023-03-31T06:51:18.8953451Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_uint8 PASSED [ 65%] 2023-03-31T06:51:18.8953613Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:18.8953773Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:18.8953930Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8954087Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex32 PASSED [ 65%] 2023-03-31T06:51:18.8954254Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:18.8954415Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:18.8954575Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:18.8954734Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int8 PASSED [ 65%] 2023-03-31T06:51:18.8954906Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:18.8955071Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_bool PASSED [ 65%] 2023-03-31T06:51:18.8955243Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:18.8955402Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex32 PASSED [ 65%] 2023-03-31T06:51:18.8955570Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:18.8955736Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:18.8955896Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:18.8956060Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:18.8956222Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:18.8956393Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:18.8956568Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex128 PASSED [ 66%] 2023-03-31T06:51:18.8956737Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:18.8956888Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:18.8957050Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:18.8957216Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex64 PASSED [ 66%] 2023-03-31T06:51:18.8957379Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:18.8957538Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:18.8957699Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:18.8957881Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:18.8958050Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:18.8958203Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_complex64 PASSED [ 66%] 2023-03-31T06:51:18.8958367Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:18.8958557Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:18.8958761Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_bool PASSED [ 66%] 2023-03-31T06:51:18.8958940Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:18.8959115Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:18.8959316Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:18.8959479Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_bool PASSED [ 66%] 2023-03-31T06:51:18.8959631Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:18.8959791Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:18.8959950Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:18.8960112Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:18.8960287Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:18.8960454Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_bool PASSED [ 66%] 2023-03-31T06:51:18.8960629Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex128 PASSED [ 66%] 2023-03-31T06:51:18.8960804Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex32 PASSED [ 66%] 2023-03-31T06:51:18.8960966Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_complex64 PASSED [ 66%] 2023-03-31T06:51:18.8961135Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:18.8961304Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:18.8961474Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:18.8961644Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:18.8961814Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:18.8961981Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:18.8962152Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:18.8962325Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:18.8962479Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:18.8962643Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int16 PASSED [ 67%] 2023-03-31T06:51:18.8962805Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int8 PASSED [ 67%] 2023-03-31T06:51:18.8962964Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:18.8963322Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:18.8963485Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:18.8963653Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:18.8963834Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:18.8964087Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:18.8964264Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:18.8964436Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:18.8964604Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:18.8964765Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:18.8964972Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:18.8965170Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:18.8965329Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:18.8965511Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int32 PASSED [ 67%] 2023-03-31T06:51:18.8965670Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int8 PASSED [ 67%] 2023-03-31T06:51:18.8965828Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_uint8 PASSED [ 67%] 2023-03-31T06:51:18.8966001Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:18.8966175Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:18.8966343Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_complex_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:18.8966511Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:18.8966673Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_bool PASSED [ 67%] 2023-03-31T06:51:18.8966839Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex32 PASSED [ 67%] 2023-03-31T06:51:18.8966993Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:18.8967155Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:18.8967315Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int16 PASSED [ 67%] 2023-03-31T06:51:18.8967473Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int32 PASSED [ 67%] 2023-03-31T06:51:18.8967633Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int8 PASSED [ 67%] 2023-03-31T06:51:18.8967799Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:18.8967964Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:18.8968129Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int16 PASSED [ 67%] 2023-03-31T06:51:18.8968275Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int32 PASSED [ 67%] 2023-03-31T06:51:18.8968435Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int64 PASSED [ 67%] 2023-03-31T06:51:18.8968594Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_bool PASSED [ 67%] 2023-03-31T06:51:18.8968757Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:18.8968923Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex32 PASSED [ 67%] 2023-03-31T06:51:18.8969082Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:18.8969243Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:18.8969406Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:18.8969552Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int64 PASSED [ 68%] 2023-03-31T06:51:18.8969706Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int8 PASSED [ 68%] 2023-03-31T06:51:18.8969867Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_bool PASSED [ 68%] 2023-03-31T06:51:18.8970035Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex128 PASSED [ 68%] 2023-03-31T06:51:18.8970240Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex32 PASSED [ 68%] 2023-03-31T06:51:18.8970406Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float16 PASSED [ 68%] 2023-03-31T06:51:18.8970565Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float64 PASSED [ 68%] 2023-03-31T06:51:18.8970726Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:18.8970904Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int64 PASSED [ 68%] 2023-03-31T06:51:18.8971092Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int8 PASSED [ 68%] 2023-03-31T06:51:18.8971254Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_uint8 PASSED [ 68%] 2023-03-31T06:51:18.8971422Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex64 PASSED [ 68%] 2023-03-31T06:51:18.8971610Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:18.8971769Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int64 PASSED [ 68%] 2023-03-31T06:51:18.8971927Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int8 PASSED [ 68%] 2023-03-31T06:51:18.8972085Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_uint8 PASSED [ 68%] 2023-03-31T06:51:18.8972236Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:51:18.8972397Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_bool PASSED [ 68%] 2023-03-31T06:51:18.8972567Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex128 PASSED [ 68%] 2023-03-31T06:51:18.8972732Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex32 PASSED [ 68%] 2023-03-31T06:51:18.8972893Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_complex64 PASSED [ 68%] 2023-03-31T06:51:18.8973058Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:18.8973217Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:18.8973376Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int64 PASSED [ 68%] 2023-03-31T06:51:18.8973532Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int8 PASSED [ 68%] 2023-03-31T06:51:18.8973678Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_bool PASSED [ 68%] 2023-03-31T06:51:18.8973838Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float16 PASSED [ 68%] 2023-03-31T06:51:18.8973997Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:18.8974154Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int64 PASSED [ 68%] 2023-03-31T06:51:18.8974373Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_bfloat16 SKIPPED (zero_ in torch.bfloat16 not supported) [ 68%] 2023-03-31T06:51:18.8974582Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_bool SKIPPED (zero_ in torch.bool not supported) [ 68%] 2023-03-31T06:51:18.8974804Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_complex128 SKIPPED (zero_ in torch.complex128 not supported) [ 68%] 2023-03-31T06:51:18.8975015Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float16 SKIPPED (zero_ in torch.float16 not supported) [ 68%] 2023-03-31T06:51:18.8975224Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float32 SKIPPED (zero_ in torch.float32 not supported) [ 68%] 2023-03-31T06:51:18.8975419Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_float64 SKIPPED (zero_ in torch.float64 not supported) [ 68%] 2023-03-31T06:51:18.8975623Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_uint8 SKIPPED (zero_ in torch.uint8 not supported) [ 68%] 2023-03-31T06:51:18.8975787Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float16 PASSED [ 68%] 2023-03-31T06:51:18.8975952Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:18.8976145Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:18.8976306Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:18.8976479Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:18.8976644Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_bool PASSED [ 69%] 2023-03-31T06:51:18.8976796Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int16 PASSED [ 69%] 2023-03-31T06:51:18.8976985Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int32 PASSED [ 69%] 2023-03-31T06:51:18.8977172Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:18.8977343Z test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:18.8977542Z test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:18.8977715Z test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:18.8977960Z test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_float32 SKIPPED (_upsample_bilinear2d_aa in torch.float32 not supported) [ 69%] 2023-03-31T06:51:18.8978196Z test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_float64 SKIPPED (_upsample_bilinear2d_aa in torch.float64 not supported) [ 69%] 2023-03-31T06:51:18.8978432Z test_decomp.py::TestDecompCPU::test_quick__upsample_bilinear2d_aa_cpu_uint8 SKIPPED (_upsample_bilinear2d_aa in torch.uint8 not supported) [ 69%] 2023-03-31T06:51:18.8978575Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:18.8978729Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:18.8978882Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float16 PASSED [ 69%] 2023-03-31T06:51:18.8979035Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:18.8979183Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:18.8979329Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_bool PASSED [ 69%] 2023-03-31T06:51:18.8979482Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:18.8979631Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:18.8979769Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int32 PASSED [ 69%] 2023-03-31T06:51:18.8979917Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:18.8980066Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:18.8980213Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_bool PASSED [ 69%] 2023-03-31T06:51:18.8980366Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:18.8980518Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:18.8980666Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:18.8980811Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int16 PASSED [ 69%] 2023-03-31T06:51:18.8980944Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int32 PASSED [ 69%] 2023-03-31T06:51:18.8981094Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:18.8981245Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:18.8981394Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex32 PASSED [ 69%] 2023-03-31T06:51:18.8981542Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:18.8981689Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_int16 PASSED [ 69%] 2023-03-31T06:51:18.8981872Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:18.8982024Z test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:18.8982173Z test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:18.8982309Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:18.8982463Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:18.8982642Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:18.8982820Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:18.8982966Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:18.8983150Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:18.8983301Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:18.8983541Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:18.8983682Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:18.8983829Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:18.8983975Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:18.8984123Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:18.8984270Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:18.8984438Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:18.8984608Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:18.8984777Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:18.8984925Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:18.8985084Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:18.8985244Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:18.8985397Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:18.8985552Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:18.8985704Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:18.8985852Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:18.8985997Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:18.8986133Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:18.8986282Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:18.8986428Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_bool PASSED [ 70%] 2023-03-31T06:51:18.8986573Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float16 PASSED [ 70%] 2023-03-31T06:51:18.8986718Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:18.8986865Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:18.8987013Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:18.8987161Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:18.8987296Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:18.8987444Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:18.8987627Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:18.8987771Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:18.8987916Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:18.8988064Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:18.8988210Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float16 PASSED [ 70%] 2023-03-31T06:51:18.8988381Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8988538Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_bool PASSED [ 71%] 2023-03-31T06:51:18.8988685Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float16 PASSED [ 71%] 2023-03-31T06:51:18.8988831Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:18.8989012Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8989158Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8989307Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:18.8989455Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8989602Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8989737Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_complex64 PASSED [ 71%] 2023-03-31T06:51:18.8989886Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_float16 PASSED [ 71%] 2023-03-31T06:51:18.8990032Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:18.8990175Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:18.8990321Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:18.8990466Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:18.8990609Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8990753Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8990891Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:18.8991038Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float16 PASSED [ 71%] 2023-03-31T06:51:18.8991186Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:18.8991337Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:18.8991483Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:18.8991633Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:18.8991780Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8991929Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8992096Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:18.8992250Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float16 PASSED [ 71%] 2023-03-31T06:51:18.8992414Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:18.8992581Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:18.8992746Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:18.8992904Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:18.8993063Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8993215Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:18.8993399Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_complex64 PASSED [ 71%] 2023-03-31T06:51:18.8993535Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:18.8993683Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:18.8993830Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:18.8994000Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:18.8994172Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:18.8994324Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:18.8994471Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_bool PASSED [ 72%] 2023-03-31T06:51:18.8994651Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_complex128 PASSED [ 72%] 2023-03-31T06:51:18.8994790Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_complex64 PASSED [ 72%] 2023-03-31T06:51:18.8994936Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:18.8995080Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:18.8995224Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.8995368Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8995520Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_float64 PASSED [ 72%] 2023-03-31T06:51:18.8995668Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.8995812Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8995949Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_bfloat16 PASSED [ 72%] 2023-03-31T06:51:18.8996096Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_bool PASSED [ 72%] 2023-03-31T06:51:18.8996244Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_complex64 PASSED [ 72%] 2023-03-31T06:51:18.8996390Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_float64 PASSED [ 72%] 2023-03-31T06:51:18.8996535Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:18.8996681Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.8996835Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_complex128 PASSED [ 72%] 2023-03-31T06:51:18.8996987Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_complex64 PASSED [ 72%] 2023-03-31T06:51:18.8997120Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_float64 PASSED [ 72%] 2023-03-31T06:51:18.8997265Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:18.8997412Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8997563Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:18.8997711Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8997875Z test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:18.8998035Z test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:18.8998189Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:18.8998328Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8998477Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:18.8998626Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:18.8998776Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.8998955Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8999124Z test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8999273Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:18.8999421Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:18.8999559Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.8999735Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:18.8999913Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float32 PASSED [ 72%] 2023-03-31T06:51:18.9000061Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float64 PASSED [ 72%] 2023-03-31T06:51:18.9000239Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:18.9000391Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:18.9000540Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:18.9000687Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex32 PASSED [ 73%] 2023-03-31T06:51:18.9000820Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:18.9000964Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:18.9001110Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:18.9001257Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:18.9001400Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:18.9001547Z test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_bfloat16 XFAIL [ 73%] 2023-03-31T06:51:18.9001695Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:18.9001842Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int16 PASSED [ 73%] 2023-03-31T06:51:18.9001986Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:18.9002122Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:18.9002269Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:18.9002412Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int32 PASSED [ 73%] 2023-03-31T06:51:18.9002560Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:18.9002708Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:18.9002859Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:18.9003191Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_bool PASSED [ 73%] 2023-03-31T06:51:18.9003346Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:18.9003481Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int16 PASSED [ 73%] 2023-03-31T06:51:18.9003629Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int32 PASSED [ 73%] 2023-03-31T06:51:18.9003775Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:18.9003923Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_bool PASSED [ 73%] 2023-03-31T06:51:18.9004074Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:18.9004225Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:18.9004371Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int16 PASSED [ 73%] 2023-03-31T06:51:18.9004520Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:18.9004654Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:18.9004864Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex32 PASSED [ 73%] 2023-03-31T06:51:18.9005013Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex64 PASSED [ 73%] 2023-03-31T06:51:18.9005159Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int16 PASSED [ 73%] 2023-03-31T06:51:18.9005304Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int32 PASSED [ 73%] 2023-03-31T06:51:18.9005502Z test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:18.9005685Z test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:18.9006124Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 73%] 2023-03-31T06:51:18.9006547Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 73%] 2023-03-31T06:51:18.9006710Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:18.9007046Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 73%] 2023-03-31T06:51:18.9007381Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 74%] 2023-03-31T06:51:18.9007719Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 74%] 2023-03-31T06:51:18.9007890Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:18.9008052Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_bool PASSED [ 74%] 2023-03-31T06:51:18.9008224Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:18.9008384Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:18.9008529Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:18.9008684Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:18.9008837Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_bool PASSED [ 74%] 2023-03-31T06:51:18.9008991Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float16 PASSED [ 74%] 2023-03-31T06:51:18.9009148Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float64 PASSED [ 74%] 2023-03-31T06:51:18.9009302Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:18.9009454Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:18.9009604Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int64 PASSED [ 74%] 2023-03-31T06:51:18.9009742Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:18.9009893Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:18.9010041Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_float32 PASSED [ 74%] 2023-03-31T06:51:18.9010194Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_float32 PASSED [ 74%] 2023-03-31T06:51:18.9010340Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_float64 PASSED [ 74%] 2023-03-31T06:51:18.9010484Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:18.9010632Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int8 PASSED [ 74%] 2023-03-31T06:51:18.9010779Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_uint8 PASSED [ 74%] 2023-03-31T06:51:18.9010928Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_float64 PASSED [ 74%] 2023-03-31T06:51:18.9011065Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:18.9011238Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:18.9011444Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_bfloat16 SKIPPED (diag in torch.bfloat16 not supported) [ 74%] 2023-03-31T06:51:18.9011645Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float16 SKIPPED (diag in torch.float16 not supported) [ 74%] 2023-03-31T06:51:18.9011840Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float64 SKIPPED (diag in torch.float64 not supported) [ 74%] 2023-03-31T06:51:18.9012062Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int64 SKIPPED (diag in torch.int64 not supported) [ 74%] 2023-03-31T06:51:18.9012278Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int8 SKIPPED (diag in torch.int8 not supported) [ 74%] 2023-03-31T06:51:18.9012470Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_uint8 SKIPPED (diag in torch.uint8 not supported) [ 74%] 2023-03-31T06:51:18.9012638Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_bool PASSED [ 74%] 2023-03-31T06:51:18.9012800Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:18.9012956Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float32 PASSED [ 74%] 2023-03-31T06:51:18.9013109Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float64 PASSED [ 74%] 2023-03-31T06:51:18.9013260Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:18.9013409Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int64 PASSED [ 74%] 2023-03-31T06:51:18.9013575Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:18.9013739Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float16 PASSED [ 74%] 2023-03-31T06:51:18.9013897Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:18.9014038Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:18.9014192Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9014343Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_bool PASSED [ 75%] 2023-03-31T06:51:18.9014497Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:18.9014646Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float16 PASSED [ 75%] 2023-03-31T06:51:18.9014794Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:18.9014948Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:18.9015099Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:18.9015235Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:18.9015404Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:18.9015563Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_bool PASSED [ 75%] 2023-03-31T06:51:18.9015734Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_complex128 PASSED [ 75%] 2023-03-31T06:51:18.9015896Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:18.9016058Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:18.9016213Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9016373Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:18.9016522Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:18.9016672Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:18.9016819Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:18.9016968Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:18.9017151Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:18.9017319Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:18.9017487Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:18.9017650Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9017827Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:18.9018022Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:18.9018191Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:18.9018357Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float16 PASSED [ 75%] 2023-03-31T06:51:18.9018549Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:18.9018714Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9018881Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:18.9019046Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float16 PASSED [ 75%] 2023-03-31T06:51:18.9019195Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:18.9019356Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9019519Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:18.9019676Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:18.9019830Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:18.9019977Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:18.9020127Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:18.9020278Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:18.9020494Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_bfloat16 SKIPPED (empty_like in torch.bfloat16 not supported) [ 76%] 2023-03-31T06:51:18.9020702Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex32 SKIPPED (empty_like in torch.complex32 not supported) [ 76%] 2023-03-31T06:51:18.9020921Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex64 SKIPPED (empty_like in torch.complex64 not supported) [ 76%] 2023-03-31T06:51:18.9021127Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int16 SKIPPED (empty_like in torch.int16 not supported) [ 76%] 2023-03-31T06:51:18.9021329Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int32 SKIPPED (empty_like in torch.int32 not supported) [ 76%] 2023-03-31T06:51:18.9021531Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int64 SKIPPED (empty_like in torch.int64 not supported) [ 76%] 2023-03-31T06:51:18.9021732Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_int8 SKIPPED (empty_like in torch.int8 not supported) [ 76%] 2023-03-31T06:51:18.9021884Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:18.9022037Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex32 PASSED [ 76%] 2023-03-31T06:51:18.9022192Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex64 PASSED [ 76%] 2023-03-31T06:51:18.9022331Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float16 PASSED [ 76%] 2023-03-31T06:51:18.9022477Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:18.9022625Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:18.9022774Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:18.9022952Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_bool PASSED [ 76%] 2023-03-31T06:51:18.9023100Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:18.9023246Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:18.9023514Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:18.9023689Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:18.9023861Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:18.9024006Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int64 PASSED [ 76%] 2023-03-31T06:51:18.9024153Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_bool PASSED [ 76%] 2023-03-31T06:51:18.9024338Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:18.9024488Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:18.9024634Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:18.9024781Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:18.9024915Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:18.9025065Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_complex128 PASSED [ 76%] 2023-03-31T06:51:18.9025213Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float16 PASSED [ 76%] 2023-03-31T06:51:18.9025363Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:18.9025507Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:18.9025655Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:18.9025804Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:18.9025952Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:18.9026084Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:18.9026229Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:18.9026375Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:18.9026522Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:18.9026669Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_uint8 PASSED [ 77%] 2023-03-31T06:51:18.9026820Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9026971Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:18.9027121Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int16 PASSED [ 77%] 2023-03-31T06:51:18.9027257Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int32 PASSED [ 77%] 2023-03-31T06:51:18.9027409Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9027554Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int64 PASSED [ 77%] 2023-03-31T06:51:18.9027699Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int8 PASSED [ 77%] 2023-03-31T06:51:18.9027853Z test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_bfloat16 XFAIL [ 77%] 2023-03-31T06:51:18.9028008Z test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float16 XFAIL [ 77%] 2023-03-31T06:51:18.9028161Z test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float32 XFAIL [ 77%] 2023-03-31T06:51:18.9028312Z test_decomp.py::TestDecompCPU::test_quick_exponential_cpu_float64 XFAIL [ 77%] 2023-03-31T06:51:18.9028445Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_bool PASSED [ 77%] 2023-03-31T06:51:18.9028627Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9028777Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:18.9028924Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:18.9029069Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:18.9029213Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int32 PASSED [ 77%] 2023-03-31T06:51:18.9029382Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int8 PASSED [ 77%] 2023-03-31T06:51:18.9029769Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9029922Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9030087Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:18.9030422Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9030745Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9031068Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9031394Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9031718Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9040005Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9040255Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9040420Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:18.9040799Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9041134Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9041452Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9041787Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9042113Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9042282Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:18.9042444Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:18.9042599Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:18.9042928Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:18.9043655Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9043987Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9044302Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9044788Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9044950Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:18.9045108Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:18.9045418Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9045794Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9046164Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9046323Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:18.9046697Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9047021Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9047347Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9047674Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9047840Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_complex128 PASSED [ 78%] 2023-03-31T06:51:18.9048001Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:18.9048142Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:18.9048468Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9048793Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9049112Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9049439Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9049600Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_complex128 PASSED [ 78%] 2023-03-31T06:51:18.9049927Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9050251Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9050580Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9050739Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:18.9050882Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:18.9051035Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:18.9051363Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9051689Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9052009Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9052367Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9052528Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:18.9052854Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9053176Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9053353Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:18.9053533Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:18.9053859Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9054209Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9054532Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:18.9054857Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9055012Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:18.9055337Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9055660Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9055990Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9056303Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9056458Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:18.9056610Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:18.9056933Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9057256Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9057585Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9057910Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9058074Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_complex128 PASSED [ 79%] 2023-03-31T06:51:18.9058396Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9058714Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9059027Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9059351Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9059672Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9059865Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_complex128 PASSED [ 79%] 2023-03-31T06:51:18.9060186Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9060340Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:18.9060494Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:18.9060858Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9061203Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9061513Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9061868Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9062194Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9062521Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9062677Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:18.9063001Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9063405Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9063764Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:18.9063921Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_bfloat16 PASSED [ 79%] 2023-03-31T06:51:18.9064077Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex128 PASSED [ 79%] 2023-03-31T06:51:18.9064214Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:18.9064363Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int16 PASSED [ 79%] 2023-03-31T06:51:18.9064512Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int32 PASSED [ 79%] 2023-03-31T06:51:18.9064661Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int64 PASSED [ 79%] 2023-03-31T06:51:18.9064815Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_complex128 PASSED [ 79%] 2023-03-31T06:51:18.9064967Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_complex64 PASSED [ 80%] 2023-03-31T06:51:18.9065118Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:18.9065268Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:18.9065400Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:18.9065548Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:18.9065693Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9065844Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:18.9065994Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:18.9066144Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:18.9066289Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:18.9066435Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9066582Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:18.9066779Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:18.9066933Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9067078Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_bool PASSED [ 80%] 2023-03-31T06:51:18.9067226Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:18.9067373Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:18.9067543Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:18.9067749Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:18.9067881Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:18.9068052Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:18.9068200Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9068348Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:18.9068493Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_bool PASSED [ 80%] 2023-03-31T06:51:18.9068640Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:18.9068785Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:18.9068931Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:18.9069065Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:18.9069210Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9069359Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:18.9069507Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:18.9069656Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:18.9069802Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:18.9069947Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:18.9070093Z test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:18.9070227Z test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:18.9070380Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:18.9070531Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex64 PASSED [ 80%] 2023-03-31T06:51:18.9070677Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:18.9070826Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9070973Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:18.9071117Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_uint8 PASSED [ 81%] 2023-03-31T06:51:18.9071261Z test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:18.9071406Z test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int32 PASSED [ 81%] 2023-03-31T06:51:18.9071538Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:18.9071683Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_bool PASSED [ 81%] 2023-03-31T06:51:18.9071832Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:18.9071977Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9072122Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:18.9072290Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:18.9072435Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int64 PASSED [ 81%] 2023-03-31T06:51:18.9072577Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:18.9072714Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float64 XFAIL [ 81%] 2023-03-31T06:51:18.9072859Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int64 XFAIL [ 81%] 2023-03-31T06:51:18.9073050Z test_decomp.py::TestDecompCPU::test_quick_grid_sampler_2d_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9073220Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_bool PASSED [ 81%] 2023-03-31T06:51:18.9073365Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9073547Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int64 PASSED [ 81%] 2023-03-31T06:51:18.9073691Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_uint8 PASSED [ 81%] 2023-03-31T06:51:18.9073846Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:18.9073982Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int32 PASSED [ 81%] 2023-03-31T06:51:18.9074132Z test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:18.9074279Z test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9074428Z test_decomp.py::TestDecompCPU::test_quick_hypot_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:18.9074579Z test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:18.9074729Z test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:18.9074880Z test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:18.9075034Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:18.9075177Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex64 PASSED [ 81%] 2023-03-31T06:51:18.9075326Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9075476Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:18.9075623Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:18.9075769Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int64 PASSED [ 81%] 2023-03-31T06:51:18.9075920Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:18.9076070Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_uint8 PASSED [ 81%] 2023-03-31T06:51:18.9076232Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex128 PASSED [ 81%] 2023-03-31T06:51:18.9076372Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:18.9076524Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:18.9076671Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:18.9076819Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:18.9076968Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9077127Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex32 PASSED [ 82%] 2023-03-31T06:51:18.9077289Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex64 PASSED [ 82%] 2023-03-31T06:51:18.9077445Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:18.9077581Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:18.9077730Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:18.9077880Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:18.9078056Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9078208Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_bool PASSED [ 82%] 2023-03-31T06:51:18.9078372Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex128 PASSED [ 82%] 2023-03-31T06:51:18.9078536Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex64 PASSED [ 82%] 2023-03-31T06:51:18.9078694Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:18.9078863Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:18.9079048Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:18.9079198Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:18.9079399Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9079549Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_bool PASSED [ 82%] 2023-03-31T06:51:18.9079700Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex32 PASSED [ 82%] 2023-03-31T06:51:18.9079845Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:18.9079990Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:18.9080123Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:18.9080276Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_complex64 PASSED [ 82%] 2023-03-31T06:51:18.9080424Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:18.9080571Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:18.9080717Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int16 PASSED [ 82%] 2023-03-31T06:51:18.9080862Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:18.9081006Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9081158Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_bfloat16 PASSED [ 82%] 2023-03-31T06:51:18.9081295Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:18.9081445Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:18.9081595Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int16 PASSED [ 82%] 2023-03-31T06:51:18.9081746Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9081897Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_bfloat16 PASSED [ 82%] 2023-03-31T06:51:18.9082045Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_bool PASSED [ 82%] 2023-03-31T06:51:18.9082198Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:18.9082351Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:18.9082497Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:18.9082633Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:18.9082775Z test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:18.9082917Z test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:18.9083275Z test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int8 PASSED [ 83%] 2023-03-31T06:51:18.9083426Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_float16 PASSED [ 83%] 2023-03-31T06:51:18.9083570Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:18.9083714Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:18.9083929Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_int32 PASSED [ 83%] 2023-03-31T06:51:18.9084061Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_int8 PASSED [ 83%] 2023-03-31T06:51:18.9084209Z test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:18.9084359Z test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:18.9084506Z test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:18.9084692Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:18.9084877Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:18.9085024Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:18.9085171Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:18.9085342Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_uint8 PASSED [ 83%] 2023-03-31T06:51:18.9085513Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:18.9085891Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 83%] 2023-03-31T06:51:18.9086059Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:18.9086221Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float16 PASSED [ 83%] 2023-03-31T06:51:18.9086381Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:18.9086723Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 83%] 2023-03-31T06:51:18.9087063Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 83%] 2023-03-31T06:51:18.9087408Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 83%] 2023-03-31T06:51:18.9087564Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:18.9087733Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float16 PASSED [ 83%] 2023-03-31T06:51:18.9087898Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:18.9088056Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:18.9088216Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:18.9088370Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float16 PASSED [ 83%] 2023-03-31T06:51:18.9088525Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:18.9088680Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:18.9088815Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_bool PASSED [ 83%] 2023-03-31T06:51:18.9088964Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:18.9089110Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int32 PASSED [ 83%] 2023-03-31T06:51:18.9089255Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:18.9089403Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int8 PASSED [ 83%] 2023-03-31T06:51:18.9089551Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_uint8 PASSED [ 83%] 2023-03-31T06:51:18.9089701Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:18.9089851Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_complex128 PASSED [ 84%] 2023-03-31T06:51:18.9089988Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:18.9090168Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:18.9090314Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:18.9090460Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:18.9090607Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:18.9090753Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:18.9090927Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:18.9091095Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:18.9091227Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:18.9091402Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:18.9091549Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:18.9091698Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_complex64 PASSED [ 84%] 2023-03-31T06:51:18.9091843Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:18.9091992Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:18.9092134Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:18.9092281Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:18.9092426Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:18.9092555Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:18.9092699Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:18.9092849Z test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float16 XFAIL [ 84%] 2023-03-31T06:51:18.9093002Z test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:18.9093165Z test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_complex128 PASSED [ 84%] 2023-03-31T06:51:18.9093315Z test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:18.9093464Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_bool PASSED [ 84%] 2023-03-31T06:51:18.9093616Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:18.9093756Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:18.9093923Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_complex128 PASSED [ 84%] 2023-03-31T06:51:18.9094083Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_complex64 PASSED [ 84%] 2023-03-31T06:51:18.9094236Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float16 PASSED [ 84%] 2023-03-31T06:51:18.9094386Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:18.9094536Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:18.9094687Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:18.9094838Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:18.9094984Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_complex64 PASSED [ 84%] 2023-03-31T06:51:18.9095138Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:18.9095290Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:18.9095439Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:18.9095591Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:18.9095750Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:18.9095946Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_bool PASSED [ 85%] 2023-03-31T06:51:18.9096194Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_complex128 PASSED [ 85%] 2023-03-31T06:51:18.9096395Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_complex64 PASSED [ 85%] 2023-03-31T06:51:18.9096553Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float16 PASSED [ 85%] 2023-03-31T06:51:18.9096741Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:18.9096917Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:18.9097068Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:18.9097217Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int16 PASSED [ 85%] 2023-03-31T06:51:18.9097398Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int32 PASSED [ 85%] 2023-03-31T06:51:18.9097546Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:18.9097686Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:18.9097843Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_complex128 PASSED [ 85%] 2023-03-31T06:51:18.9097996Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_complex64 PASSED [ 85%] 2023-03-31T06:51:18.9098146Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:18.9098297Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:18.9098449Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:18.9098601Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:18.9098753Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_bool PASSED [ 85%] 2023-03-31T06:51:18.9098893Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:18.9099044Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int32 PASSED [ 85%] 2023-03-31T06:51:18.9099194Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:18.9099342Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_uint8 PASSED [ 85%] 2023-03-31T06:51:18.9099487Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_bool PASSED [ 85%] 2023-03-31T06:51:18.9099632Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:18.9099778Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:18.9099920Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_uint8 PASSED [ 85%] 2023-03-31T06:51:18.9100057Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_bool PASSED [ 85%] 2023-03-31T06:51:18.9100219Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex64 PASSED [ 85%] 2023-03-31T06:51:18.9100377Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float16 PASSED [ 85%] 2023-03-31T06:51:18.9100527Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:18.9100677Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int16 PASSED [ 85%] 2023-03-31T06:51:18.9100827Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int32 PASSED [ 85%] 2023-03-31T06:51:18.9100976Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:18.9101127Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_uint8 PASSED [ 85%] 2023-03-31T06:51:18.9101266Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:18.9101415Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int16 PASSED [ 85%] 2023-03-31T06:51:18.9101563Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:18.9101740Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_uint8 PASSED [ 85%] 2023-03-31T06:51:18.9101891Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_complex128 PASSED [ 85%] 2023-03-31T06:51:18.9102040Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_complex64 PASSED [ 86%] 2023-03-31T06:51:18.9102187Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float16 PASSED [ 86%] 2023-03-31T06:51:18.9102333Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:18.9102610Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_complex128 SKIPPED (meshgrid in torch.complex128 not supported) [ 86%] 2023-03-31T06:51:18.9102862Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float16 SKIPPED (meshgrid in torch.float16 not supported) [ 86%] 2023-03-31T06:51:18.9103106Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int16 SKIPPED (meshgrid in torch.int16 not supported) [ 86%] 2023-03-31T06:51:18.9103319Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int32 SKIPPED (meshgrid in torch.int32 not supported) [ 86%] 2023-03-31T06:51:18.9103618Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int64 SKIPPED (meshgrid in torch.int64 not supported) [ 86%] 2023-03-31T06:51:18.9103831Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_int8 SKIPPED (meshgrid in torch.int8 not supported) [ 86%] 2023-03-31T06:51:18.9104051Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_uint8 SKIPPED (meshgrid in torch.uint8 not supported) [ 86%] 2023-03-31T06:51:18.9104284Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_bfloat16 SKIPPED (meshgrid in torch.bfloat16 not supported) [ 86%] 2023-03-31T06:51:18.9104505Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_bool SKIPPED (meshgrid in torch.bool not supported) [ 86%] 2023-03-31T06:51:18.9104746Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_complex128 SKIPPED (meshgrid in torch.complex128 not supported) [ 86%] 2023-03-31T06:51:18.9104974Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float32 SKIPPED (meshgrid in torch.float32 not supported) [ 86%] 2023-03-31T06:51:18.9105183Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float64 SKIPPED (meshgrid in torch.float64 not supported) [ 86%] 2023-03-31T06:51:18.9105409Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_uint8 SKIPPED (meshgrid in torch.uint8 not supported) [ 86%] 2023-03-31T06:51:18.9105568Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:18.9105723Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:18.9105875Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int16 PASSED [ 86%] 2023-03-31T06:51:18.9106028Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:18.9106179Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:18.9106330Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex64 PASSED [ 86%] 2023-03-31T06:51:18.9106477Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float16 PASSED [ 86%] 2023-03-31T06:51:18.9106612Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:18.9106755Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int16 PASSED [ 86%] 2023-03-31T06:51:18.9106902Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:18.9107049Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int8 PASSED [ 86%] 2023-03-31T06:51:18.9107195Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:18.9107343Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:18.9107531Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_complex128 PASSED [ 86%] 2023-03-31T06:51:18.9107679Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:18.9107810Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int16 PASSED [ 86%] 2023-03-31T06:51:18.9107981Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:18.9108150Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:18.9108344Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:18.9108538Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:18.9108706Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int8 PASSED [ 86%] 2023-03-31T06:51:18.9108900Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:18.9109072Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:18.9109225Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:18.9109386Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int32 PASSED [ 87%] 2023-03-31T06:51:18.9109552Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9109717Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_uint8 PASSED [ 87%] 2023-03-31T06:51:18.9109886Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_float64 PASSED [ 87%] 2023-03-31T06:51:18.9110047Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:18.9110209Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9110371Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_uint8 PASSED [ 87%] 2023-03-31T06:51:18.9110516Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:18.9110667Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:18.9110817Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int32 PASSED [ 87%] 2023-03-31T06:51:18.9110967Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9111117Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_uint8 PASSED [ 87%] 2023-03-31T06:51:18.9111268Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_bool PASSED [ 87%] 2023-03-31T06:51:18.9111422Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float16 PASSED [ 87%] 2023-03-31T06:51:18.9111570Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:18.9111707Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9111870Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:18.9112021Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_bool PASSED [ 87%] 2023-03-31T06:51:18.9112184Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex128 PASSED [ 87%] 2023-03-31T06:51:18.9112345Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex32 PASSED [ 87%] 2023-03-31T06:51:18.9112505Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_complex64 PASSED [ 87%] 2023-03-31T06:51:18.9112662Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float64 PASSED [ 87%] 2023-03-31T06:51:18.9112815Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:18.9112967Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int32 PASSED [ 87%] 2023-03-31T06:51:18.9113109Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9113291Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_uint8 PASSED [ 87%] 2023-03-31T06:51:18.9113521Z test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_bfloat16 SKIPPED (native_batch_norm in torch.bfloat16 not supported) [ 87%] 2023-03-31T06:51:18.9113746Z test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_float64 SKIPPED (native_batch_norm in torch.float64 not supported) [ 87%] 2023-03-31T06:51:18.9113917Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_bool PASSED [ 87%] 2023-03-31T06:51:18.9114154Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float32 PASSED [ 87%] 2023-03-31T06:51:18.9114347Z test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:18.9114511Z test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_float32 PASSED [ 87%] 2023-03-31T06:51:18.9114676Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:18.9114829Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_bool PASSED [ 87%] 2023-03-31T06:51:18.9114979Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_complex128 PASSED [ 87%] 2023-03-31T06:51:18.9115125Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int64 PASSED [ 87%] 2023-03-31T06:51:18.9115269Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:18.9115414Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_uint8 PASSED [ 87%] 2023-03-31T06:51:18.9115566Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex128 PASSED [ 88%] 2023-03-31T06:51:18.9115718Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex64 PASSED [ 88%] 2023-03-31T06:51:18.9115853Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:18.9116001Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:18.9116147Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int32 PASSED [ 88%] 2023-03-31T06:51:18.9116293Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int8 PASSED [ 88%] 2023-03-31T06:51:18.9116440Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_uint8 PASSED [ 88%] 2023-03-31T06:51:18.9116649Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float16 SKIPPED (new_empty in torch.float16 not supported) [ 88%] 2023-03-31T06:51:18.9116851Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int16 SKIPPED (new_empty in torch.int16 not supported) [ 88%] 2023-03-31T06:51:18.9117048Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int32 SKIPPED (new_empty in torch.int32 not supported) [ 88%] 2023-03-31T06:51:18.9117240Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int64 SKIPPED (new_empty in torch.int64 not supported) [ 88%] 2023-03-31T06:51:18.9117420Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_uint8 SKIPPED (new_empty in torch.uint8 not supported) [ 88%] 2023-03-31T06:51:18.9117644Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_bool SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9117868Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex128 SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9118093Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex64 SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9118315Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float16 SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9118538Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float32 SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9118756Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int64 SKIPPED (Expected: new_empty_strided is not comparable) [ 88%] 2023-03-31T06:51:18.9118912Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:18.9119097Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:18.9119251Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int16 PASSED [ 88%] 2023-03-31T06:51:18.9119389Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int32 PASSED [ 88%] 2023-03-31T06:51:18.9119536Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int64 PASSED [ 88%] 2023-03-31T06:51:18.9119686Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_uint8 PASSED [ 88%] 2023-03-31T06:51:18.9119862Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:18.9120043Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex32 PASSED [ 88%] 2023-03-31T06:51:18.9120193Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:18.9120378Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:18.9120526Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int32 PASSED [ 88%] 2023-03-31T06:51:18.9120664Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:18.9120823Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex128 PASSED [ 88%] 2023-03-31T06:51:18.9120976Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex32 PASSED [ 88%] 2023-03-31T06:51:18.9121124Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:18.9121274Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int16 PASSED [ 88%] 2023-03-31T06:51:18.9121426Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int32 PASSED [ 88%] 2023-03-31T06:51:18.9121573Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int8 PASSED [ 88%] 2023-03-31T06:51:18.9121723Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_uint8 PASSED [ 88%] 2023-03-31T06:51:18.9121862Z test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:18.9122053Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:18.9122240Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9122408Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9122576Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9122754Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9122929Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9123343Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9123546Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9123710Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9123885Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9124059Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9124239Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9124417Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9124591Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9124764Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9124935Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9125095Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9125354Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int64 PASSED [ 89%] 2023-03-31T06:51:18.9125524Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int8 PASSED [ 89%] 2023-03-31T06:51:18.9125698Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9125869Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float16 PASSED [ 89%] 2023-03-31T06:51:18.9126076Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9126278Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9126453Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9126642Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9126814Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9127241Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 89%] 2023-03-31T06:51:18.9127412Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9127593Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9127776Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_complex64 PASSED [ 89%] 2023-03-31T06:51:18.9127948Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int32 PASSED [ 89%] 2023-03-31T06:51:18.9128121Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int8 PASSED [ 89%] 2023-03-31T06:51:18.9128291Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9128515Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_float64 SKIPPED (nn.functional.relu6 in torch.float64 not supported) [ 89%] 2023-03-31T06:51:18.9128744Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int8 SKIPPED (nn.functional.relu6 in torch.int8 not supported) [ 89%] 2023-03-31T06:51:18.9128974Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_uint8 SKIPPED (nn.functional.relu6 in torch.uint8 not supported) [ 89%] 2023-03-31T06:51:18.9129143Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:18.9129314Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:18.9129480Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_uint8 PASSED [ 89%] 2023-03-31T06:51:18.9129651Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9129821Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:18.9129996Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_float64 PASSED [ 90%] 2023-03-31T06:51:18.9130156Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:18.9130330Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_float64 PASSED [ 90%] 2023-03-31T06:51:18.9130501Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_complex64 PASSED [ 90%] 2023-03-31T06:51:18.9130668Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float16 PASSED [ 90%] 2023-03-31T06:51:18.9130837Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:18.9131033Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 90%] 2023-03-31T06:51:18.9131246Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 90%] 2023-03-31T06:51:18.9131485Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 90%] 2023-03-31T06:51:18.9131683Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 90%] 2023-03-31T06:51:18.9131869Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 90%] 2023-03-31T06:51:18.9132103Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 90%] 2023-03-31T06:51:18.9132333Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 90%] 2023-03-31T06:51:18.9132540Z test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 90%] 2023-03-31T06:51:18.9132767Z test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 90%] 2023-03-31T06:51:18.9132934Z test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float32 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:18.9133099Z test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float64 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:18.9133263Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_complex64 XFAIL [ 90%] 2023-03-31T06:51:18.9133425Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float64 XFAIL [ 90%] 2023-03-31T06:51:18.9133596Z test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float16 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:18.9133753Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_bfloat16 PASSED [ 90%] 2023-03-31T06:51:18.9133908Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex128 PASSED [ 90%] 2023-03-31T06:51:18.9134064Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex32 PASSED [ 90%] 2023-03-31T06:51:18.9134217Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_complex64 PASSED [ 90%] 2023-03-31T06:51:18.9134367Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float16 PASSED [ 90%] 2023-03-31T06:51:18.9134515Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float64 PASSED [ 90%] 2023-03-31T06:51:18.9134662Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int64 PASSED [ 90%] 2023-03-31T06:51:18.9134794Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_uint8 PASSED [ 90%] 2023-03-31T06:51:18.9134947Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_bfloat16 PASSED [ 90%] 2023-03-31T06:51:18.9135097Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_bool PASSED [ 90%] 2023-03-31T06:51:18.9135249Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex32 PASSED [ 90%] 2023-03-31T06:51:18.9135401Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex64 PASSED [ 90%] 2023-03-31T06:51:18.9135552Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float16 PASSED [ 90%] 2023-03-31T06:51:18.9135703Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:18.9135853Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_float64 PASSED [ 90%] 2023-03-31T06:51:18.9135987Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int16 PASSED [ 90%] 2023-03-31T06:51:18.9136134Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int64 PASSED [ 90%] 2023-03-31T06:51:18.9136278Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int8 PASSED [ 90%] 2023-03-31T06:51:18.9136432Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_bfloat16 PASSED [ 90%] 2023-03-31T06:51:18.9136582Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_bool PASSED [ 91%] 2023-03-31T06:51:18.9136733Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:18.9136884Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:18.9137058Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:18.9137208Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_complex128 PASSED [ 91%] 2023-03-31T06:51:18.9137340Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:18.9137485Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_uint8 PASSED [ 91%] 2023-03-31T06:51:18.9137633Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:18.9137807Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_bool PASSED [ 91%] 2023-03-31T06:51:18.9137979Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:18.9138126Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int8 PASSED [ 91%] 2023-03-31T06:51:18.9138276Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex32 XFAIL [ 91%] 2023-03-31T06:51:18.9138453Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float16 XFAIL [ 91%] 2023-03-31T06:51:18.9138587Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float32 XFAIL [ 91%] 2023-03-31T06:51:18.9138733Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_float64 XFAIL [ 91%] 2023-03-31T06:51:18.9138884Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_bool PASSED [ 91%] 2023-03-31T06:51:18.9139045Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_complex128 PASSED [ 91%] 2023-03-31T06:51:18.9139199Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:18.9139353Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int16 PASSED [ 91%] 2023-03-31T06:51:18.9139503Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:18.9139652Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int32 PASSED [ 91%] 2023-03-31T06:51:18.9139791Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_uint8 PASSED [ 91%] 2023-03-31T06:51:18.9139941Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:18.9140092Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_complex128 PASSED [ 91%] 2023-03-31T06:51:18.9140241Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:18.9140388Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:18.9140534Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:18.9140683Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int16 PASSED [ 91%] 2023-03-31T06:51:18.9140829Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int32 PASSED [ 91%] 2023-03-31T06:51:18.9140962Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:18.9141109Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_int8 PASSED [ 91%] 2023-03-31T06:51:18.9141259Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_uint8 PASSED [ 91%] 2023-03-31T06:51:18.9141405Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex32 PASSED [ 91%] 2023-03-31T06:51:18.9141548Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int16 PASSED [ 91%] 2023-03-31T06:51:18.9141692Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int32 PASSED [ 91%] 2023-03-31T06:51:18.9141836Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:18.9141982Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_uint8 PASSED [ 91%] 2023-03-31T06:51:18.9142119Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_complex128 PASSED [ 91%] 2023-03-31T06:51:18.9142263Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:18.9142407Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:18.9142551Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:18.9142725Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_uint8 PASSED [ 92%] 2023-03-31T06:51:18.9142876Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:18.9143024Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_complex64 PASSED [ 92%] 2023-03-31T06:51:18.9143169Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:18.9143301Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:18.9143577Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:18.9143759Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_uint8 PASSED [ 92%] 2023-03-31T06:51:18.9143907Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:51:18.9144085Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:18.9144231Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:18.9144379Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:18.9144523Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:18.9144654Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_uint8 PASSED [ 92%] 2023-03-31T06:51:18.9144804Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:51:18.9145158Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:18.9145311Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:18.9145462Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex64 PASSED [ 92%] 2023-03-31T06:51:18.9145613Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_float16 PASSED [ 92%] 2023-03-31T06:51:18.9145762Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:18.9145910Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:18.9146240Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:18.9146554Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:18.9146706Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:51:18.9146860Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex32 PASSED [ 92%] 2023-03-31T06:51:18.9147008Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:18.9147155Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:18.9147301Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:18.9147449Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:18.9147593Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:18.9147729Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_bool PASSED [ 92%] 2023-03-31T06:51:18.9147881Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_complex64 PASSED [ 92%] 2023-03-31T06:51:18.9148032Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:18.9148181Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:18.9148329Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:18.9148476Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:18.9148626Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:18.9148805Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_uint8 PASSED [ 92%] 2023-03-31T06:51:18.9148941Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:51:18.9149088Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_bool PASSED [ 93%] 2023-03-31T06:51:18.9149232Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:18.9149381Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:18.9149555Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:18.9149727Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:18.9149876Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:18.9150049Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:18.9150184Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:18.9150332Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_complex64 PASSED [ 93%] 2023-03-31T06:51:18.9150478Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:18.9150620Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:18.9150765Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:18.9150913Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_bool PASSED [ 93%] 2023-03-31T06:51:18.9151064Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_complex128 PASSED [ 93%] 2023-03-31T06:51:18.9151211Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_complex64 PASSED [ 93%] 2023-03-31T06:51:18.9151346Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:18.9151492Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:18.9151639Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:18.9151780Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:18.9151929Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:18.9152073Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_bool PASSED [ 93%] 2023-03-31T06:51:18.9152224Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex32 PASSED [ 93%] 2023-03-31T06:51:18.9152376Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex64 PASSED [ 93%] 2023-03-31T06:51:18.9152510Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float16 PASSED [ 93%] 2023-03-31T06:51:18.9152657Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:18.9152803Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:18.9152948Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:18.9153092Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:18.9153237Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:18.9153397Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:18.9153556Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:18.9153709Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:18.9153848Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:18.9154000Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_bool PASSED [ 93%] 2023-03-31T06:51:18.9154160Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:18.9154350Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:18.9154502Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:18.9154654Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:18.9154806Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:18.9154956Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:18.9155119Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_bool PASSED [ 94%] 2023-03-31T06:51:18.9155295Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:18.9155443Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:18.9155592Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:18.9155767Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:18.9155920Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:18.9156070Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:18.9156218Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:18.9156359Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:18.9156524Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:18.9156688Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:18.9156846Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:18.9157001Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:18.9157220Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_bool SKIPPED (special.ndtr in torch.bool not supported) [ 94%] 2023-03-31T06:51:18.9157443Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_float64 SKIPPED (special.ndtr in torch.float64 not supported) [ 94%] 2023-03-31T06:51:18.9157655Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int64 SKIPPED (special.ndtr in torch.int64 not supported) [ 94%] 2023-03-31T06:51:18.9157868Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_uint8 SKIPPED (special.ndtr in torch.uint8 not supported) [ 94%] 2023-03-31T06:51:18.9158015Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_bool PASSED [ 94%] 2023-03-31T06:51:18.9158172Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:18.9158325Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:18.9158492Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_bfloat16 PASSED [ 94%] 2023-03-31T06:51:18.9158654Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_bool PASSED [ 94%] 2023-03-31T06:51:18.9158820Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float16 PASSED [ 94%] 2023-03-31T06:51:18.9158979Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:18.9159135Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:18.9159283Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:18.9159446Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:18.9159603Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_bool PASSED [ 94%] 2023-03-31T06:51:18.9159759Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:18.9159919Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:18.9160075Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:18.9160303Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_bfloat16 PASSED [ 94%] 2023-03-31T06:51:18.9160454Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_bool PASSED [ 94%] 2023-03-31T06:51:18.9160596Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex32 PASSED [ 94%] 2023-03-31T06:51:18.9160746Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex64 PASSED [ 94%] 2023-03-31T06:51:18.9160896Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_float16 PASSED [ 94%] 2023-03-31T06:51:18.9161072Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:18.9161255Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:18.9161402Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:18.9161584Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:18.9161752Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:18.9161902Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:18.9162060Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int16 PASSED [ 95%] 2023-03-31T06:51:18.9162219Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:18.9162385Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:18.9162552Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:18.9162712Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:18.9162868Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:18.9163266Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int32 PASSED [ 95%] 2023-03-31T06:51:18.9163412Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:18.9163566Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:18.9163715Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_bool PASSED [ 95%] 2023-03-31T06:51:18.9163862Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int32 PASSED [ 95%] 2023-03-31T06:51:18.9164012Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:18.9164166Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_bool PASSED [ 95%] 2023-03-31T06:51:18.9164322Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:18.9164473Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:18.9164609Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int16 PASSED [ 95%] 2023-03-31T06:51:18.9164760Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int32 PASSED [ 95%] 2023-03-31T06:51:18.9164908Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:18.9165057Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:18.9165207Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:18.9165375Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:18.9165538Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_bool PASSED [ 95%] 2023-03-31T06:51:18.9165710Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:18.9165872Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:18.9166012Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:18.9166167Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex32 PASSED [ 95%] 2023-03-31T06:51:18.9166381Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:18.9166532Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:18.9166678Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int32 PASSED [ 95%] 2023-03-31T06:51:18.9166822Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:18.9166973Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:18.9167155Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:18.9167327Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:18.9167479Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:18.9167685Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:18.9167857Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:18.9168024Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:18.9168187Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float16 PASSED [ 96%] 2023-03-31T06:51:18.9168345Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:18.9168501Z test_decomp.py::TestDecompCPU::test_quick_std_mean_unbiased_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:18.9168650Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:18.9168816Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:18.9168975Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float16 PASSED [ 96%] 2023-03-31T06:51:18.9169135Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:18.9169293Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:18.9169445Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:18.9169597Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float16 PASSED [ 96%] 2023-03-31T06:51:18.9169745Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:18.9169879Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:18.9170027Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:18.9170172Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:18.9170318Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_uint8 PASSED [ 96%] 2023-03-31T06:51:18.9170462Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_bool PASSED [ 96%] 2023-03-31T06:51:18.9170611Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:18.9170757Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:18.9170899Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:18.9171030Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_bool PASSED [ 96%] 2023-03-31T06:51:18.9171177Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:18.9171321Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:18.9171465Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:18.9171608Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:18.9171751Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:18.9171898Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_bool PASSED [ 96%] 2023-03-31T06:51:18.9172077Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:18.9172213Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:18.9172358Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:18.9172502Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:18.9172645Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:18.9172815Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_uint8 PASSED [ 96%] 2023-03-31T06:51:18.9172995Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:18.9173139Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:18.9173314Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:18.9173446Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_uint8 PASSED [ 96%] 2023-03-31T06:51:18.9173594Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:18.9173737Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:18.9173882Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:18.9174035Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:18.9174188Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_bool PASSED [ 97%] 2023-03-31T06:51:18.9174346Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex32 PASSED [ 97%] 2023-03-31T06:51:18.9174500Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex64 PASSED [ 97%] 2023-03-31T06:51:18.9174649Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float16 PASSED [ 97%] 2023-03-31T06:51:18.9174789Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:18.9174940Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:18.9175090Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:18.9175240Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:18.9175389Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:18.9175537Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:18.9175684Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_bool PASSED [ 97%] 2023-03-31T06:51:18.9175834Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_complex128 PASSED [ 97%] 2023-03-31T06:51:18.9175967Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float16 PASSED [ 97%] 2023-03-31T06:51:18.9176113Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:18.9176258Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:18.9176402Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:18.9176555Z test_decomp.py::TestDecompCPU::test_quick_tril_indices_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:18.9176699Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_bool PASSED [ 97%] 2023-03-31T06:51:18.9176848Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_complex128 PASSED [ 97%] 2023-03-31T06:51:18.9176999Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_complex64 PASSED [ 97%] 2023-03-31T06:51:18.9177136Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:18.9177280Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:18.9177430Z test_decomp.py::TestDecompCPU::test_quick_triu_indices_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:18.9177584Z test_decomp.py::TestDecompCPU::test_quick_triu_indices_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:18.9177771Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:18.9177917Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:18.9178068Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:18.9178216Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_bool PASSED [ 97%] 2023-03-31T06:51:18.9178354Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex32 PASSED [ 97%] 2023-03-31T06:51:18.9178548Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:18.9178720Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:18.9178865Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:18.9179040Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:18.9179200Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:18.9179352Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_bool PASSED [ 97%] 2023-03-31T06:51:18.9179514Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex128 PASSED [ 97%] 2023-03-31T06:51:18.9179660Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex64 PASSED [ 97%] 2023-03-31T06:51:18.9179811Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int16 PASSED [ 98%] 2023-03-31T06:51:18.9179964Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_uint8 PASSED [ 98%] 2023-03-31T06:51:18.9180113Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_bool PASSED [ 98%] 2023-03-31T06:51:18.9180265Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex32 PASSED [ 98%] 2023-03-31T06:51:18.9180420Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9180569Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:18.9180718Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:18.9180852Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int64 PASSED [ 98%] 2023-03-31T06:51:18.9180998Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int8 PASSED [ 98%] 2023-03-31T06:51:18.9181146Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_uint8 PASSED [ 98%] 2023-03-31T06:51:18.9181298Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_bfloat16 XFAIL [ 98%] 2023-03-31T06:51:18.9181449Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float16 XFAIL [ 98%] 2023-03-31T06:51:18.9181597Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float32 XFAIL [ 98%] 2023-03-31T06:51:18.9181749Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_bool PASSED [ 98%] 2023-03-31T06:51:18.9181914Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9182063Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:18.9182221Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:18.9182377Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:18.9182528Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int32 PASSED [ 98%] 2023-03-31T06:51:18.9182680Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_uint8 PASSED [ 98%] 2023-03-31T06:51:18.9182833Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_bool PASSED [ 98%] 2023-03-31T06:51:18.9182995Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:18.9183147Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9183287Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:18.9183573Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:18.9183731Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int64 PASSED [ 98%] 2023-03-31T06:51:18.9183881Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int8 PASSED [ 98%] 2023-03-31T06:51:18.9184032Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_uint8 PASSED [ 98%] 2023-03-31T06:51:18.9184179Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:18.9184371Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:18.9184546Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9184679Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:18.9184859Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:18.9185011Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9185181Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:18.9185350Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:18.9185514Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:18.9185676Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:18.9185841Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:18.9186001Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:18.9186136Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_bool PASSED [ 99%] 2023-03-31T06:51:18.9186288Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex32 PASSED [ 99%] 2023-03-31T06:51:18.9186435Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_int64 PASSED [ 99%] 2023-03-31T06:51:18.9186583Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_uint8 PASSED [ 99%] 2023-03-31T06:51:18.9186734Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:51:18.9186885Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex128 PASSED [ 99%] 2023-03-31T06:51:18.9187032Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_float16 PASSED [ 99%] 2023-03-31T06:51:18.9187178Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_int16 PASSED [ 99%] 2023-03-31T06:51:18.9187312Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_int32 PASSED [ 99%] 2023-03-31T06:51:18.9187460Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_int64 PASSED [ 99%] 2023-03-31T06:51:18.9187609Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:51:18.9187756Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int16 PASSED [ 99%] 2023-03-31T06:51:18.9187903Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int64 PASSED [ 99%] 2023-03-31T06:51:18.9188112Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_complex128 SKIPPED (zero_ in torch.complex128 not supported) [ 99%] 2023-03-31T06:51:18.9188318Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_complex64 SKIPPED (zero_ in torch.complex64 not supported) [ 99%] 2023-03-31T06:51:18.9188518Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float32 SKIPPED (zero_ in torch.float32 not supported) [ 99%] 2023-03-31T06:51:18.9188698Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int32 SKIPPED (zero_ in torch.int32 not supported) [ 99%] 2023-03-31T06:51:18.9188891Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int8 SKIPPED (zero_ in torch.int8 not supported) [ 99%] 2023-03-31T06:51:18.9189076Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_uint8 SKIPPED (zero_ in torch.uint8 not supported) [ 99%] 2023-03-31T06:51:18.9189231Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:51:18.9189410Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_bool PASSED [ 99%] 2023-03-31T06:51:18.9189565Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex32 PASSED [ 99%] 2023-03-31T06:51:18.9189716Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float16 PASSED [ 99%] 2023-03-31T06:51:18.9189865Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:18.9190038Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int8 PASSED [ 99%] 2023-03-31T06:51:18.9190204Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_bfloat16 PASSED [ 99%] 2023-03-31T06:51:18.9190355Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_bool PASSED [ 99%] 2023-03-31T06:51:18.9190515Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex128 PASSED [ 99%] 2023-03-31T06:51:18.9190703Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex32 PASSED [ 99%] 2023-03-31T06:51:18.9190864Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_complex64 PASSED [ 99%] 2023-03-31T06:51:18.9191016Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float16 PASSED [ 99%] 2023-03-31T06:51:18.9191168Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:18.9191318Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int64 PASSED [ 99%] 2023-03-31T06:51:18.9191487Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_eval_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:18.9191669Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_eval_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:18.9191851Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_train_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:18.9192031Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_train_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:18.9192215Z test_decomp.py::DecompContiguousTestsCPU::test_contiguous_softmax_cpu PASSED [ 99%] 2023-03-31T06:51:18.9192369Z test_decomp.py::HasDecompTest::test_has_decomposition PASSED [100%] 2023-03-31T06:51:18.9192381Z 2023-03-31T06:51:18.9192767Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-d4f07a676c877163.xml - 2023-03-31T06:51:18.9192912Z === 3608 passed, 303 skipped, 4 deselected, 17 xfailed in 1585.34s (0:26:25) === 2023-03-31T06:51:18.9193096Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:51:18.9193106Z 2023-03-31T06:51:18.9193472Z ##[endgroup] 2023-03-31T06:51:18.9193777Z FINISHED PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_geq5cc7g.log) 2023-03-31T06:51:18.9193784Z 2023-03-31T06:51:21.1422206Z 2023-03-31T06:51:21.1422780Z Expand the folded group to see the log file of test_decomp 2023-03-31T06:51:21.1423922Z ##[group]PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_99_jfdlp.log) 2023-03-31T06:51:21.1429765Z Test results will be stored in test-reports/python-pytest/test_decomp/test_decomp-9412a89c8914e16a.xml 2023-03-31T06:51:21.1430530Z ============================= test session starts ============================== 2023-03-31T06:51:21.1431242Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:51:21.1431734Z cachedir: .pytest_cache 2023-03-31T06:51:21.1432521Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:51:21.1433200Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:51:21.1434017Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:51:21.1434638Z collecting ... collected 7785 items / 4 deselected / 7781 selected 2023-03-31T06:51:21.1958412Z Running 3849 items in this shard: test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_shapes_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_grid_sampler_2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_histogram_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_histogramdd_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_histogramdd_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_istft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nanquantile_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_ctc_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_grid_sample_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_area_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_grad_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_grad_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multi_margin_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multi_margin_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_margin_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_margin_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pdist_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_complex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_pca_lowrank_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polar_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_quantile_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_bartlett_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_bartlett_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_blackman_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_cosine_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_gaussian_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_hamming_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hamming_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hamming_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hann_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_kaiser_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_kaiser_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_nuttall_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_tril_indices_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_indices_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_triu_indices_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_real_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_as_real_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_bool, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int8, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int64, test/test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_abs_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_acos_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_add_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_addr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_all_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_amax_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_any_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_atan_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_cat_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_clone_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_cos_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_expand_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_eye_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fill_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_frac_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_full_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_ge_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_grid_sampler_2d_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_le_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_log_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_logit_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_mv_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_normal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_permute_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_prod_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_select_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_split_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_stack_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_std_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_sum_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_t_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_tril_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_tril_indices_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_triu_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_var_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_view_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex32, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_where_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_bfloat16, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_bool, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float16, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int64, test/test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int16, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int32, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int8, test/test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_uint8, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_train_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_eval_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_eval_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_train_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_train_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_eval_mode_cpu_float32, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_eval_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_train_mode_cpu_float64, test/test_decomp.py::TestDecompCPU::test_uniform_cpu, test/test_decomp.py::DecompContiguousTestsCPU::test_contiguous_log_softmax_cpu, test/test_decomp.py::DecompAmpTestsCPU::test_amp_batch_norm_backward_cpu 2023-03-31T06:51:21.2296036Z 2023-03-31T06:51:21.2298690Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:21.2299073Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_bool PASSED [ 0%] 2023-03-31T06:51:21.2299745Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex128 PASSED [ 0%] 2023-03-31T06:51:21.2300151Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex32 PASSED [ 0%] 2023-03-31T06:51:21.2300482Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_complex64 PASSED [ 0%] 2023-03-31T06:51:21.2300876Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_float16 PASSED [ 0%] 2023-03-31T06:51:21.2301213Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int32 PASSED [ 0%] 2023-03-31T06:51:21.2301528Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_int64 PASSED [ 0%] 2023-03-31T06:51:21.2301849Z test_decomp.py::TestDecompCPU::test_comprehensive_H_cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:21.2302194Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_complex128 PASSED [ 0%] 2023-03-31T06:51:21.2302529Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float16 PASSED [ 0%] 2023-03-31T06:51:21.2302850Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_float64 PASSED [ 0%] 2023-03-31T06:51:21.2303177Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int16 PASSED [ 0%] 2023-03-31T06:51:21.2303597Z test_decomp.py::TestDecompCPU::test_comprehensive_T_cpu_int32 PASSED [ 0%] 2023-03-31T06:51:21.2303936Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:21.2304293Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_bool PASSED [ 0%] 2023-03-31T06:51:21.2304658Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex32 PASSED [ 0%] 2023-03-31T06:51:21.2305022Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_complex64 PASSED [ 0%] 2023-03-31T06:51:21.2305366Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:21.2305715Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int64 PASSED [ 0%] 2023-03-31T06:51:21.2306066Z test_decomp.py::TestDecompCPU::test_comprehensive___getitem___cpu_int8 PASSED [ 0%] 2023-03-31T06:51:21.2306408Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:21.2306761Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_complex128 PASSED [ 0%] 2023-03-31T06:51:21.2307115Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float16 PASSED [ 0%] 2023-03-31T06:51:21.2307465Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float32 PASSED [ 0%] 2023-03-31T06:51:21.2307802Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_float64 PASSED [ 0%] 2023-03-31T06:51:21.2308147Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:21.2308492Z test_decomp.py::TestDecompCPU::test_comprehensive___radd___cpu_int8 PASSED [ 0%] 2023-03-31T06:51:21.2308820Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int32 PASSED [ 0%] 2023-03-31T06:51:21.2309164Z test_decomp.py::TestDecompCPU::test_comprehensive___rand___cpu_int8 PASSED [ 0%] 2023-03-31T06:51:21.2309518Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_complex128 PASSED [ 0%] 2023-03-31T06:51:21.2309878Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_complex64 PASSED [ 0%] 2023-03-31T06:51:21.2310218Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_float64 PASSED [ 0%] 2023-03-31T06:51:21.2310622Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int16 PASSED [ 0%] 2023-03-31T06:51:21.2310968Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int32 PASSED [ 0%] 2023-03-31T06:51:21.2311295Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_int8 PASSED [ 0%] 2023-03-31T06:51:21.2311637Z test_decomp.py::TestDecompCPU::test_comprehensive___rdiv___cpu_uint8 PASSED [ 0%] 2023-03-31T06:51:21.2311986Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_bfloat16 PASSED [ 0%] 2023-03-31T06:51:21.2312406Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_complex128 PASSED [ 1%] 2023-03-31T06:51:21.2312790Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_complex64 PASSED [ 1%] 2023-03-31T06:51:21.2313147Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_float32 PASSED [ 1%] 2023-03-31T06:51:21.2313531Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2313889Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int32 PASSED [ 1%] 2023-03-31T06:51:21.2314226Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:21.2314574Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:21.2314924Z test_decomp.py::TestDecompCPU::test_comprehensive___rmatmul___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:21.2315261Z test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:21.2315611Z test_decomp.py::TestDecompCPU::test_comprehensive___rmod___cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2315963Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_bfloat16 PASSED [ 1%] 2023-03-31T06:51:21.2316307Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_bool PASSED [ 1%] 2023-03-31T06:51:21.2316645Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_complex128 PASSED [ 1%] 2023-03-31T06:51:21.2317004Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_complex64 PASSED [ 1%] 2023-03-31T06:51:21.2317357Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_float16 PASSED [ 1%] 2023-03-31T06:51:21.2317691Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:21.2318034Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:21.2318373Z test_decomp.py::TestDecompCPU::test_comprehensive___rmul___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:21.2318715Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int32 PASSED [ 1%] 2023-03-31T06:51:21.2319036Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:21.2319373Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_int8 PASSED [ 1%] 2023-03-31T06:51:21.2319711Z test_decomp.py::TestDecompCPU::test_comprehensive___ror___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:21.2320038Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2320384Z test_decomp.py::TestDecompCPU::test_comprehensive___rpow___cpu_int32 PASSED [ 1%] 2023-03-31T06:51:21.2320732Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_complex128 PASSED [ 1%] 2023-03-31T06:51:21.2321083Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float32 PASSED [ 1%] 2023-03-31T06:51:21.2321415Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2321761Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int32 PASSED [ 1%] 2023-03-31T06:51:21.2322101Z test_decomp.py::TestDecompCPU::test_comprehensive___rsub___cpu_int64 PASSED [ 1%] 2023-03-31T06:51:21.2322433Z test_decomp.py::TestDecompCPU::test_comprehensive___rxor___cpu_uint8 PASSED [ 1%] 2023-03-31T06:51:21.2322803Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_lengths_cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2323422Z test_decomp.py::TestDecompCPU::test_comprehensive__segment_reduce_offsets_cpu_float16 PASSED [ 1%] 2023-03-31T06:51:21.2323886Z test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_bfloat16 SKIPPED (_softmax_backward_data in torch.bfloat16 not supported) [ 1%] 2023-03-31T06:51:21.2324327Z test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_float32 PASSED [ 1%] 2023-03-31T06:51:21.2324715Z test_decomp.py::TestDecompCPU::test_comprehensive__softmax_backward_data_cpu_float64 PASSED [ 1%] 2023-03-31T06:51:21.2325211Z test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_float32 SKIPPED (_upsample_bilinear2d_aa in torch.float32 not supported) [ 1%] 2023-03-31T06:51:21.2325770Z test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_float64 SKIPPED (_upsample_bilinear2d_aa in torch.float64 not supported) [ 1%] 2023-03-31T06:51:21.2326314Z test_decomp.py::TestDecompCPU::test_comprehensive__upsample_bilinear2d_aa_cpu_uint8 SKIPPED (_upsample_bilinear2d_aa in torch.uint8 not supported) [ 1%] 2023-03-31T06:51:21.2326745Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float16 PASSED [ 2%] 2023-03-31T06:51:21.2327087Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:21.2327426Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:21.2327745Z test_decomp.py::TestDecompCPU::test_comprehensive_abs_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:21.2328090Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:21.2328440Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:21.2328760Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:21.2329087Z test_decomp.py::TestDecompCPU::test_comprehensive_acos_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:21.2329421Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_bool PASSED [ 2%] 2023-03-31T06:51:21.2329773Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:21.2330120Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:21.2330469Z test_decomp.py::TestDecompCPU::test_comprehensive_acosh_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:21.2330820Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:21.2331155Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_complex32 PASSED [ 2%] 2023-03-31T06:51:21.2331501Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float16 PASSED [ 2%] 2023-03-31T06:51:21.2331838Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:21.2332170Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:21.2332490Z test_decomp.py::TestDecompCPU::test_comprehensive_add_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:21.2332836Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:21.2333191Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:21.2333535Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:21.2333885Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int16 PASSED [ 2%] 2023-03-31T06:51:21.2334225Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:21.2334569Z test_decomp.py::TestDecompCPU::test_comprehensive_addbmm_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:21.2334908Z test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:21.2335263Z test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:21.2335620Z test_decomp.py::TestDecompCPU::test_comprehensive_addcdiv_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:21.2353602Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_bfloat16 PASSED [ 2%] 2023-03-31T06:51:21.2354145Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:21.2354518Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:21.2354868Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_float64 PASSED [ 2%] 2023-03-31T06:51:21.2355222Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int32 PASSED [ 2%] 2023-03-31T06:51:21.2355617Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:21.2355995Z test_decomp.py::TestDecompCPU::test_comprehensive_addcmul_cpu_uint8 PASSED [ 2%] 2023-03-31T06:51:21.2356349Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_complex128 PASSED [ 2%] 2023-03-31T06:51:21.2356750Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_complex64 PASSED [ 2%] 2023-03-31T06:51:21.2357108Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_float32 PASSED [ 2%] 2023-03-31T06:51:21.2357438Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int64 PASSED [ 2%] 2023-03-31T06:51:21.2357779Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_cpu_int8 PASSED [ 2%] 2023-03-31T06:51:21.2358147Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:21.2358517Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_complex128 PASSED [ 3%] 2023-03-31T06:51:21.2358907Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:21.2359285Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:21.2359662Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:21.2360024Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:21.2360394Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:21.2360761Z test_decomp.py::TestDecompCPU::test_comprehensive_addmm_decomposed_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:21.2361125Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_complex64 PASSED [ 3%] 2023-03-31T06:51:21.2361465Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:21.2361813Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:21.2362165Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:21.2362492Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:21.2362831Z test_decomp.py::TestDecompCPU::test_comprehensive_addmv_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:21.2363401Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_complex128 PASSED [ 3%] 2023-03-31T06:51:21.2363755Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float16 PASSED [ 3%] 2023-03-31T06:51:21.2364088Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:21.2364427Z test_decomp.py::TestDecompCPU::test_comprehensive_addr_cpu_int16 PASSED [ 3%] 2023-03-31T06:51:21.2364776Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_complex128 PASSED [ 3%] 2023-03-31T06:51:21.2365109Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_float16 PASSED [ 3%] 2023-03-31T06:51:21.2365452Z test_decomp.py::TestDecompCPU::test_comprehensive_all_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:21.2365800Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float16 PASSED [ 3%] 2023-03-31T06:51:21.2366155Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:21.2366501Z test_decomp.py::TestDecompCPU::test_comprehensive_allclose_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:21.2366951Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_bfloat16 PASSED [ 3%] 2023-03-31T06:51:21.2367300Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:21.2367630Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:21.2367971Z test_decomp.py::TestDecompCPU::test_comprehensive_amax_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:21.2368311Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_bool PASSED [ 3%] 2023-03-31T06:51:21.2368694Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_float16 PASSED [ 3%] 2023-03-31T06:51:21.2369057Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:21.2369390Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:21.2369785Z test_decomp.py::TestDecompCPU::test_comprehensive_amin_cpu_int8 PASSED [ 3%] 2023-03-31T06:51:21.2370122Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_float32 PASSED [ 3%] 2023-03-31T06:51:21.2370470Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_float64 PASSED [ 3%] 2023-03-31T06:51:21.2370821Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int32 PASSED [ 3%] 2023-03-31T06:51:21.2371169Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_int64 PASSED [ 3%] 2023-03-31T06:51:21.2371500Z test_decomp.py::TestDecompCPU::test_comprehensive_aminmax_cpu_uint8 PASSED [ 3%] 2023-03-31T06:51:21.2371845Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_bool PASSED [ 4%] 2023-03-31T06:51:21.2372189Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:21.2372521Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:21.2372867Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:21.2373207Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:21.2373550Z test_decomp.py::TestDecompCPU::test_comprehensive_angle_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:21.2373877Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2374215Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:21.2374553Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:21.2374874Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:21.2375208Z test_decomp.py::TestDecompCPU::test_comprehensive_any_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:21.2375550Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2375903Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_float16 PASSED [ 4%] 2023-03-31T06:51:21.2376234Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:21.2376578Z test_decomp.py::TestDecompCPU::test_comprehensive_arange_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:21.2376923Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2377259Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:21.2377601Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:21.2377942Z test_decomp.py::TestDecompCPU::test_comprehensive_argmax_cpu_uint8 PASSED [ 4%] 2023-03-31T06:51:21.2378290Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2378629Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:21.2378973Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:21.2379322Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:21.2379682Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:21.2380025Z test_decomp.py::TestDecompCPU::test_comprehensive_argmin_cpu_int64 PASSED [ 4%] 2023-03-31T06:51:21.2380376Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2380733Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:21.2381068Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:21.2381443Z test_decomp.py::TestDecompCPU::test_comprehensive_argsort_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:21.2381864Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_bfloat16 PASSED [ 4%] 2023-03-31T06:51:21.2382206Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_bool PASSED [ 4%] 2023-03-31T06:51:21.2382590Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_complex128 PASSED [ 4%] 2023-03-31T06:51:21.2382957Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_complex64 PASSED [ 4%] 2023-03-31T06:51:21.2383317Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:21.2383769Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_float64 PASSED [ 4%] 2023-03-31T06:51:21.2384122Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int16 PASSED [ 4%] 2023-03-31T06:51:21.2384473Z test_decomp.py::TestDecompCPU::test_comprehensive_argwhere_cpu_int8 PASSED [ 4%] 2023-03-31T06:51:21.2384816Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_float32 PASSED [ 4%] 2023-03-31T06:51:21.2385176Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int32 PASSED [ 4%] 2023-03-31T06:51:21.2385529Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:21.2385908Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:21.2386291Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_bool PASSED [ 5%] 2023-03-31T06:51:21.2386687Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex128 PASSED [ 5%] 2023-03-31T06:51:21.2387092Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_complex64 PASSED [ 5%] 2023-03-31T06:51:21.2387476Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_float32 PASSED [ 5%] 2023-03-31T06:51:21.2387868Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_partial_views_cpu_int64 PASSED [ 5%] 2023-03-31T06:51:21.2388312Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_bfloat16 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2388803Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex128 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2389288Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_complex32 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2389771Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float16 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2390249Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float32 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2390733Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_float64 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2391201Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int32 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2391675Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int64 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2392210Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_int8 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2392680Z test_decomp.py::TestDecompCPU::test_comprehensive_as_strided_scatter_cpu_uint8 SKIPPED (Expected: new_empty_strided is not comparable) [ 5%] 2023-03-31T06:51:21.2393090Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:21.2393425Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_bool PASSED [ 5%] 2023-03-31T06:51:21.2393808Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_complex64 PASSED [ 5%] 2023-03-31T06:51:21.2394183Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:21.2394513Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:21.2394879Z test_decomp.py::TestDecompCPU::test_comprehensive_asin_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:21.2395228Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:21.2395572Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_bool PASSED [ 5%] 2023-03-31T06:51:21.2395897Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:21.2396225Z test_decomp.py::TestDecompCPU::test_comprehensive_asinh_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:21.2396566Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_float32 PASSED [ 5%] 2023-03-31T06:51:21.2396901Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_float64 PASSED [ 5%] 2023-03-31T06:51:21.2397244Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_int32 PASSED [ 5%] 2023-03-31T06:51:21.2397584Z test_decomp.py::TestDecompCPU::test_comprehensive_atan2_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:21.2397920Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_bool PASSED [ 5%] 2023-03-31T06:51:21.2398254Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_complex128 PASSED [ 5%] 2023-03-31T06:51:21.2398599Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int16 PASSED [ 5%] 2023-03-31T06:51:21.2398932Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_int8 PASSED [ 5%] 2023-03-31T06:51:21.2399252Z test_decomp.py::TestDecompCPU::test_comprehensive_atan_cpu_uint8 PASSED [ 5%] 2023-03-31T06:51:21.2399593Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_bfloat16 PASSED [ 5%] 2023-03-31T06:51:21.2399948Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_complex128 PASSED [ 5%] 2023-03-31T06:51:21.2400309Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:21.2400643Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:21.2400981Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:21.2401323Z test_decomp.py::TestDecompCPU::test_comprehensive_atanh_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:21.2401660Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:21.2402023Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex128 PASSED [ 6%] 2023-03-31T06:51:21.2402386Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex32 PASSED [ 6%] 2023-03-31T06:51:21.2402748Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:21.2403279Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float16 PASSED [ 6%] 2023-03-31T06:51:21.2403638Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:21.2403991Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:21.2404333Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:21.2404686Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:21.2405111Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:21.2405466Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_1d_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:21.2405804Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_2d_cpu_float16 PASSED [ 6%] 2023-03-31T06:51:21.2406160Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:21.2406565Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex128 PASSED [ 6%] 2023-03-31T06:51:21.2406958Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:21.2407321Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float16 PASSED [ 6%] 2023-03-31T06:51:21.2407675Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:21.2408073Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_float64 PASSED [ 6%] 2023-03-31T06:51:21.2408415Z test_decomp.py::TestDecompCPU::test_comprehensive_atleast_3d_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:21.2408765Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:21.2409121Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_complex128 PASSED [ 6%] 2023-03-31T06:51:21.2409460Z test_decomp.py::TestDecompCPU::test_comprehensive_baddbmm_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:21.2409818Z test_decomp.py::TestDecompCPU::test_comprehensive_bernoulli_cpu_float32 PASSED [ 6%] 2023-03-31T06:51:21.2410174Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_bfloat16 PASSED [ 6%] 2023-03-31T06:51:21.2410522Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_bool PASSED [ 6%] 2023-03-31T06:51:21.2410862Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_complex64 PASSED [ 6%] 2023-03-31T06:51:21.2411219Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_float16 PASSED [ 6%] 2023-03-31T06:51:21.2411568Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:21.2411897Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:21.2412242Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int64 PASSED [ 6%] 2023-03-31T06:51:21.2412587Z test_decomp.py::TestDecompCPU::test_comprehensive_bfloat16_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:21.2412935Z test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int16 PASSED [ 6%] 2023-03-31T06:51:21.2413273Z test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int32 PASSED [ 6%] 2023-03-31T06:51:21.2413618Z test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_int8 PASSED [ 6%] 2023-03-31T06:51:21.2413965Z test_decomp.py::TestDecompCPU::test_comprehensive_bincount_cpu_uint8 PASSED [ 6%] 2023-03-31T06:51:21.2414305Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_and_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:21.2414668Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:21.2415039Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:21.2415407Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_left_shift_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:21.2415753Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:21.2416111Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:21.2416466Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_not_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:21.2416817Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:21.2417158Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:21.2417554Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:21.2417909Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:21.2418247Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_or_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:21.2418611Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:21.2418983Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_right_shift_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:21.2419394Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:21.2419755Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:21.2420105Z test_decomp.py::TestDecompCPU::test_comprehensive_bitwise_xor_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:21.2420492Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_complex64 PASSED [ 7%] 2023-03-31T06:51:21.2420839Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:21.2421194Z test_decomp.py::TestDecompCPU::test_comprehensive_block_diag_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:21.2421542Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_float32 PASSED [ 7%] 2023-03-31T06:51:21.2421882Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:21.2422210Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:21.2422547Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:21.2422886Z test_decomp.py::TestDecompCPU::test_comprehensive_bmm_cpu_uint8 PASSED [ 7%] 2023-03-31T06:51:21.2423216Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_float32 PASSED [ 7%] 2023-03-31T06:51:21.2423672Z test_decomp.py::TestDecompCPU::test_comprehensive_bool_cpu_int16 PASSED [ 7%] 2023-03-31T06:51:21.2424031Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_shapes_cpu_float32 PASSED [ 7%] 2023-03-31T06:51:21.2424406Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_bool PASSED [ 7%] 2023-03-31T06:51:21.2424772Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_complex128 PASSED [ 7%] 2023-03-31T06:51:21.2425151Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_tensors_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:21.2425527Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_complex128 PASSED [ 7%] 2023-03-31T06:51:21.2425882Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:21.2426247Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int32 PASSED [ 7%] 2023-03-31T06:51:21.2426604Z test_decomp.py::TestDecompCPU::test_comprehensive_broadcast_to_cpu_int8 PASSED [ 7%] 2023-03-31T06:51:21.2426965Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float16 PASSED [ 7%] 2023-03-31T06:51:21.2427304Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_float64 PASSED [ 7%] 2023-03-31T06:51:21.2427664Z test_decomp.py::TestDecompCPU::test_comprehensive_bucketize_cpu_int64 PASSED [ 7%] 2023-03-31T06:51:21.2428015Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_complex128 PASSED [ 8%] 2023-03-31T06:51:21.2428354Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_complex64 PASSED [ 8%] 2023-03-31T06:51:21.2428702Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:21.2429050Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:21.2429389Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:21.2429714Z test_decomp.py::TestDecompCPU::test_comprehensive_byte_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:21.2430072Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_complex64 PASSED [ 8%] 2023-03-31T06:51:21.2430491Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:21.2430847Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:21.2431208Z test_decomp.py::TestDecompCPU::test_comprehensive_cartesian_prod_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:21.2431557Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_bool PASSED [ 8%] 2023-03-31T06:51:21.2431900Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex32 PASSED [ 8%] 2023-03-31T06:51:21.2432267Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_complex64 PASSED [ 8%] 2023-03-31T06:51:21.2432641Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:21.2432977Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:21.2433334Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_float64 PASSED [ 8%] 2023-03-31T06:51:21.2433671Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:21.2433999Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:21.2434335Z test_decomp.py::TestDecompCPU::test_comprehensive_cat_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:21.2434667Z test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:21.2435017Z test_decomp.py::TestDecompCPU::test_comprehensive_cauchy_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:21.2435371Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:21.2435718Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_complex32 PASSED [ 8%] 2023-03-31T06:51:21.2436070Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float16 PASSED [ 8%] 2023-03-31T06:51:21.2436420Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_float64 PASSED [ 8%] 2023-03-31T06:51:21.2436773Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:21.2437107Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:21.2437452Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int64 PASSED [ 8%] 2023-03-31T06:51:21.2437799Z test_decomp.py::TestDecompCPU::test_comprehensive_cdouble_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:21.2438130Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_bfloat16 PASSED [ 8%] 2023-03-31T06:51:21.2438472Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int16 PASSED [ 8%] 2023-03-31T06:51:21.2438811Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int32 PASSED [ 8%] 2023-03-31T06:51:21.2439148Z test_decomp.py::TestDecompCPU::test_comprehensive_ceil_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:21.2439481Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_complex32 PASSED [ 8%] 2023-03-31T06:51:21.2439835Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float32 PASSED [ 8%] 2023-03-31T06:51:21.2440183Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_float64 PASSED [ 8%] 2023-03-31T06:51:21.2440518Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_int8 PASSED [ 8%] 2023-03-31T06:51:21.2440859Z test_decomp.py::TestDecompCPU::test_comprehensive_cfloat_cpu_uint8 PASSED [ 8%] 2023-03-31T06:51:21.2441198Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_bool PASSED [ 8%] 2023-03-31T06:51:21.2441555Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:21.2441900Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex32 PASSED [ 9%] 2023-03-31T06:51:21.2442251Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:21.2442603Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_float16 PASSED [ 9%] 2023-03-31T06:51:21.2442935Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int32 PASSED [ 9%] 2023-03-31T06:51:21.2443489Z test_decomp.py::TestDecompCPU::test_comprehensive_chalf_cpu_int64 PASSED [ 9%] 2023-03-31T06:51:21.2443829Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_bool PASSED [ 9%] 2023-03-31T06:51:21.2444178Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:21.2444518Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_complex32 PASSED [ 9%] 2023-03-31T06:51:21.2444867Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float16 PASSED [ 9%] 2023-03-31T06:51:21.2445249Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:21.2445613Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int16 PASSED [ 9%] 2023-03-31T06:51:21.2445950Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int32 PASSED [ 9%] 2023-03-31T06:51:21.2446328Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_int64 PASSED [ 9%] 2023-03-31T06:51:21.2446665Z test_decomp.py::TestDecompCPU::test_comprehensive_char_cpu_uint8 PASSED [ 9%] 2023-03-31T06:51:21.2447002Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:21.2447366Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:21.2447720Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:21.2448076Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_inverse_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:21.2448453Z test_decomp.py::TestDecompCPU::test_comprehensive_cholesky_solve_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:21.2448816Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:51:21.2449166Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex128 PASSED [ 9%] 2023-03-31T06:51:21.2449507Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_complex64 PASSED [ 9%] 2023-03-31T06:51:21.2449860Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:21.2450204Z test_decomp.py::TestDecompCPU::test_comprehensive_chunk_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:21.2450553Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_bfloat16 PASSED [ 9%] 2023-03-31T06:51:21.2450890Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:21.2451232Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_float64 PASSED [ 9%] 2023-03-31T06:51:21.2451577Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int16 PASSED [ 9%] 2023-03-31T06:51:21.2451908Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_cpu_int32 PASSED [ 9%] 2023-03-31T06:51:21.2452254Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_bool PASSED [ 9%] 2023-03-31T06:51:21.2452606Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_float16 PASSED [ 9%] 2023-03-31T06:51:21.2452960Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int64 PASSED [ 9%] 2023-03-31T06:51:21.2453298Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_int8 PASSED [ 9%] 2023-03-31T06:51:21.2453643Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_max_cpu_uint8 PASSED [ 9%] 2023-03-31T06:51:21.2453995Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_bool PASSED [ 9%] 2023-03-31T06:51:21.2454330Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float16 PASSED [ 9%] 2023-03-31T06:51:21.2454686Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_float32 PASSED [ 9%] 2023-03-31T06:51:21.2455036Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:21.2455385Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:21.2455719Z test_decomp.py::TestDecompCPU::test_comprehensive_clamp_min_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:21.2456099Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_bool PASSED [ 10%] 2023-03-31T06:51:21.2456452Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex128 PASSED [ 10%] 2023-03-31T06:51:21.2456792Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_complex64 PASSED [ 10%] 2023-03-31T06:51:21.2457134Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:21.2457473Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:21.2457842Z test_decomp.py::TestDecompCPU::test_comprehensive_clone_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:21.2458208Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:21.2458574Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_complex32 PASSED [ 10%] 2023-03-31T06:51:21.2458983Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:21.2459332Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:21.2459694Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:21.2460046Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:21.2460404Z test_decomp.py::TestDecompCPU::test_comprehensive_column_stack_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:21.2460754Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_bfloat16 PASSED [ 10%] 2023-03-31T06:51:21.2461118Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_bool PASSED [ 10%] 2023-03-31T06:51:21.2461489Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_complex128 PASSED [ 10%] 2023-03-31T06:51:21.2461849Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_complex64 PASSED [ 10%] 2023-03-31T06:51:21.2462218Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:21.2462582Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:21.2462943Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:21.2463288Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int64 PASSED [ 10%] 2023-03-31T06:51:21.2463749Z test_decomp.py::TestDecompCPU::test_comprehensive_combinations_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:21.2464112Z test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:21.2464451Z test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:21.2464809Z test_decomp.py::TestDecompCPU::test_comprehensive_complex_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:21.2465164Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_complex32 PASSED [ 10%] 2023-03-31T06:51:21.2465514Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int16 PASSED [ 10%] 2023-03-31T06:51:21.2465843Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_int8 PASSED [ 10%] 2023-03-31T06:51:21.2466181Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:21.2466532Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_bool PASSED [ 10%] 2023-03-31T06:51:21.2466888Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_complex128 PASSED [ 10%] 2023-03-31T06:51:21.2467258Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float16 PASSED [ 10%] 2023-03-31T06:51:21.2467621Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float32 PASSED [ 10%] 2023-03-31T06:51:21.2467987Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_float64 PASSED [ 10%] 2023-03-31T06:51:21.2468336Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_int32 PASSED [ 10%] 2023-03-31T06:51:21.2468696Z test_decomp.py::TestDecompCPU::test_comprehensive_conj_physical_cpu_uint8 PASSED [ 10%] 2023-03-31T06:51:21.2469103Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_bool PASSED [ 11%] 2023-03-31T06:51:21.2469480Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:21.2469840Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float16 PASSED [ 11%] 2023-03-31T06:51:21.2470212Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:21.2470604Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:21.2470979Z test_decomp.py::TestDecompCPU::test_comprehensive_constant_pad_nd_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:21.2471343Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:21.2471736Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex32 PASSED [ 11%] 2023-03-31T06:51:21.2472109Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:21.2472457Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:21.2472813Z test_decomp.py::TestDecompCPU::test_comprehensive_contiguous_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:21.2473177Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:21.2473519Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_bool PASSED [ 11%] 2023-03-31T06:51:21.2473872Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:21.2474226Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int32 PASSED [ 11%] 2023-03-31T06:51:21.2474575Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:21.2474911Z test_decomp.py::TestDecompCPU::test_comprehensive_copysign_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:21.2475268Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:21.2475615Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int32 PASSED [ 11%] 2023-03-31T06:51:21.2475948Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:21.2476293Z test_decomp.py::TestDecompCPU::test_comprehensive_corrcoef_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:21.2476635Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:21.2476981Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_bool PASSED [ 11%] 2023-03-31T06:51:21.2477316Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:21.2477658Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_int8 PASSED [ 11%] 2023-03-31T06:51:21.2477998Z test_decomp.py::TestDecompCPU::test_comprehensive_cos_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:21.2478330Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:21.2478680Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_float32 PASSED [ 11%] 2023-03-31T06:51:21.2479021Z test_decomp.py::TestDecompCPU::test_comprehensive_cosh_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:21.2479377Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_complex128 PASSED [ 11%] 2023-03-31T06:51:21.2479732Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float16 PASSED [ 11%] 2023-03-31T06:51:21.2480096Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:21.2480462Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_int64 PASSED [ 11%] 2023-03-31T06:51:21.2480809Z test_decomp.py::TestDecompCPU::test_comprehensive_count_nonzero_cpu_uint8 PASSED [ 11%] 2023-03-31T06:51:21.2481166Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_complex64 PASSED [ 11%] 2023-03-31T06:51:21.2481510Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_float64 PASSED [ 11%] 2023-03-31T06:51:21.2481881Z test_decomp.py::TestDecompCPU::test_comprehensive_cov_cpu_int16 PASSED [ 11%] 2023-03-31T06:51:21.2482216Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_bfloat16 PASSED [ 11%] 2023-03-31T06:51:21.2482570Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:21.2482920Z test_decomp.py::TestDecompCPU::test_comprehensive_cross_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:21.2483500Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_bool PASSED [ 12%] 2023-03-31T06:51:21.2483888Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:21.2484235Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:21.2484579Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:21.2484947Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:21.2485291Z test_decomp.py::TestDecompCPU::test_comprehensive_cummax_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:21.2485634Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_bool PASSED [ 12%] 2023-03-31T06:51:21.2485965Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:21.2486312Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:21.2486652Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:21.2486993Z test_decomp.py::TestDecompCPU::test_comprehensive_cummin_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:21.2487339Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:21.2487699Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_complex64 PASSED [ 12%] 2023-03-31T06:51:21.2488059Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:21.2488401Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int16 PASSED [ 12%] 2023-03-31T06:51:21.2488748Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:21.2489093Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int64 PASSED [ 12%] 2023-03-31T06:51:21.2489437Z test_decomp.py::TestDecompCPU::test_comprehensive_cumprod_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:21.2489776Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:21.2490137Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_complex64 PASSED [ 12%] 2023-03-31T06:51:21.2490491Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:21.2490835Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:21.2491164Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_int8 PASSED [ 12%] 2023-03-31T06:51:21.2491506Z test_decomp.py::TestDecompCPU::test_comprehensive_cumsum_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:21.2491871Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_float32 PASSED [ 12%] 2023-03-31T06:51:21.2492247Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_float64 PASSED [ 12%] 2023-03-31T06:51:21.2492627Z test_decomp.py::TestDecompCPU::test_comprehensive_cumulative_trapezoid_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:21.2492992Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_float16 PASSED [ 12%] 2023-03-31T06:51:21.2493343Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_int32 PASSED [ 12%] 2023-03-31T06:51:21.2493678Z test_decomp.py::TestDecompCPU::test_comprehensive_deg2rad_cpu_uint8 PASSED [ 12%] 2023-03-31T06:51:21.2494076Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_bfloat16 SKIPPED (diag in torch.bfloat16 not supported) [ 12%] 2023-03-31T06:51:21.2494549Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_bool SKIPPED (diag in torch.bool not supported) [ 12%] 2023-03-31T06:51:21.2494977Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_complex128 SKIPPED (diag in torch.complex128 not supported) [ 12%] 2023-03-31T06:51:21.2495416Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float16 SKIPPED (diag in torch.float16 not supported) [ 12%] 2023-03-31T06:51:21.2495843Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_float32 SKIPPED (diag in torch.float32 not supported) [ 12%] 2023-03-31T06:51:21.2496293Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_cpu_int8 SKIPPED (diag in torch.int8 not supported) [ 12%] 2023-03-31T06:51:21.2496692Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_bfloat16 PASSED [ 12%] 2023-03-31T06:51:21.2497052Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float16 PASSED [ 13%] 2023-03-31T06:51:21.2497439Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:21.2497802Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:21.2498139Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:21.2498492Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_int8 PASSED [ 13%] 2023-03-31T06:51:21.2498847Z test_decomp.py::TestDecompCPU::test_comprehensive_diag_embed_cpu_uint8 PASSED [ 13%] 2023-03-31T06:51:21.2499186Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:21.2499544Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_bool PASSED [ 13%] 2023-03-31T06:51:21.2499899Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_complex64 PASSED [ 13%] 2023-03-31T06:51:21.2500260Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:21.2500600Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_float64 PASSED [ 13%] 2023-03-31T06:51:21.2500949Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:21.2501302Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:21.2501636Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_int8 PASSED [ 13%] 2023-03-31T06:51:21.2501987Z test_decomp.py::TestDecompCPU::test_comprehensive_diagflat_cpu_uint8 PASSED [ 13%] 2023-03-31T06:51:21.2502346Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:21.2502720Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_complex32 PASSED [ 13%] 2023-03-31T06:51:21.2503078Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:21.2503537Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_float64 PASSED [ 13%] 2023-03-31T06:51:21.2503906Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:21.2504261Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:21.2504629Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_copy_cpu_int8 PASSED [ 13%] 2023-03-31T06:51:21.2504991Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:21.2505350Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_bool PASSED [ 13%] 2023-03-31T06:51:21.2505695Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex32 PASSED [ 13%] 2023-03-31T06:51:21.2506054Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_complex64 PASSED [ 13%] 2023-03-31T06:51:21.2506413Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:21.2506768Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int16 PASSED [ 13%] 2023-03-31T06:51:21.2507107Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:21.2507502Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_int64 PASSED [ 13%] 2023-03-31T06:51:21.2507852Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_cpu_uint8 PASSED [ 13%] 2023-03-31T06:51:21.2508201Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_bfloat16 PASSED [ 13%] 2023-03-31T06:51:21.2508581Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_complex128 PASSED [ 13%] 2023-03-31T06:51:21.2509006Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_complex64 PASSED [ 13%] 2023-03-31T06:51:21.2509411Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float32 PASSED [ 13%] 2023-03-31T06:51:21.2509769Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_float64 PASSED [ 13%] 2023-03-31T06:51:21.2510169Z test_decomp.py::TestDecompCPU::test_comprehensive_diagonal_scatter_cpu_int32 PASSED [ 13%] 2023-03-31T06:51:21.2510529Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_bool PASSED [ 13%] 2023-03-31T06:51:21.2510867Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_complex64 PASSED [ 14%] 2023-03-31T06:51:21.2511212Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:21.2511557Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:21.2511898Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2512224Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:21.2512561Z test_decomp.py::TestDecompCPU::test_comprehensive_diff_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2512907Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2513240Z test_decomp.py::TestDecompCPU::test_comprehensive_digamma_cpu_int32 PASSED [ 14%] 2023-03-31T06:51:21.2513589Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:21.2513943Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_complex64 PASSED [ 14%] 2023-03-31T06:51:21.2514286Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:21.2514616Z test_decomp.py::TestDecompCPU::test_comprehensive_dist_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:21.2514977Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:21.2515355Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2515717Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:21.2516083Z test_decomp.py::TestDecompCPU::test_comprehensive_div_floor_rounding_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2516470Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_complex128 PASSED [ 14%] 2023-03-31T06:51:21.2516855Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:21.2517218Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:21.2517593Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_float64 PASSED [ 14%] 2023-03-31T06:51:21.2517966Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2518341Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:21.2518704Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_int8 PASSED [ 14%] 2023-03-31T06:51:21.2519076Z test_decomp.py::TestDecompCPU::test_comprehensive_div_no_rounding_mode_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2519447Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:21.2519813Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float16 PASSED [ 14%] 2023-03-31T06:51:21.2520225Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_float32 PASSED [ 14%] 2023-03-31T06:51:21.2520599Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2520968Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_int32 PASSED [ 14%] 2023-03-31T06:51:21.2521320Z test_decomp.py::TestDecompCPU::test_comprehensive_div_trunc_rounding_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2521705Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_complex64 PASSED [ 14%] 2023-03-31T06:51:21.2522076Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2522404Z test_decomp.py::TestDecompCPU::test_comprehensive_dot_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2522783Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_bool PASSED [ 14%] 2023-03-31T06:51:21.2523317Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int16 PASSED [ 14%] 2023-03-31T06:51:21.2523667Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_int64 PASSED [ 14%] 2023-03-31T06:51:21.2523998Z test_decomp.py::TestDecompCPU::test_comprehensive_double_cpu_uint8 PASSED [ 14%] 2023-03-31T06:51:21.2524348Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_bfloat16 PASSED [ 14%] 2023-03-31T06:51:21.2524697Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_bool PASSED [ 14%] 2023-03-31T06:51:21.2525037Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:21.2525399Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_complex64 PASSED [ 15%] 2023-03-31T06:51:21.2525749Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:21.2526092Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:21.2526428Z test_decomp.py::TestDecompCPU::test_comprehensive_dsplit_cpu_uint8 PASSED [ 15%] 2023-03-31T06:51:21.2526782Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_complex32 PASSED [ 15%] 2023-03-31T06:51:21.2527134Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float16 PASSED [ 15%] 2023-03-31T06:51:21.2527470Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:21.2527813Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:21.2528162Z test_decomp.py::TestDecompCPU::test_comprehensive_dstack_cpu_uint8 PASSED [ 15%] 2023-03-31T06:51:21.2528515Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:21.2528855Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int64 PASSED [ 15%] 2023-03-31T06:51:21.2529197Z test_decomp.py::TestDecompCPU::test_comprehensive_einsum_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:21.2529602Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_complex64 SKIPPED (empty in torch.complex64 not supported) [ 15%] 2023-03-31T06:51:21.2530036Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float16 SKIPPED (empty in torch.float16 not supported) [ 15%] 2023-03-31T06:51:21.2530469Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_float64 SKIPPED (empty in torch.float64 not supported) [ 15%] 2023-03-31T06:51:21.2530894Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int64 SKIPPED (empty in torch.int64 not supported) [ 15%] 2023-03-31T06:51:21.2531319Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_cpu_int8 SKIPPED (empty in torch.int8 not supported) [ 15%] 2023-03-31T06:51:21.2531756Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_complex64 SKIPPED (empty_like in torch.complex64 not supported) [ 15%] 2023-03-31T06:51:21.2532223Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float16 SKIPPED (empty_like in torch.float16 not supported) [ 15%] 2023-03-31T06:51:21.2532745Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float32 SKIPPED (empty_like in torch.float32 not supported) [ 15%] 2023-03-31T06:51:21.2533195Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_float64 SKIPPED (empty_like in torch.float64 not supported) [ 15%] 2023-03-31T06:51:21.2533622Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_like_cpu_int64 SKIPPED (empty_like in torch.int64 not supported) [ 15%] 2023-03-31T06:51:21.2534021Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_bool PASSED [ 15%] 2023-03-31T06:51:21.2534431Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:21.2534847Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float16 PASSED [ 15%] 2023-03-31T06:51:21.2535203Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_float64 PASSED [ 15%] 2023-03-31T06:51:21.2535607Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:21.2535974Z test_decomp.py::TestDecompCPU::test_comprehensive_empty_permuted_cpu_uint8 PASSED [ 15%] 2023-03-31T06:51:21.2536328Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:21.2536663Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex128 PASSED [ 15%] 2023-03-31T06:51:21.2537012Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_complex32 PASSED [ 15%] 2023-03-31T06:51:21.2537357Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_float32 PASSED [ 15%] 2023-03-31T06:51:21.2537683Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int32 PASSED [ 15%] 2023-03-31T06:51:21.2538021Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_int8 PASSED [ 15%] 2023-03-31T06:51:21.2538356Z test_decomp.py::TestDecompCPU::test_comprehensive_eq_cpu_uint8 PASSED [ 15%] 2023-03-31T06:51:21.2538705Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_bfloat16 PASSED [ 15%] 2023-03-31T06:51:21.2539046Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_complex64 PASSED [ 15%] 2023-03-31T06:51:21.2539397Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_float64 PASSED [ 16%] 2023-03-31T06:51:21.2539740Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:21.2540064Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:21.2540400Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:21.2540738Z test_decomp.py::TestDecompCPU::test_comprehensive_equal_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:21.2541084Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:21.2541410Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_bool PASSED [ 16%] 2023-03-31T06:51:21.2541749Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:21.2542085Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:21.2542406Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:21.2542734Z test_decomp.py::TestDecompCPU::test_comprehensive_erf_cpu_int8 PASSED [ 16%] 2023-03-31T06:51:21.2543075Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:21.2543516Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_bool PASSED [ 16%] 2023-03-31T06:51:21.2543849Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_float64 PASSED [ 16%] 2023-03-31T06:51:21.2544191Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:21.2544529Z test_decomp.py::TestDecompCPU::test_comprehensive_erfc_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:21.2544861Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:21.2545202Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int16 PASSED [ 16%] 2023-03-31T06:51:21.2545577Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:21.2545916Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_int8 PASSED [ 16%] 2023-03-31T06:51:21.2546240Z test_decomp.py::TestDecompCPU::test_comprehensive_exp2_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:21.2546581Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:21.2546920Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_bool PASSED [ 16%] 2023-03-31T06:51:21.2547301Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_complex128 PASSED [ 16%] 2023-03-31T06:51:21.2547681Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:21.2548016Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:21.2548402Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:21.2548726Z test_decomp.py::TestDecompCPU::test_comprehensive_exp_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:21.2549070Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_bool PASSED [ 16%] 2023-03-31T06:51:21.2549424Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float16 PASSED [ 16%] 2023-03-31T06:51:21.2549766Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_float32 PASSED [ 16%] 2023-03-31T06:51:21.2550122Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_int32 PASSED [ 16%] 2023-03-31T06:51:21.2550472Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_as_cpu_uint8 PASSED [ 16%] 2023-03-31T06:51:21.2550826Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:21.2551165Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_float16 PASSED [ 16%] 2023-03-31T06:51:21.2551515Z test_decomp.py::TestDecompCPU::test_comprehensive_expand_cpu_int64 PASSED [ 16%] 2023-03-31T06:51:21.2551866Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_bfloat16 PASSED [ 16%] 2023-03-31T06:51:21.2552197Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_bool PASSED [ 16%] 2023-03-31T06:51:21.2552551Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_complex128 PASSED [ 16%] 2023-03-31T06:51:21.2552909Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:21.2553256Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2553589Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:21.2553929Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:21.2554266Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:21.2554594Z test_decomp.py::TestDecompCPU::test_comprehensive_expm1_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:21.2554945Z test_decomp.py::TestDecompCPU::test_comprehensive_exponential_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2555307Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_complex128 PASSED [ 17%] 2023-03-31T06:51:21.2555658Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:21.2555988Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:21.2556323Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2556661Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:21.2556984Z test_decomp.py::TestDecompCPU::test_comprehensive_eye_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:21.2557322Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_bool PASSED [ 17%] 2023-03-31T06:51:21.2557674Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_complex128 PASSED [ 17%] 2023-03-31T06:51:21.2558037Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_complex64 PASSED [ 17%] 2023-03-31T06:51:21.2558413Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:21.2558763Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:21.2559105Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int64 PASSED [ 17%] 2023-03-31T06:51:21.2559433Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:21.2559805Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft2_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:21.2560175Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:21.2560525Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2560858Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:21.2561233Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:21.2561577Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fft_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:21.2561913Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_float32 PASSED [ 17%] 2023-03-31T06:51:21.2562263Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2562613Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int16 PASSED [ 17%] 2023-03-31T06:51:21.2562961Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:21.2563475Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftn_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:21.2563839Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_complex32 PASSED [ 17%] 2023-03-31T06:51:21.2564213Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_float16 PASSED [ 17%] 2023-03-31T06:51:21.2564564Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_int8 PASSED [ 17%] 2023-03-31T06:51:21.2564924Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_fftshift_cpu_uint8 PASSED [ 17%] 2023-03-31T06:51:21.2565277Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_bool PASSED [ 17%] 2023-03-31T06:51:21.2565630Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_float64 PASSED [ 17%] 2023-03-31T06:51:21.2565970Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int32 PASSED [ 17%] 2023-03-31T06:51:21.2566318Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:21.2566665Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft2_cpu_uint8 PASSED [ 18%] 2023-03-31T06:51:21.2567005Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:21.2567370Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:21.2567717Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:21.2568048Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:21.2568391Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfft_cpu_uint8 PASSED [ 18%] 2023-03-31T06:51:21.2568742Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:21.2569092Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:21.2569449Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:21.2569794Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:21.2570140Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:21.2570468Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_hfftn_cpu_uint8 PASSED [ 18%] 2023-03-31T06:51:21.2570816Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_bool PASSED [ 18%] 2023-03-31T06:51:21.2571241Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:21.2571587Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:21.2571938Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:21.2572286Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:21.2572667Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:21.2573030Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft2_cpu_uint8 PASSED [ 18%] 2023-03-31T06:51:21.2573376Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_bool PASSED [ 18%] 2023-03-31T06:51:21.2573726Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:21.2574107Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_float32 PASSED [ 18%] 2023-03-31T06:51:21.2574458Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:21.2574809Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:21.2575153Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:21.2575485Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifft_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:21.2575840Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:21.2576199Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:21.2576537Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:21.2576880Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftn_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:21.2577245Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex128 PASSED [ 18%] 2023-03-31T06:51:21.2577617Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_complex64 PASSED [ 18%] 2023-03-31T06:51:21.2577970Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float16 PASSED [ 18%] 2023-03-31T06:51:21.2578332Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_float64 PASSED [ 18%] 2023-03-31T06:51:21.2578693Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int16 PASSED [ 18%] 2023-03-31T06:51:21.2579049Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int32 PASSED [ 18%] 2023-03-31T06:51:21.2579394Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int64 PASSED [ 18%] 2023-03-31T06:51:21.2579751Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ifftshift_cpu_int8 PASSED [ 18%] 2023-03-31T06:51:21.2580108Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:21.2580445Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:21.2580794Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft2_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:21.2581144Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:21.2581497Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:21.2581833Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:21.2582179Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:21.2582533Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfft_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:21.2582860Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_bool PASSED [ 19%] 2023-03-31T06:51:21.2583217Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:21.2583674Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:21.2584116Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:21.2584455Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_ihfftn_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:21.2584801Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_bool PASSED [ 19%] 2023-03-31T06:51:21.2585161Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_complex128 PASSED [ 19%] 2023-03-31T06:51:21.2585534Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:21.2585914Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:21.2586257Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:21.2586629Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft2_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:21.2586963Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_bool PASSED [ 19%] 2023-03-31T06:51:21.2587316Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_complex128 PASSED [ 19%] 2023-03-31T06:51:21.2587678Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_complex64 PASSED [ 19%] 2023-03-31T06:51:21.2588021Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfft_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:21.2588375Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_complex128 PASSED [ 19%] 2023-03-31T06:51:21.2588738Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_complex64 PASSED [ 19%] 2023-03-31T06:51:21.2589094Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:21.2589432Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_irfftn_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:21.2589785Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:21.2590135Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:21.2590466Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft2_cpu_uint8 PASSED [ 19%] 2023-03-31T06:51:21.2590814Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:21.2591160Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_float64 PASSED [ 19%] 2023-03-31T06:51:21.2591504Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int16 PASSED [ 19%] 2023-03-31T06:51:21.2591837Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:21.2592181Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int64 PASSED [ 19%] 2023-03-31T06:51:21.2592526Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfft_cpu_int8 PASSED [ 19%] 2023-03-31T06:51:21.2592863Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_float32 PASSED [ 19%] 2023-03-31T06:51:21.2593215Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int32 PASSED [ 19%] 2023-03-31T06:51:21.2593560Z test_decomp.py::TestDecompCPU::test_comprehensive_fft_rfftn_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:21.2593906Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_bool PASSED [ 20%] 2023-03-31T06:51:21.2594239Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_complex64 PASSED [ 20%] 2023-03-31T06:51:21.2594582Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float16 PASSED [ 20%] 2023-03-31T06:51:21.2594929Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:21.2595260Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:21.2595598Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:21.2595940Z test_decomp.py::TestDecompCPU::test_comprehensive_fill_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:21.2596286Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_bfloat16 PASSED [ 20%] 2023-03-31T06:51:21.2596663Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_complex64 PASSED [ 20%] 2023-03-31T06:51:21.2597020Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float16 PASSED [ 20%] 2023-03-31T06:51:21.2597366Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:21.2597699Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:21.2598092Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:21.2598463Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:21.2598807Z test_decomp.py::TestDecompCPU::test_comprehensive_flatten_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:21.2599139Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:21.2599521Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_complex64 PASSED [ 20%] 2023-03-31T06:51:21.2599874Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float16 PASSED [ 20%] 2023-03-31T06:51:21.2600217Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:21.2600547Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:21.2600885Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:21.2601227Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:21.2601552Z test_decomp.py::TestDecompCPU::test_comprehensive_flip_cpu_int8 PASSED [ 20%] 2023-03-31T06:51:21.2601889Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_bool PASSED [ 20%] 2023-03-31T06:51:21.2602247Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:21.2602592Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_float64 PASSED [ 20%] 2023-03-31T06:51:21.2602945Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:21.2603479Z test_decomp.py::TestDecompCPU::test_comprehensive_fliplr_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:21.2603828Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_bfloat16 PASSED [ 20%] 2023-03-31T06:51:21.2604172Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:21.2604526Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_float32 PASSED [ 20%] 2023-03-31T06:51:21.2604878Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:21.2605220Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:21.2605550Z test_decomp.py::TestDecompCPU::test_comprehensive_flipud_cpu_uint8 PASSED [ 20%] 2023-03-31T06:51:21.2605901Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_complex128 PASSED [ 20%] 2023-03-31T06:51:21.2606247Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int16 PASSED [ 20%] 2023-03-31T06:51:21.2606576Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int32 PASSED [ 20%] 2023-03-31T06:51:21.2606906Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_int64 PASSED [ 20%] 2023-03-31T06:51:21.2607244Z test_decomp.py::TestDecompCPU::test_comprehensive_float_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:21.2607579Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_bool PASSED [ 21%] 2023-03-31T06:51:21.2607938Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2608298Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2608655Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:21.2608993Z test_decomp.py::TestDecompCPU::test_comprehensive_float_power_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:21.2609395Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2609742Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:21.2610082Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_cpu_int8 PASSED [ 21%] 2023-03-31T06:51:21.2610418Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2610776Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:21.2611168Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:21.2611544Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:21.2611898Z test_decomp.py::TestDecompCPU::test_comprehensive_floor_divide_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:21.2612283Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2612629Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2612956Z test_decomp.py::TestDecompCPU::test_comprehensive_fmax_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:21.2613291Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2613632Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:21.2613959Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:21.2614298Z test_decomp.py::TestDecompCPU::test_comprehensive_fmin_cpu_uint8 PASSED [ 21%] 2023-03-31T06:51:21.2614649Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:21.2614992Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2615320Z test_decomp.py::TestDecompCPU::test_comprehensive_fmod_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:21.2615658Z test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:21.2615999Z test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2616327Z test_decomp.py::TestDecompCPU::test_comprehensive_frac_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:21.2616669Z test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:21.2617014Z test_decomp.py::TestDecompCPU::test_comprehensive_frexp_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2617360Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_bfloat16 PASSED [ 21%] 2023-03-31T06:51:21.2617702Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex128 PASSED [ 21%] 2023-03-31T06:51:21.2618053Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_complex64 PASSED [ 21%] 2023-03-31T06:51:21.2618398Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float16 PASSED [ 21%] 2023-03-31T06:51:21.2618728Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float32 PASSED [ 21%] 2023-03-31T06:51:21.2619067Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_float64 PASSED [ 21%] 2023-03-31T06:51:21.2619409Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int16 PASSED [ 21%] 2023-03-31T06:51:21.2619748Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int32 PASSED [ 21%] 2023-03-31T06:51:21.2620070Z test_decomp.py::TestDecompCPU::test_comprehensive_full_cpu_int64 PASSED [ 21%] 2023-03-31T06:51:21.2620414Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2620775Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_complex128 PASSED [ 22%] 2023-03-31T06:51:21.2621129Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_complex64 PASSED [ 22%] 2023-03-31T06:51:21.2621484Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2621835Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int16 PASSED [ 22%] 2023-03-31T06:51:21.2622213Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_int64 PASSED [ 22%] 2023-03-31T06:51:21.2622554Z test_decomp.py::TestDecompCPU::test_comprehensive_full_like_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2622899Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float16 PASSED [ 22%] 2023-03-31T06:51:21.2623243Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2623678Z test_decomp.py::TestDecompCPU::test_comprehensive_gather_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:21.2624049Z test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int16 PASSED [ 22%] 2023-03-31T06:51:21.2624414Z test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:21.2624748Z test_decomp.py::TestDecompCPU::test_comprehensive_gcd_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2625122Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2625457Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float16 PASSED [ 22%] 2023-03-31T06:51:21.2625794Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2626114Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_int64 PASSED [ 22%] 2023-03-31T06:51:21.2626441Z test_decomp.py::TestDecompCPU::test_comprehensive_ge_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2626778Z test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2627124Z test_decomp.py::TestDecompCPU::test_comprehensive_geqrf_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:21.2627465Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2627818Z test_decomp.py::TestDecompCPU::test_comprehensive_gradient_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:21.2628178Z test_decomp.py::TestDecompCPU::test_comprehensive_grid_sampler_2d_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:21.2628520Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2628855Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2629193Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:21.2629526Z test_decomp.py::TestDecompCPU::test_comprehensive_gt_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2629854Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2630196Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_bool PASSED [ 22%] 2023-03-31T06:51:21.2630541Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_complex64 PASSED [ 22%] 2023-03-31T06:51:21.2630875Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:21.2631218Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int32 PASSED [ 22%] 2023-03-31T06:51:21.2631556Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_int8 PASSED [ 22%] 2023-03-31T06:51:21.2631894Z test_decomp.py::TestDecompCPU::test_comprehensive_half_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2632231Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2632591Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float32 PASSED [ 22%] 2023-03-31T06:51:21.2632948Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_float64 PASSED [ 22%] 2023-03-31T06:51:21.2633288Z test_decomp.py::TestDecompCPU::test_comprehensive_heaviside_cpu_uint8 PASSED [ 22%] 2023-03-31T06:51:21.2633640Z test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_bfloat16 PASSED [ 22%] 2023-03-31T06:51:21.2633990Z test_decomp.py::TestDecompCPU::test_comprehensive_histc_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2634338Z test_decomp.py::TestDecompCPU::test_comprehensive_histogram_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2634686Z test_decomp.py::TestDecompCPU::test_comprehensive_histogramdd_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2635082Z test_decomp.py::TestDecompCPU::test_comprehensive_histogramdd_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2635444Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_complex32 PASSED [ 23%] 2023-03-31T06:51:21.2635783Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2636132Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2636507Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:21.2636875Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:21.2637206Z test_decomp.py::TestDecompCPU::test_comprehensive_hsplit_cpu_uint8 PASSED [ 23%] 2023-03-31T06:51:21.2637552Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:21.2637938Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_bool PASSED [ 23%] 2023-03-31T06:51:21.2638272Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float16 PASSED [ 23%] 2023-03-31T06:51:21.2638614Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2638962Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:21.2639302Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_int64 PASSED [ 23%] 2023-03-31T06:51:21.2639625Z test_decomp.py::TestDecompCPU::test_comprehensive_hstack_cpu_uint8 PASSED [ 23%] 2023-03-31T06:51:21.2639967Z test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2640316Z test_decomp.py::TestDecompCPU::test_comprehensive_hypot_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2640645Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:21.2640980Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_bool PASSED [ 23%] 2023-03-31T06:51:21.2641313Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2641645Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2641965Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int16 PASSED [ 23%] 2023-03-31T06:51:21.2642292Z test_decomp.py::TestDecompCPU::test_comprehensive_i0_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:21.2642633Z test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2642969Z test_decomp.py::TestDecompCPU::test_comprehensive_igamma_cpu_float64 PASSED [ 23%] 2023-03-31T06:51:21.2643502Z test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:21.2643861Z test_decomp.py::TestDecompCPU::test_comprehensive_igammac_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2644215Z test_decomp.py::TestDecompCPU::test_comprehensive_imag_cpu_complex128 PASSED [ 23%] 2023-03-31T06:51:21.2644558Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_bfloat16 PASSED [ 23%] 2023-03-31T06:51:21.2644908Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_bool PASSED [ 23%] 2023-03-31T06:51:21.2645262Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex128 PASSED [ 23%] 2023-03-31T06:51:21.2645611Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex32 PASSED [ 23%] 2023-03-31T06:51:21.2645970Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_complex64 PASSED [ 23%] 2023-03-31T06:51:21.2646331Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_float32 PASSED [ 23%] 2023-03-31T06:51:21.2646684Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int32 PASSED [ 23%] 2023-03-31T06:51:21.2647019Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_int8 PASSED [ 23%] 2023-03-31T06:51:21.2647370Z test_decomp.py::TestDecompCPU::test_comprehensive_index_add_cpu_uint8 PASSED [ 24%] 2023-03-31T06:51:21.2647790Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:21.2648142Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:21.2648499Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int16 PASSED [ 24%] 2023-03-31T06:51:21.2648848Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:21.2649199Z test_decomp.py::TestDecompCPU::test_comprehensive_index_copy_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:21.2649570Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:21.2649972Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:21.2650333Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:21.2650714Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int16 PASSED [ 24%] 2023-03-31T06:51:21.2651067Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int32 PASSED [ 24%] 2023-03-31T06:51:21.2651415Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:21.2651766Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:21.2652102Z test_decomp.py::TestDecompCPU::test_comprehensive_index_fill_cpu_uint8 PASSED [ 24%] 2023-03-31T06:51:21.2652453Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_bool PASSED [ 24%] 2023-03-31T06:51:21.2652812Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:21.2653177Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:21.2653523Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float32 PASSED [ 24%] 2023-03-31T06:51:21.2653883Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:21.2654236Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int16 PASSED [ 24%] 2023-03-31T06:51:21.2654570Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int32 PASSED [ 24%] 2023-03-31T06:51:21.2654912Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:21.2655260Z test_decomp.py::TestDecompCPU::test_comprehensive_index_put_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:21.2655614Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:21.2655962Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int16 PASSED [ 24%] 2023-03-31T06:51:21.2656320Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int32 PASSED [ 24%] 2023-03-31T06:51:21.2656678Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:21.2657024Z test_decomp.py::TestDecompCPU::test_comprehensive_index_reduce_cpu_uint8 PASSED [ 24%] 2023-03-31T06:51:21.2657395Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:21.2657763Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float16 PASSED [ 24%] 2023-03-31T06:51:21.2658123Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float32 PASSED [ 24%] 2023-03-31T06:51:21.2658470Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:21.2658832Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int64 PASSED [ 24%] 2023-03-31T06:51:21.2659188Z test_decomp.py::TestDecompCPU::test_comprehensive_index_select_cpu_int8 PASSED [ 24%] 2023-03-31T06:51:21.2659530Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_bfloat16 PASSED [ 24%] 2023-03-31T06:51:21.2659882Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_complex128 PASSED [ 24%] 2023-03-31T06:51:21.2660236Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_complex64 PASSED [ 24%] 2023-03-31T06:51:21.2660614Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_float32 PASSED [ 24%] 2023-03-31T06:51:21.2660948Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_float64 PASSED [ 24%] 2023-03-31T06:51:21.2661291Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:21.2661631Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:21.2661959Z test_decomp.py::TestDecompCPU::test_comprehensive_inner_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:21.2662327Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:21.2662699Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float16 PASSED [ 25%] 2023-03-31T06:51:21.2663034Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:21.2663499Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:21.2663840Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:21.2664172Z test_decomp.py::TestDecompCPU::test_comprehensive_int_cpu_int8 PASSED [ 25%] 2023-03-31T06:51:21.2664508Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_complex128 PASSED [ 25%] 2023-03-31T06:51:21.2664861Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float16 PASSED [ 25%] 2023-03-31T06:51:21.2665211Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:21.2665563Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:21.2665897Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:21.2666241Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:21.2666589Z test_decomp.py::TestDecompCPU::test_comprehensive_isclose_cpu_int8 PASSED [ 25%] 2023-03-31T06:51:21.2666932Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:21.2667284Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_bool PASSED [ 25%] 2023-03-31T06:51:21.2667640Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex128 PASSED [ 25%] 2023-03-31T06:51:21.2668003Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_complex32 PASSED [ 25%] 2023-03-31T06:51:21.2668346Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float32 PASSED [ 25%] 2023-03-31T06:51:21.2668702Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:21.2669058Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:21.2669393Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:21.2669742Z test_decomp.py::TestDecompCPU::test_comprehensive_isfinite_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:21.2670091Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:21.2670434Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int16 PASSED [ 25%] 2023-03-31T06:51:21.2670761Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int32 PASSED [ 25%] 2023-03-31T06:51:21.2671095Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_int64 PASSED [ 25%] 2023-03-31T06:51:21.2671430Z test_decomp.py::TestDecompCPU::test_comprehensive_isin_cpu_uint8 PASSED [ 25%] 2023-03-31T06:51:21.2671769Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex128 PASSED [ 25%] 2023-03-31T06:51:21.2672128Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex32 PASSED [ 25%] 2023-03-31T06:51:21.2672478Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:21.2672830Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float16 PASSED [ 25%] 2023-03-31T06:51:21.2673164Z test_decomp.py::TestDecompCPU::test_comprehensive_isinf_cpu_float64 PASSED [ 25%] 2023-03-31T06:51:21.2673544Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_bfloat16 PASSED [ 25%] 2023-03-31T06:51:21.2673901Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_complex128 PASSED [ 25%] 2023-03-31T06:51:21.2674241Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:21.2674587Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float16 PASSED [ 26%] 2023-03-31T06:51:21.2674956Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_float32 PASSED [ 26%] 2023-03-31T06:51:21.2675328Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_int32 PASSED [ 26%] 2023-03-31T06:51:21.2675653Z test_decomp.py::TestDecompCPU::test_comprehensive_isnan_cpu_uint8 PASSED [ 26%] 2023-03-31T06:51:21.2675999Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float16 PASSED [ 26%] 2023-03-31T06:51:21.2676425Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float32 PASSED [ 26%] 2023-03-31T06:51:21.2676764Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_float64 PASSED [ 26%] 2023-03-31T06:51:21.2677113Z test_decomp.py::TestDecompCPU::test_comprehensive_isneginf_cpu_int8 PASSED [ 26%] 2023-03-31T06:51:21.2677461Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_bfloat16 PASSED [ 26%] 2023-03-31T06:51:21.2677814Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_float64 PASSED [ 26%] 2023-03-31T06:51:21.2678155Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int16 PASSED [ 26%] 2023-03-31T06:51:21.2678503Z test_decomp.py::TestDecompCPU::test_comprehensive_isposinf_cpu_int64 PASSED [ 26%] 2023-03-31T06:51:21.2678849Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_bool PASSED [ 26%] 2023-03-31T06:51:21.2679183Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float16 PASSED [ 26%] 2023-03-31T06:51:21.2679346Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float32 PASSED [ 26%] 2023-03-31T06:51:21.2679514Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_float64 PASSED [ 26%] 2023-03-31T06:51:21.2679674Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int32 PASSED [ 26%] 2023-03-31T06:51:21.2679835Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int64 PASSED [ 26%] 2023-03-31T06:51:21.2679997Z test_decomp.py::TestDecompCPU::test_comprehensive_isreal_cpu_int8 PASSED [ 26%] 2023-03-31T06:51:21.2680161Z test_decomp.py::TestDecompCPU::test_comprehensive_istft_cpu_complex64 PASSED [ 26%] 2023-03-31T06:51:21.2680375Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_bool SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2680592Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_complex128 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2680796Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_complex64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2681007Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2681216Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_float32 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2681421Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_2inputs_2outputs_cpu_int8 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2681638Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_bool SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2681858Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_float32 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2682069Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2682312Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int32 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2682522Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_4inputs_with_extra_args_cpu_int8 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2682720Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2682903Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_float64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2683349Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int16 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2683580Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int32 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2683776Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_cpu_int8 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2684030Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_complex64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2695774Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_binary_return_by_ref_cpu_int64 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2696074Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_complex128 SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T06:51:21.2696283Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:21.2696498Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_float32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:21.2696702Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int16 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:21.2696896Z test_decomp.py::TestDecompCPU::test_comprehensive_jiterator_unary_cpu_int32 SKIPPED (Only runs on cuda) [ 27%] 2023-03-31T06:51:21.2697072Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_complex128 PASSED [ 27%] 2023-03-31T06:51:21.2697234Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_complex64 PASSED [ 27%] 2023-03-31T06:51:21.2697398Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float16 PASSED [ 27%] 2023-03-31T06:51:21.2697563Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_float64 PASSED [ 27%] 2023-03-31T06:51:21.2697724Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_int32 PASSED [ 27%] 2023-03-31T06:51:21.2697886Z test_decomp.py::TestDecompCPU::test_comprehensive_kron_cpu_uint8 PASSED [ 27%] 2023-03-31T06:51:21.2698057Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_float32 PASSED [ 27%] 2023-03-31T06:51:21.2698224Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_float64 PASSED [ 27%] 2023-03-31T06:51:21.2698389Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:21.2698543Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_int32 PASSED [ 27%] 2023-03-31T06:51:21.2698706Z test_decomp.py::TestDecompCPU::test_comprehensive_kthvalue_cpu_uint8 PASSED [ 27%] 2023-03-31T06:51:21.2698864Z test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:21.2699021Z test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int64 PASSED [ 27%] 2023-03-31T06:51:21.2699177Z test_decomp.py::TestDecompCPU::test_comprehensive_lcm_cpu_int8 PASSED [ 27%] 2023-03-31T06:51:21.2699345Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_complex64 PASSED [ 27%] 2023-03-31T06:51:21.2699511Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float16 PASSED [ 27%] 2023-03-31T06:51:21.2699676Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float32 PASSED [ 27%] 2023-03-31T06:51:21.2699826Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_float64 PASSED [ 27%] 2023-03-31T06:51:21.2699991Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:21.2700316Z test_decomp.py::TestDecompCPU::test_comprehensive_ldexp_cpu_int8 PASSED [ 27%] 2023-03-31T06:51:21.2700477Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:21.2700631Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_bool PASSED [ 27%] 2023-03-31T06:51:21.2700786Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float16 PASSED [ 27%] 2023-03-31T06:51:21.2700939Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_float32 PASSED [ 27%] 2023-03-31T06:51:21.2701133Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:21.2701305Z test_decomp.py::TestDecompCPU::test_comprehensive_le_cpu_uint8 PASSED [ 27%] 2023-03-31T06:51:21.2701469Z test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:21.2701672Z test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_complex128 PASSED [ 27%] 2023-03-31T06:51:21.2701836Z test_decomp.py::TestDecompCPU::test_comprehensive_lerp_cpu_float64 PASSED [ 27%] 2023-03-31T06:51:21.2702003Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_bfloat16 PASSED [ 27%] 2023-03-31T06:51:21.2702164Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int16 PASSED [ 27%] 2023-03-31T06:51:21.2702325Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int32 PASSED [ 27%] 2023-03-31T06:51:21.2702483Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int64 PASSED [ 27%] 2023-03-31T06:51:21.2702629Z test_decomp.py::TestDecompCPU::test_comprehensive_lgamma_cpu_int8 PASSED [ 27%] 2023-03-31T06:51:21.2702807Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2702979Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2703148Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cond_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2703325Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2703618Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2703787Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int64 PASSED [ 28%] 2023-03-31T06:51:21.2703956Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cross_cpu_int8 PASSED [ 28%] 2023-03-31T06:51:21.2704129Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2704288Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2704477Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2704660Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_det_singular_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2704839Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_bfloat16 PASSED [ 28%] 2023-03-31T06:51:21.2705022Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_complex32 PASSED [ 28%] 2023-03-31T06:51:21.2705198Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2705371Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_diagonal_cpu_uint8 PASSED [ 28%] 2023-03-31T06:51:21.2705542Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eig_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2705704Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigh_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2705887Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2706063Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2706236Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvals_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2706413Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_eigvalsh_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2706648Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2706846Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2707039Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_householder_product_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2707211Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2707392Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2707588Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2707757Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_inv_ex_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2707971Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2708150Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2708329Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2708513Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2708697Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2708878Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_factor_ex_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2709041Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_float32 PASSED [ 28%] 2023-03-31T06:51:21.2709215Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_ldl_solve_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2709391Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_complex128 PASSED [ 28%] 2023-03-31T06:51:21.2709563Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_cpu_float64 PASSED [ 28%] 2023-03-31T06:51:21.2709759Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_complex64 PASSED [ 28%] 2023-03-31T06:51:21.2709949Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2710134Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lstsq_grad_oriented_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2710309Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:21.2710467Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2710650Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:21.2710830Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2711005Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2711185Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2711364Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_factor_ex_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2711536Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_lu_solve_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2711715Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_bfloat16 PASSED [ 29%] 2023-03-31T06:51:21.2711895Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_norm_cpu_float16 PASSED [ 29%] 2023-03-31T06:51:21.2712067Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:21.2712250Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2712473Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2712654Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_power_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2712832Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2713029Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_matrix_rank_hermitian_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2713234Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2713437Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2713608Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int16 PASSED [ 29%] 2023-03-31T06:51:21.2713764Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_int8 PASSED [ 29%] 2023-03-31T06:51:21.2713979Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_multi_dot_cpu_uint8 PASSED [ 29%] 2023-03-31T06:51:21.2714152Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_bfloat16 PASSED [ 29%] 2023-03-31T06:51:21.2714320Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2714523Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2714722Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float16 PASSED [ 29%] 2023-03-31T06:51:21.2714922Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_norm_subgradients_at_zero_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2715096Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2715252Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2715440Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2715622Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_hermitian_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2715879Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_pinv_singular_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 29%] 2023-03-31T06:51:21.2716050Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_complex128 PASSED [ 29%] 2023-03-31T06:51:21.2716222Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_complex64 PASSED [ 29%] 2023-03-31T06:51:21.2716391Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_float32 PASSED [ 29%] 2023-03-31T06:51:21.2716552Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_qr_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2716726Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_slogdet_cpu_float64 PASSED [ 29%] 2023-03-31T06:51:21.2716885Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2717066Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:21.2717242Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_ex_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2717431Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_solve_triangular_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2717601Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2717773Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svd_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2717953Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:21.2718127Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2718299Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_svdvals_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2718493Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorinv_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2718679Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2718857Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2719032Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_tensorsolve_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2719232Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2719439Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int32 PASSED [ 30%] 2023-03-31T06:51:21.2719608Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vander_cpu_int64 PASSED [ 30%] 2023-03-31T06:51:21.2719786Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:21.2719977Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2720153Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2720324Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vecdot_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2720507Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:21.2720688Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2720865Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float16 PASSED [ 30%] 2023-03-31T06:51:21.2721045Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2721222Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_vector_norm_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2721396Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2721553Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_float16 PASSED [ 30%] 2023-03-31T06:51:21.2721719Z test_decomp.py::TestDecompCPU::test_comprehensive_linspace_cpu_int8 PASSED [ 30%] 2023-03-31T06:51:21.2721886Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2722044Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int64 PASSED [ 30%] 2023-03-31T06:51:21.2722203Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_int8 PASSED [ 30%] 2023-03-31T06:51:21.2722367Z test_decomp.py::TestDecompCPU::test_comprehensive_log10_cpu_uint8 PASSED [ 30%] 2023-03-31T06:51:21.2722536Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_bfloat16 PASSED [ 30%] 2023-03-31T06:51:21.2722707Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_complex128 PASSED [ 30%] 2023-03-31T06:51:21.2722860Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_complex64 PASSED [ 30%] 2023-03-31T06:51:21.2723295Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_float32 PASSED [ 30%] 2023-03-31T06:51:21.2723464Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_float64 PASSED [ 30%] 2023-03-31T06:51:21.2723622Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int16 PASSED [ 30%] 2023-03-31T06:51:21.2723783Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int32 PASSED [ 30%] 2023-03-31T06:51:21.2723942Z test_decomp.py::TestDecompCPU::test_comprehensive_log1p_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:21.2724100Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_bool PASSED [ 31%] 2023-03-31T06:51:21.2724265Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:21.2724405Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:21.2724563Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:21.2724718Z test_decomp.py::TestDecompCPU::test_comprehensive_log2_cpu_int8 PASSED [ 31%] 2023-03-31T06:51:21.2724939Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:21.2725096Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_bool PASSED [ 31%] 2023-03-31T06:51:21.2725260Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_complex64 PASSED [ 31%] 2023-03-31T06:51:21.2725417Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2725610Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:21.2725786Z test_decomp.py::TestDecompCPU::test_comprehensive_log_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:21.2725955Z test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float16 PASSED [ 31%] 2023-03-31T06:51:21.2726123Z test_decomp.py::TestDecompCPU::test_comprehensive_log_normal_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2726335Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2726524Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:21.2726716Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:21.2726903Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float16 PASSED [ 31%] 2023-03-31T06:51:21.2727083Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:21.2727267Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:21.2727438Z test_decomp.py::TestDecompCPU::test_comprehensive_log_softmax_with_dtype_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:21.2727617Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2727788Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp2_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:21.2727959Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:21.2728135Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:21.2728303Z test_decomp.py::TestDecompCPU::test_comprehensive_logaddexp_cpu_float64 PASSED [ 31%] 2023-03-31T06:51:21.2728481Z test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:21.2728660Z test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_complex64 PASSED [ 31%] 2023-03-31T06:51:21.2728822Z test_decomp.py::TestDecompCPU::test_comprehensive_logcumsumexp_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2728994Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:21.2729163Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_float32 PASSED [ 31%] 2023-03-31T06:51:21.2729332Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:21.2729498Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_and_cpu_uint8 PASSED [ 31%] 2023-03-31T06:51:21.2729669Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_bfloat16 PASSED [ 31%] 2023-03-31T06:51:21.2729843Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_complex128 PASSED [ 31%] 2023-03-31T06:51:21.2730009Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int16 PASSED [ 31%] 2023-03-31T06:51:21.2730178Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int32 PASSED [ 31%] 2023-03-31T06:51:21.2730330Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int64 PASSED [ 31%] 2023-03-31T06:51:21.2730499Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_not_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2730668Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_bool PASSED [ 32%] 2023-03-31T06:51:21.2730873Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:21.2731049Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:21.2731216Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float16 PASSED [ 32%] 2023-03-31T06:51:21.2731383Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:21.2731549Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:21.2731726Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:21.2731914Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_or_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2732081Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_bool PASSED [ 32%] 2023-03-31T06:51:21.2732296Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:21.2732470Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:21.2732642Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:21.2732808Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:21.2732975Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:21.2733127Z test_decomp.py::TestDecompCPU::test_comprehensive_logical_xor_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2733293Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:21.2733455Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:21.2733615Z test_decomp.py::TestDecompCPU::test_comprehensive_logit_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2733789Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:21.2733956Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:21.2734120Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:21.2734283Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2734450Z test_decomp.py::TestDecompCPU::test_comprehensive_logspace_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:21.2734600Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_bool PASSED [ 32%] 2023-03-31T06:51:21.2734771Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:21.2734939Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int16 PASSED [ 32%] 2023-03-31T06:51:21.2735107Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_int8 PASSED [ 32%] 2023-03-31T06:51:21.2735272Z test_decomp.py::TestDecompCPU::test_comprehensive_logsumexp_cpu_uint8 PASSED [ 32%] 2023-03-31T06:51:21.2735433Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_bool PASSED [ 32%] 2023-03-31T06:51:21.2735601Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex128 PASSED [ 32%] 2023-03-31T06:51:21.2735766Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex32 PASSED [ 32%] 2023-03-31T06:51:21.2735914Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_complex64 PASSED [ 32%] 2023-03-31T06:51:21.2736074Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_float32 PASSED [ 32%] 2023-03-31T06:51:21.2736237Z test_decomp.py::TestDecompCPU::test_comprehensive_long_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:21.2736398Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_bfloat16 PASSED [ 32%] 2023-03-31T06:51:21.2736560Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_float64 PASSED [ 32%] 2023-03-31T06:51:21.2736717Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int32 PASSED [ 32%] 2023-03-31T06:51:21.2736873Z test_decomp.py::TestDecompCPU::test_comprehensive_lt_cpu_int64 PASSED [ 32%] 2023-03-31T06:51:21.2737065Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:21.2737210Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2737381Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:21.2737548Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_solve_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2737757Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:21.2737994Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:21.2738163Z test_decomp.py::TestDecompCPU::test_comprehensive_lu_unpack_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2738320Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_bool PASSED [ 33%] 2023-03-31T06:51:21.2738512Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex32 PASSED [ 33%] 2023-03-31T06:51:21.2738662Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:21.2738819Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2738974Z test_decomp.py::TestDecompCPU::test_comprehensive_mH_cpu_int16 PASSED [ 33%] 2023-03-31T06:51:21.2739135Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:21.2739300Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_complex32 PASSED [ 33%] 2023-03-31T06:51:21.2739461Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2739616Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2739773Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:21.2739916Z test_decomp.py::TestDecompCPU::test_comprehensive_mT_cpu_int8 PASSED [ 33%] 2023-03-31T06:51:21.2740089Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2740259Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2740426Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int32 PASSED [ 33%] 2023-03-31T06:51:21.2740587Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:21.2740753Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_int8 PASSED [ 33%] 2023-03-31T06:51:21.2740921Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amax_cpu_uint8 PASSED [ 33%] 2023-03-31T06:51:21.2741094Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:21.2741263Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float16 PASSED [ 33%] 2023-03-31T06:51:21.2741419Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2741587Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_amin_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:21.2741759Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2741929Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmax_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2742103Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_bfloat16 PASSED [ 33%] 2023-03-31T06:51:21.2742276Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2742446Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_int64 PASSED [ 33%] 2023-03-31T06:51:21.2742610Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_argmin_cpu_uint8 PASSED [ 33%] 2023-03-31T06:51:21.2742778Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_complex128 PASSED [ 33%] 2023-03-31T06:51:21.2742986Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_complex64 PASSED [ 33%] 2023-03-31T06:51:21.2743161Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_float32 PASSED [ 33%] 2023-03-31T06:51:21.2743417Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_float64 PASSED [ 33%] 2023-03-31T06:51:21.2743611Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:21.2743783Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:21.2743991Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int64 PASSED [ 34%] 2023-03-31T06:51:21.2744190Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:21.2744358Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumprod_cpu_uint8 PASSED [ 34%] 2023-03-31T06:51:21.2744548Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_complex128 PASSED [ 34%] 2023-03-31T06:51:21.2744726Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_complex64 PASSED [ 34%] 2023-03-31T06:51:21.2744895Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:21.2745064Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_cumsum_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:21.2745232Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:21.2745407Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_complex128 PASSED [ 34%] 2023-03-31T06:51:21.2745578Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:21.2745744Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:21.2745897Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:21.2746062Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:21.2746230Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int64 PASSED [ 34%] 2023-03-31T06:51:21.2746398Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_fill_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:21.2746576Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_log_softmax_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:21.2746755Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:21.2746935Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:21.2747113Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logaddexp_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:21.2747310Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_float64 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:21.2747492Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int16 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:21.2747686Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int32 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:21.2747874Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_logsumexp_cpu_int8 SKIPPED (Skipped!) [ 34%] 2023-03-31T06:51:21.2748049Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_complex128 PASSED [ 34%] 2023-03-31T06:51:21.2748223Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_complex64 PASSED [ 34%] 2023-03-31T06:51:21.2748393Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:21.2748565Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:21.2748731Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int16 PASSED [ 34%] 2023-03-31T06:51:21.2748883Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int32 PASSED [ 34%] 2023-03-31T06:51:21.2749052Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_mean_cpu_int8 PASSED [ 34%] 2023-03-31T06:51:21.2749252Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_median_cpu_float64 PASSED [ 34%] 2023-03-31T06:51:21.2749425Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_bfloat16 PASSED [ 34%] 2023-03-31T06:51:21.2749594Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float16 PASSED [ 34%] 2023-03-31T06:51:21.2749758Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_norm_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:21.2749965Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_complex128 PASSED [ 34%] 2023-03-31T06:51:21.2750170Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_normalize_cpu_float32 PASSED [ 34%] 2023-03-31T06:51:21.2750347Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_complex128 PASSED [ 34%] 2023-03-31T06:51:21.2750528Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2750697Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:21.2750860Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:21.2751027Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_prod_cpu_int8 PASSED [ 35%] 2023-03-31T06:51:21.2751201Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:21.2751370Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_bool PASSED [ 35%] 2023-03-31T06:51:21.2751552Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_complex128 PASSED [ 35%] 2023-03-31T06:51:21.2751728Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float16 PASSED [ 35%] 2023-03-31T06:51:21.2751884Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:21.2752056Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2752229Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_scatter_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:21.2752396Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_bool PASSED [ 35%] 2023-03-31T06:51:21.2752573Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_complex128 PASSED [ 35%] 2023-03-31T06:51:21.2752753Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:21.2752924Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float16 PASSED [ 35%] 2023-03-31T06:51:21.2753099Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2753254Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_select_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:21.2753429Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:21.2753604Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmax_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:21.2753774Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_softmin_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2753945Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_bfloat16 PASSED [ 35%] 2023-03-31T06:51:21.2754116Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:21.2754285Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2754453Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:21.2754619Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int32 PASSED [ 35%] 2023-03-31T06:51:21.2754770Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_int8 PASSED [ 35%] 2023-03-31T06:51:21.2754939Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_std_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:21.2755139Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_complex64 PASSED [ 35%] 2023-03-31T06:51:21.2755308Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float16 PASSED [ 35%] 2023-03-31T06:51:21.2755475Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float32 PASSED [ 35%] 2023-03-31T06:51:21.2755642Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2755809Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_int8 PASSED [ 35%] 2023-03-31T06:51:21.2756000Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_sum_cpu_uint8 PASSED [ 35%] 2023-03-31T06:51:21.2756187Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float16 PASSED [ 35%] 2023-03-31T06:51:21.2756355Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_float64 PASSED [ 35%] 2023-03-31T06:51:21.2756562Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int16 PASSED [ 35%] 2023-03-31T06:51:21.2756729Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int32 PASSED [ 35%] 2023-03-31T06:51:21.2756891Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_int64 PASSED [ 35%] 2023-03-31T06:51:21.2757050Z test_decomp.py::TestDecompCPU::test_comprehensive_masked_var_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:21.2757214Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2757376Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_complex64 PASSED [ 36%] 2023-03-31T06:51:21.2757527Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:21.2757690Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:21.2757852Z test_decomp.py::TestDecompCPU::test_comprehensive_matmul_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:21.2758029Z test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_complex128 PASSED [ 36%] 2023-03-31T06:51:21.2758203Z test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_complex64 PASSED [ 36%] 2023-03-31T06:51:21.2758370Z test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:21.2758537Z test_decomp.py::TestDecompCPU::test_comprehensive_matrix_exp_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:21.2758706Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2758874Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float16 PASSED [ 36%] 2023-03-31T06:51:21.2759029Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:21.2759197Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int16 PASSED [ 36%] 2023-03-31T06:51:21.2759360Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:21.2759525Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:21.2759689Z test_decomp.py::TestDecompCPU::test_comprehensive_max_binary_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:21.2759885Z test_decomp.py::TestDecompCPU::test_comprehensive_max_pool2d_with_indices_backward_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:21.2760067Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2760245Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_bool PASSED [ 36%] 2023-03-31T06:51:21.2760414Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:21.2760597Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:21.2760773Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:21.2760952Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_no_dim_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:21.2761180Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2761362Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_bool PASSED [ 36%] 2023-03-31T06:51:21.2761546Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float16 PASSED [ 36%] 2023-03-31T06:51:21.2761725Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:21.2761927Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:21.2762119Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int16 PASSED [ 36%] 2023-03-31T06:51:21.2762295Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_int64 PASSED [ 36%] 2023-03-31T06:51:21.2762503Z test_decomp.py::TestDecompCPU::test_comprehensive_max_reduction_with_dim_cpu_uint8 PASSED [ 36%] 2023-03-31T06:51:21.2762670Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int32 PASSED [ 36%] 2023-03-31T06:51:21.2762834Z test_decomp.py::TestDecompCPU::test_comprehensive_maximum_cpu_int8 PASSED [ 36%] 2023-03-31T06:51:21.2762996Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2763368Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float16 PASSED [ 36%] 2023-03-31T06:51:21.2763528Z test_decomp.py::TestDecompCPU::test_comprehensive_mean_cpu_float64 PASSED [ 36%] 2023-03-31T06:51:21.2763685Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_bfloat16 PASSED [ 36%] 2023-03-31T06:51:21.2763852Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_float32 PASSED [ 36%] 2023-03-31T06:51:21.2764013Z test_decomp.py::TestDecompCPU::test_comprehensive_median_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:21.2764257Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_bfloat16 SKIPPED (meshgrid in torch.bfloat16 not supported) [ 37%] 2023-03-31T06:51:21.2764490Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_bool SKIPPED (meshgrid in torch.bool not supported) [ 37%] 2023-03-31T06:51:21.2764737Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_complex128 SKIPPED (meshgrid in torch.complex128 not supported) [ 37%] 2023-03-31T06:51:21.2764981Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_complex64 SKIPPED (meshgrid in torch.complex64 not supported) [ 37%] 2023-03-31T06:51:21.2765218Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_float64 SKIPPED (meshgrid in torch.float64 not supported) [ 37%] 2023-03-31T06:51:21.2765451Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int32 SKIPPED (meshgrid in torch.int32 not supported) [ 37%] 2023-03-31T06:51:21.2765683Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int64 SKIPPED (meshgrid in torch.int64 not supported) [ 37%] 2023-03-31T06:51:21.2765900Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_int8 SKIPPED (meshgrid in torch.int8 not supported) [ 37%] 2023-03-31T06:51:21.2766129Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_list_of_tensors_cpu_uint8 SKIPPED (meshgrid in torch.uint8 not supported) [ 37%] 2023-03-31T06:51:21.2766370Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_bfloat16 SKIPPED (meshgrid in torch.bfloat16 not supported) [ 37%] 2023-03-31T06:51:21.2766614Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_complex64 SKIPPED (meshgrid in torch.complex64 not supported) [ 37%] 2023-03-31T06:51:21.2766851Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_float32 SKIPPED (meshgrid in torch.float32 not supported) [ 37%] 2023-03-31T06:51:21.2767087Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int64 SKIPPED (meshgrid in torch.int64 not supported) [ 37%] 2023-03-31T06:51:21.2767387Z test_decomp.py::TestDecompCPU::test_comprehensive_meshgrid_variadic_tensors_cpu_int8 SKIPPED (meshgrid in torch.int8 not supported) [ 37%] 2023-03-31T06:51:21.2767560Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:21.2767726Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_bool PASSED [ 37%] 2023-03-31T06:51:21.2767895Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float16 PASSED [ 37%] 2023-03-31T06:51:21.2768087Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_float64 PASSED [ 37%] 2023-03-31T06:51:21.2768292Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:21.2768457Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:21.2768659Z test_decomp.py::TestDecompCPU::test_comprehensive_min_binary_cpu_uint8 PASSED [ 37%] 2023-03-31T06:51:21.2768842Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:21.2769020Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_bool PASSED [ 37%] 2023-03-31T06:51:21.2769202Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float16 PASSED [ 37%] 2023-03-31T06:51:21.2769382Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:21.2769562Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int32 PASSED [ 37%] 2023-03-31T06:51:21.2769727Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_no_dim_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:21.2769911Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_bfloat16 PASSED [ 37%] 2023-03-31T06:51:21.2770095Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_float64 PASSED [ 37%] 2023-03-31T06:51:21.2770279Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int32 PASSED [ 37%] 2023-03-31T06:51:21.2770458Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:21.2770635Z test_decomp.py::TestDecompCPU::test_comprehensive_min_reduction_with_dim_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:21.2770802Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_bool PASSED [ 37%] 2023-03-31T06:51:21.2770966Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int16 PASSED [ 37%] 2023-03-31T06:51:21.2771120Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int32 PASSED [ 37%] 2023-03-31T06:51:21.2771285Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int64 PASSED [ 37%] 2023-03-31T06:51:21.2771448Z test_decomp.py::TestDecompCPU::test_comprehensive_minimum_cpu_int8 PASSED [ 37%] 2023-03-31T06:51:21.2771610Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2771775Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_complex128 PASSED [ 38%] 2023-03-31T06:51:21.2771935Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_complex64 PASSED [ 38%] 2023-03-31T06:51:21.2772091Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:21.2772249Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:21.2772392Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:21.2772554Z test_decomp.py::TestDecompCPU::test_comprehensive_mm_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:21.2772711Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_bool PASSED [ 38%] 2023-03-31T06:51:21.2772870Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float16 PASSED [ 38%] 2023-03-31T06:51:21.2773029Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:21.2773185Z test_decomp.py::TestDecompCPU::test_comprehensive_mode_cpu_int32 PASSED [ 38%] 2023-03-31T06:51:21.2773383Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_bool PASSED [ 38%] 2023-03-31T06:51:21.2773555Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_complex64 PASSED [ 38%] 2023-03-31T06:51:21.2773707Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float16 PASSED [ 38%] 2023-03-31T06:51:21.2773874Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:21.2774069Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:21.2774261Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:21.2774422Z test_decomp.py::TestDecompCPU::test_comprehensive_movedim_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:21.2774587Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2774773Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_bool PASSED [ 38%] 2023-03-31T06:51:21.2774933Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:21.2775092Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:21.2775235Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_int8 PASSED [ 38%] 2023-03-31T06:51:21.2775393Z test_decomp.py::TestDecompCPU::test_comprehensive_msort_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:21.2775550Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_bool PASSED [ 38%] 2023-03-31T06:51:21.2775714Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_complex64 PASSED [ 38%] 2023-03-31T06:51:21.2775873Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:21.2776026Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int64 PASSED [ 38%] 2023-03-31T06:51:21.2776183Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_int8 PASSED [ 38%] 2023-03-31T06:51:21.2776343Z test_decomp.py::TestDecompCPU::test_comprehensive_mul_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:21.2776504Z test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2776675Z test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_float32 PASSED [ 38%] 2023-03-31T06:51:21.2776842Z test_decomp.py::TestDecompCPU::test_comprehensive_multinomial_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:21.2776997Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2777155Z test_decomp.py::TestDecompCPU::test_comprehensive_mv_cpu_int16 PASSED [ 38%] 2023-03-31T06:51:21.2777341Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2777521Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_1_cpu_uint8 PASSED [ 38%] 2023-03-31T06:51:21.2777708Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_bfloat16 PASSED [ 38%] 2023-03-31T06:51:21.2777879Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_float64 PASSED [ 38%] 2023-03-31T06:51:21.2778058Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:21.2778235Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:21.2778412Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2778590Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_3_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:21.2778775Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:21.2778954Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:21.2779136Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2779340Z test_decomp.py::TestDecompCPU::test_comprehensive_mvlgamma_mvlgamma_p_5_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:21.2779498Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:21.2779668Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float16 PASSED [ 39%] 2023-03-31T06:51:21.2779835Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:21.2780000Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:21.2780201Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:21.2780394Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:21.2780558Z test_decomp.py::TestDecompCPU::test_comprehensive_nan_to_num_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2780753Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmean_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:21.2780904Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:21.2781072Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2781235Z test_decomp.py::TestDecompCPU::test_comprehensive_nanmedian_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:21.2781406Z test_decomp.py::TestDecompCPU::test_comprehensive_nanquantile_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:21.2781570Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float16 PASSED [ 39%] 2023-03-31T06:51:21.2781731Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_float64 PASSED [ 39%] 2023-03-31T06:51:21.2781896Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:21.2782057Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int64 PASSED [ 39%] 2023-03-31T06:51:21.2782206Z test_decomp.py::TestDecompCPU::test_comprehensive_nansum_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2782381Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_complex32 PASSED [ 39%] 2023-03-31T06:51:21.2782552Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float16 PASSED [ 39%] 2023-03-31T06:51:21.2782720Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:21.2782887Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int16 PASSED [ 39%] 2023-03-31T06:51:21.2783050Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int32 PASSED [ 39%] 2023-03-31T06:51:21.2783218Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_copy_cpu_int8 PASSED [ 39%] 2023-03-31T06:51:21.2783488Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_complex128 PASSED [ 39%] 2023-03-31T06:51:21.2783656Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float16 PASSED [ 39%] 2023-03-31T06:51:21.2783806Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_float32 PASSED [ 39%] 2023-03-31T06:51:21.2783971Z test_decomp.py::TestDecompCPU::test_comprehensive_narrow_cpu_uint8 PASSED [ 39%] 2023-03-31T06:51:21.2784212Z test_decomp.py::TestDecompCPU::test_comprehensive_native_batch_norm_cpu_float32 SKIPPED (native_batch_norm in torch.float32 not supported) [ 39%] 2023-03-31T06:51:21.2784403Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_bfloat16 PASSED [ 39%] 2023-03-31T06:51:21.2784583Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_bool PASSED [ 39%] 2023-03-31T06:51:21.2784773Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_float16 PASSED [ 40%] 2023-03-31T06:51:21.2784960Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int32 PASSED [ 40%] 2023-03-31T06:51:21.2785146Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_int64 PASSED [ 40%] 2023-03-31T06:51:21.2785331Z test_decomp.py::TestDecompCPU::test_comprehensive_native_dropout_backward_cpu_uint8 PASSED [ 40%] 2023-03-31T06:51:21.2785533Z test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_float32 PASSED [ 40%] 2023-03-31T06:51:21.2785713Z test_decomp.py::TestDecompCPU::test_comprehensive_native_layer_norm_cpu_float64 PASSED [ 40%] 2023-03-31T06:51:21.2785874Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_float64 PASSED [ 40%] 2023-03-31T06:51:21.2786031Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int64 PASSED [ 40%] 2023-03-31T06:51:21.2786258Z test_decomp.py::TestDecompCPU::test_comprehensive_ne_cpu_int8 PASSED [ 40%] 2023-03-31T06:51:21.2786441Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_float32 PASSED [ 40%] 2023-03-31T06:51:21.2786597Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int32 PASSED [ 40%] 2023-03-31T06:51:21.2786753Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int64 PASSED [ 40%] 2023-03-31T06:51:21.2786925Z test_decomp.py::TestDecompCPU::test_comprehensive_neg_cpu_int8 PASSED [ 40%] 2023-03-31T06:51:21.2787152Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_bfloat16 SKIPPED (new_empty in torch.bfloat16 not supported) [ 40%] 2023-03-31T06:51:21.2787367Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_bool SKIPPED (new_empty in torch.bool not supported) [ 40%] 2023-03-31T06:51:21.2787598Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex128 SKIPPED (new_empty in torch.complex128 not supported) [ 40%] 2023-03-31T06:51:21.2787827Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex32 SKIPPED (new_empty in torch.complex32 not supported) [ 40%] 2023-03-31T06:51:21.2788053Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_complex64 SKIPPED (new_empty in torch.complex64 not supported) [ 40%] 2023-03-31T06:51:21.2788274Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_float16 SKIPPED (new_empty in torch.float16 not supported) [ 40%] 2023-03-31T06:51:21.2788491Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int16 SKIPPED (new_empty in torch.int16 not supported) [ 40%] 2023-03-31T06:51:21.2788704Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_cpu_int64 SKIPPED (new_empty in torch.int64 not supported) [ 40%] 2023-03-31T06:51:21.2788941Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_complex64 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2789166Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float16 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2789404Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float32 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2789636Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_float64 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2789875Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int64 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2790103Z test_decomp.py::TestDecompCPU::test_comprehensive_new_empty_strided_cpu_int8 SKIPPED (Expected: new_empty_strided is not comparable) [ 40%] 2023-03-31T06:51:21.2790273Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_bfloat16 PASSED [ 40%] 2023-03-31T06:51:21.2790437Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_bool PASSED [ 40%] 2023-03-31T06:51:21.2790610Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_complex32 PASSED [ 40%] 2023-03-31T06:51:21.2790779Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float32 PASSED [ 40%] 2023-03-31T06:51:21.2790930Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_float64 PASSED [ 40%] 2023-03-31T06:51:21.2791094Z test_decomp.py::TestDecompCPU::test_comprehensive_new_full_cpu_uint8 PASSED [ 40%] 2023-03-31T06:51:21.2791262Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_bfloat16 PASSED [ 40%] 2023-03-31T06:51:21.2791460Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex128 PASSED [ 40%] 2023-03-31T06:51:21.2791630Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_complex64 PASSED [ 40%] 2023-03-31T06:51:21.2791793Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_float64 PASSED [ 40%] 2023-03-31T06:51:21.2791957Z test_decomp.py::TestDecompCPU::test_comprehensive_new_ones_cpu_int16 PASSED [ 40%] 2023-03-31T06:51:21.2792143Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_bool PASSED [ 40%] 2023-03-31T06:51:21.2792339Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex128 PASSED [ 41%] 2023-03-31T06:51:21.2792496Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex32 PASSED [ 41%] 2023-03-31T06:51:21.2792690Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_complex64 PASSED [ 41%] 2023-03-31T06:51:21.2792856Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float16 PASSED [ 41%] 2023-03-31T06:51:21.2793021Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2793179Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2793343Z test_decomp.py::TestDecompCPU::test_comprehensive_new_zeros_cpu_int32 PASSED [ 41%] 2023-03-31T06:51:21.2793512Z test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2793678Z test_decomp.py::TestDecompCPU::test_comprehensive_nextafter_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2793864Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool1d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2794064Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2794263Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2794462Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool2d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2794656Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_avg_pool3d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2794851Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2795052Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2795248Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool1d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2795447Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2795627Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2795822Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_adaptive_max_pool2d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2796021Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2796215Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_alpha_dropout_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2796408Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2796597Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool1d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2796782Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2796967Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool2d_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:21.2797152Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2797343Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2797529Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_avg_pool3d_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:21.2797719Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2797906Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_float32 PASSED [ 41%] 2023-03-31T06:51:21.2798114Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_batch_norm_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2798325Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_bfloat16 PASSED [ 41%] 2023-03-31T06:51:21.2798511Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_float64 PASSED [ 41%] 2023-03-31T06:51:21.2798730Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int16 PASSED [ 41%] 2023-03-31T06:51:21.2798915Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int32 PASSED [ 41%] 2023-03-31T06:51:21.2799082Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int64 PASSED [ 41%] 2023-03-31T06:51:21.2799264Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_int8 PASSED [ 41%] 2023-03-31T06:51:21.2799443Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_bilinear_cpu_uint8 PASSED [ 42%] 2023-03-31T06:51:21.2799643Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2799859Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2800072Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2800263Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_complex128 PASSED [ 42%] 2023-03-31T06:51:21.2800443Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2800619Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv1d_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:21.2800788Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2800967Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2801145Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv2d_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:21.2801340Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose1d_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:21.2801538Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_complex128 PASSED [ 42%] 2023-03-31T06:51:21.2801739Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose2d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2801934Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2802131Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_complex128 PASSED [ 42%] 2023-03-31T06:51:21.2802328Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_complex64 PASSED [ 42%] 2023-03-31T06:51:21.2802511Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_conv_transpose3d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2802712Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2802908Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int64 PASSED [ 42%] 2023-03-31T06:51:21.2803292Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_embedding_loss_cpu_int8 PASSED [ 42%] 2023-03-31T06:51:21.2803551Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cosine_similarity_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2803745Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_cross_entropy_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2803930Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_ctc_loss_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2804119Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2804340Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2804542Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout2d_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2804721Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2804935Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout3d_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2805120Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2805304Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_dropout_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2805484Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_elu_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2805676Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2805870Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float16 PASSED [ 42%] 2023-03-31T06:51:21.2806063Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_bag_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2806236Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2806424Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_embedding_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2806640Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_bfloat16 PASSED [ 42%] 2023-03-31T06:51:21.2806853Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_float32 PASSED [ 42%] 2023-03-31T06:51:21.2807066Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cpu_float64 PASSED [ 42%] 2023-03-31T06:51:21.2807283Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_bool PASSED [ 43%] 2023-03-31T06:51:21.2807506Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_complex128 PASSED [ 43%] 2023-03-31T06:51:21.2807725Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_complex64 PASSED [ 43%] 2023-03-31T06:51:21.2807944Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_float16 PASSED [ 43%] 2023-03-31T06:51:21.2808142Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cpu_int32 PASSED [ 43%] 2023-03-31T06:51:21.2808342Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool3d_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2808541Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_fractional_max_pool3d_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2808722Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2808900Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2809075Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_glu_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2809265Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_grid_sample_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2809482Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2809670Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2809839Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_group_norm_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2810028Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardshrink_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2810261Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2810471Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2810657Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardsigmoid_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2810868Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2811052Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hardtanh_cpu_int16 PASSED [ 43%] 2023-03-31T06:51:21.2811253Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_hinge_embedding_loss_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2811445Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2811624Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_instance_norm_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2811821Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_area_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2812021Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2812220Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bicubic_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2812424Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2812623Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2812820Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_bilinear_cpu_uint8 PASSED [ 43%] 2023-03-31T06:51:21.2813017Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2813214Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2813398Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_linear_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2813598Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_interpolate_nearest_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2813780Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2813962Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_float32 PASSED [ 43%] 2023-03-31T06:51:21.2814142Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_kl_div_cpu_float64 PASSED [ 43%] 2023-03-31T06:51:21.2814326Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_bfloat16 PASSED [ 43%] 2023-03-31T06:51:21.2814511Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_complex128 PASSED [ 43%] 2023-03-31T06:51:21.2814692Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2814874Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_l1_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2815048Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:21.2815236Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_layer_norm_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2815449Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_leaky_relu_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2815633Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:21.2815816Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_complex128 PASSED [ 44%] 2023-03-31T06:51:21.2815995Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int16 PASSED [ 44%] 2023-03-31T06:51:21.2816200Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_int8 PASSED [ 44%] 2023-03-31T06:51:21.2816401Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_linear_cpu_uint8 PASSED [ 44%] 2023-03-31T06:51:21.2816598Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_local_response_norm_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2816805Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_logsigmoid_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:21.2817003Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2817196Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_int64 PASSED [ 44%] 2023-03-31T06:51:21.2817390Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_margin_ranking_loss_cpu_uint8 PASSED [ 44%] 2023-03-31T06:51:21.2817578Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:21.2817767Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool1d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2817950Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_bfloat16 PASSED [ 44%] 2023-03-31T06:51:21.2818134Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool2d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2818322Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_pool3d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2818500Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2818692Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool1d_grad_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2818878Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2819075Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool2d_grad_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2819265Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2819453Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_max_unpool3d_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2819633Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2819813Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mish_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2819999Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_mse_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2820178Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multi_margin_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2820370Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multi_margin_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2820573Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_margin_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2820775Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_multilabel_margin_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2820955Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_float32 PASSED [ 44%] 2023-03-31T06:51:21.2821136Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_nll_loss_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2821349Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_normalize_cpu_float64 PASSED [ 44%] 2023-03-31T06:51:21.2821594Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_bool SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:21.2821844Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_complex128 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:21.2822121Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_complex64 SKIPPED (Expected: new_empty_strided is not comparable) [ 44%] 2023-03-31T06:51:21.2822392Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float32 SKIPPED (Expected: new_empty_strided is not comparable) [ 45%] 2023-03-31T06:51:21.2822659Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_float64 SKIPPED (Expected: new_empty_strided is not comparable) [ 45%] 2023-03-31T06:51:21.2822903Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int16 SKIPPED (Expected: new_empty_strided is not comparable) [ 45%] 2023-03-31T06:51:21.2823136Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_int64 SKIPPED (Expected: new_empty_strided is not comparable) [ 45%] 2023-03-31T06:51:21.2823475Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_circular_cpu_uint8 SKIPPED (Expected: new_empty_strided is not comparable) [ 45%] 2023-03-31T06:51:21.2823673Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_bool PASSED [ 45%] 2023-03-31T06:51:21.2823862Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float16 PASSED [ 45%] 2023-03-31T06:51:21.2824056Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:21.2824248Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int32 PASSED [ 45%] 2023-03-31T06:51:21.2824433Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:21.2824606Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_constant_cpu_uint8 PASSED [ 45%] 2023-03-31T06:51:21.2824797Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_complex64 PASSED [ 45%] 2023-03-31T06:51:21.2824987Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_reflect_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:21.2825185Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_complex128 PASSED [ 45%] 2023-03-31T06:51:21.2825375Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pad_replicate_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:21.2825574Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:21.2825775Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_complex128 PASSED [ 45%] 2023-03-31T06:51:21.2825972Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_complex64 PASSED [ 45%] 2023-03-31T06:51:21.2826168Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:21.2826347Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pairwise_distance_cpu_uint8 PASSED [ 45%] 2023-03-31T06:51:21.2826529Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pdist_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:21.2826723Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_bfloat16 PASSED [ 45%] 2023-03-31T06:51:21.2826919Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_complex128 PASSED [ 45%] 2023-03-31T06:51:21.2827111Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_float16 PASSED [ 45%] 2023-03-31T06:51:21.2827336Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int32 PASSED [ 45%] 2023-03-31T06:51:21.2827527Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_shuffle_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:21.2827718Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_bool PASSED [ 45%] 2023-03-31T06:51:21.2827917Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_complex128 PASSED [ 45%] 2023-03-31T06:51:21.2828125Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_float16 PASSED [ 45%] 2023-03-31T06:51:21.2828340Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int32 PASSED [ 45%] 2023-03-31T06:51:21.2828525Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:21.2828744Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_pixel_unshuffle_cpu_uint8 PASSED [ 45%] 2023-03-31T06:51:21.2828937Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_float64 PASSED [ 45%] 2023-03-31T06:51:21.2829126Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int16 PASSED [ 45%] 2023-03-31T06:51:21.2829314Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int64 PASSED [ 45%] 2023-03-31T06:51:21.2829503Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_int8 PASSED [ 45%] 2023-03-31T06:51:21.2829696Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_poisson_nll_loss_cpu_uint8 PASSED [ 45%] 2023-03-31T06:51:21.2829863Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_float32 PASSED [ 45%] 2023-03-31T06:51:21.2830046Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_prelu_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2830299Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_bfloat16 SKIPPED (nn.functional.relu6 in torch.bfloat16 not supported) [ 46%] 2023-03-31T06:51:21.2830544Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_float32 SKIPPED (nn.functional.relu6 in torch.float32 not supported) [ 46%] 2023-03-31T06:51:21.2830780Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int64 SKIPPED (nn.functional.relu6 in torch.int64 not supported) [ 46%] 2023-03-31T06:51:21.2831020Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_int8 SKIPPED (nn.functional.relu6 in torch.int8 not supported) [ 46%] 2023-03-31T06:51:21.2831261Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu6_cpu_uint8 SKIPPED (nn.functional.relu6 in torch.uint8 not supported) [ 46%] 2023-03-31T06:51:21.2831446Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2831628Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:21.2831792Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2831971Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:21.2832146Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int32 PASSED [ 46%] 2023-03-31T06:51:21.2832324Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:21.2832502Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_relu_cpu_uint8 PASSED [ 46%] 2023-03-31T06:51:21.2832683Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2832862Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_rrelu_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2833154Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_bfloat16 SKIPPED (Skipped) [ 46%] 2023-03-31T06:51:21.2833380Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_scaled_dot_product_attention_cpu_float64 SKIPPED (Skipped) [ 46%] 2023-03-31T06:51:21.2833546Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_selu_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2833742Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_complex_cpu_complex64 PASSED [ 46%] 2023-03-31T06:51:21.2833946Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2834150Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_silu_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2834340Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:21.2834560Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_smooth_l1_loss_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2834759Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_soft_margin_loss_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2834945Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2835128Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:21.2835296Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2835498Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2835702Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_complex128 PASSED [ 46%] 2023-03-31T06:51:21.2835899Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float16 PASSED [ 46%] 2023-03-31T06:51:21.2836095Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2836291Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int16 PASSED [ 46%] 2023-03-31T06:51:21.2836484Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int32 PASSED [ 46%] 2023-03-31T06:51:21.2836674Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softmin_with_dtype_cpu_int8 PASSED [ 46%] 2023-03-31T06:51:21.2836860Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softplus_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2837039Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2837228Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_float32 PASSED [ 46%] 2023-03-31T06:51:21.2837411Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softshrink_cpu_float64 PASSED [ 46%] 2023-03-31T06:51:21.2837597Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_bfloat16 PASSED [ 46%] 2023-03-31T06:51:21.2837780Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:21.2837965Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int32 PASSED [ 47%] 2023-03-31T06:51:21.2838146Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_softsign_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2838337Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_bfloat16 PASSED [ 47%] 2023-03-31T06:51:21.2838521Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int16 PASSED [ 47%] 2023-03-31T06:51:21.2838685Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2838865Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_tanhshrink_cpu_uint8 PASSED [ 47%] 2023-03-31T06:51:21.2839075Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_bfloat16 PASSED [ 47%] 2023-03-31T06:51:21.2839259Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:21.2839436Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:21.2839620Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int32 PASSED [ 47%] 2023-03-31T06:51:21.2839837Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_threshold_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2840066Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_complex128 PASSED [ 47%] 2023-03-31T06:51:21.2840266Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:21.2840484Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:21.2840681Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2840876Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_loss_cpu_uint8 PASSED [ 47%] 2023-03-31T06:51:21.2841091Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_bfloat16 PASSED [ 47%] 2023-03-31T06:51:21.2841309Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_complex128 PASSED [ 47%] 2023-03-31T06:51:21.2841526Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:21.2841743Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:21.2841955Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2842167Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cpu_int8 PASSED [ 47%] 2023-03-31T06:51:21.2842340Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_complex128 PASSED [ 47%] 2023-03-31T06:51:21.2842524Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_complex64 PASSED [ 47%] 2023-03-31T06:51:21.2842704Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:21.2842888Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_unfold_cpu_float32 PASSED [ 47%] 2023-03-31T06:51:21.2843281Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_bilinear_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:21.2843482Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_bfloat16 PASSED [ 47%] 2023-03-31T06:51:21.2843681Z test_decomp.py::TestDecompCPU::test_comprehensive_nn_functional_upsample_nearest_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:21.2843851Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_bfloat16 PASSED [ 47%] 2023-03-31T06:51:21.2844016Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_bool PASSED [ 47%] 2023-03-31T06:51:21.2844171Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_complex32 PASSED [ 47%] 2023-03-31T06:51:21.2844337Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float16 PASSED [ 47%] 2023-03-31T06:51:21.2844504Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_float64 PASSED [ 47%] 2023-03-31T06:51:21.2844670Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int16 PASSED [ 47%] 2023-03-31T06:51:21.2844833Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int32 PASSED [ 47%] 2023-03-31T06:51:21.2844995Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_int64 PASSED [ 47%] 2023-03-31T06:51:21.2845162Z test_decomp.py::TestDecompCPU::test_comprehensive_nonzero_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:21.2845431Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 48%] 2023-03-31T06:51:21.2845656Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 48%] 2023-03-31T06:51:21.2845855Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 48%] 2023-03-31T06:51:21.2846098Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 48%] 2023-03-31T06:51:21.2846340Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_fro_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 48%] 2023-03-31T06:51:21.2846553Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 48%] 2023-03-31T06:51:21.2846801Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 48%] 2023-03-31T06:51:21.2847010Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_inf_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 48%] 2023-03-31T06:51:21.2847215Z test_decomp.py::TestDecompCPU::test_comprehensive_norm_nuc_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 48%] 2023-03-31T06:51:21.2847393Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:21.2847576Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_complex128 PASSED [ 48%] 2023-03-31T06:51:21.2847741Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_in_place_cpu_float64 PASSED [ 48%] 2023-03-31T06:51:21.2847936Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_bfloat16 SKIPPED (Skipped!) [ 48%] 2023-03-31T06:51:21.2848134Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float16 SKIPPED (Skipped!) [ 48%] 2023-03-31T06:51:21.2848326Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float32 SKIPPED (Skipped!) [ 48%] 2023-03-31T06:51:21.2848515Z test_decomp.py::TestDecompCPU::test_comprehensive_normal_number_mean_cpu_float64 SKIPPED (Skipped!) [ 48%] 2023-03-31T06:51:21.2848678Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:21.2848847Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_complex64 PASSED [ 48%] 2023-03-31T06:51:21.2849011Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:21.2849174Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int16 PASSED [ 48%] 2023-03-31T06:51:21.2849320Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int32 PASSED [ 48%] 2023-03-31T06:51:21.2849480Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:21.2849642Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_int8 PASSED [ 48%] 2023-03-31T06:51:21.2849802Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:21.2849972Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:21.2850136Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_bool PASSED [ 48%] 2023-03-31T06:51:21.2850308Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex128 PASSED [ 48%] 2023-03-31T06:51:21.2850480Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex32 PASSED [ 48%] 2023-03-31T06:51:21.2850639Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_complex64 PASSED [ 48%] 2023-03-31T06:51:21.2850806Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:21.2850969Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:21.2851131Z test_decomp.py::TestDecompCPU::test_comprehensive_ones_like_cpu_uint8 PASSED [ 48%] 2023-03-31T06:51:21.2851325Z test_decomp.py::TestDecompCPU::test_comprehensive_ormqr_cpu_float32 PASSED [ 48%] 2023-03-31T06:51:21.2851492Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_bfloat16 PASSED [ 48%] 2023-03-31T06:51:21.2851652Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_bool PASSED [ 48%] 2023-03-31T06:51:21.2851826Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_complex128 PASSED [ 48%] 2023-03-31T06:51:21.2852000Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int32 PASSED [ 48%] 2023-03-31T06:51:21.2852186Z test_decomp.py::TestDecompCPU::test_comprehensive_outer_cpu_int64 PASSED [ 48%] 2023-03-31T06:51:21.2852356Z test_decomp.py::TestDecompCPU::test_comprehensive_pca_lowrank_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2852526Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:21.2852718Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_bool PASSED [ 49%] 2023-03-31T06:51:21.2852891Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_complex128 PASSED [ 49%] 2023-03-31T06:51:21.2853055Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:21.2853219Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:21.2853369Z test_decomp.py::TestDecompCPU::test_comprehensive_permute_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:21.2853547Z test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_complex128 PASSED [ 49%] 2023-03-31T06:51:21.2853718Z test_decomp.py::TestDecompCPU::test_comprehensive_pinverse_cpu_complex64 PASSED [ 49%] 2023-03-31T06:51:21.2853879Z test_decomp.py::TestDecompCPU::test_comprehensive_polar_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2854069Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:21.2854256Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2854441Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:21.2854622Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:21.2854808Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int64 PASSED [ 49%] 2023-03-31T06:51:21.2854977Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:21.2855156Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_0_cpu_uint8 PASSED [ 49%] 2023-03-31T06:51:21.2855343Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2855522Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:21.2855703Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:21.2855884Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_1_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:21.2856071Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:21.2856254Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:21.2856433Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int16 PASSED [ 49%] 2023-03-31T06:51:21.2856598Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int32 PASSED [ 49%] 2023-03-31T06:51:21.2856777Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_2_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:21.2856958Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_bool PASSED [ 49%] 2023-03-31T06:51:21.2857144Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:21.2857364Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_3_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:21.2857553Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:21.2857739Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2857920Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int64 PASSED [ 49%] 2023-03-31T06:51:21.2858125Z test_decomp.py::TestDecompCPU::test_comprehensive_polygamma_polygamma_n_4_cpu_int8 PASSED [ 49%] 2023-03-31T06:51:21.2858311Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_bfloat16 PASSED [ 49%] 2023-03-31T06:51:21.2858483Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_complex32 PASSED [ 49%] 2023-03-31T06:51:21.2858686Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float16 PASSED [ 49%] 2023-03-31T06:51:21.2858854Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float32 PASSED [ 49%] 2023-03-31T06:51:21.2859016Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_float64 PASSED [ 49%] 2023-03-31T06:51:21.2859185Z test_decomp.py::TestDecompCPU::test_comprehensive_positive_cpu_uint8 PASSED [ 50%] 2023-03-31T06:51:21.2859348Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:21.2859510Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:21.2859660Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:21.2859820Z test_decomp.py::TestDecompCPU::test_comprehensive_pow_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:21.2859983Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_bfloat16 PASSED [ 50%] 2023-03-31T06:51:21.2860144Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_bool PASSED [ 50%] 2023-03-31T06:51:21.2860315Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:21.2860482Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_complex64 PASSED [ 50%] 2023-03-31T06:51:21.2860641Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:21.2860798Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:21.2860941Z test_decomp.py::TestDecompCPU::test_comprehensive_prod_cpu_int8 PASSED [ 50%] 2023-03-31T06:51:21.2861107Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_bfloat16 PASSED [ 50%] 2023-03-31T06:51:21.2861267Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_bool PASSED [ 50%] 2023-03-31T06:51:21.2861432Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_complex128 PASSED [ 50%] 2023-03-31T06:51:21.2861590Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int16 PASSED [ 50%] 2023-03-31T06:51:21.2861747Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:21.2861903Z test_decomp.py::TestDecompCPU::test_comprehensive_put_cpu_uint8 PASSED [ 50%] 2023-03-31T06:51:21.2862074Z test_decomp.py::TestDecompCPU::test_comprehensive_quantile_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:21.2862223Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_bool PASSED [ 50%] 2023-03-31T06:51:21.2862387Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:21.2862552Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:21.2862719Z test_decomp.py::TestDecompCPU::test_comprehensive_rad2deg_cpu_uint8 PASSED [ 50%] 2023-03-31T06:51:21.2862888Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_bfloat16 PASSED [ 50%] 2023-03-31T06:51:21.2863060Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_complex32 PASSED [ 50%] 2023-03-31T06:51:21.2863230Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:21.2863542Z test_decomp.py::TestDecompCPU::test_comprehensive_rand_like_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:21.2863703Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_bfloat16 PASSED [ 50%] 2023-03-31T06:51:21.2863867Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:21.2864029Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:21.2864235Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int32 PASSED [ 50%] 2023-03-31T06:51:21.2864425Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:21.2864587Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_cpu_uint8 PASSED [ 50%] 2023-03-31T06:51:21.2864758Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float16 PASSED [ 50%] 2023-03-31T06:51:21.2864961Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float32 PASSED [ 50%] 2023-03-31T06:51:21.2865133Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:21.2865287Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int64 PASSED [ 50%] 2023-03-31T06:51:21.2865455Z test_decomp.py::TestDecompCPU::test_comprehensive_randint_like_cpu_int8 PASSED [ 50%] 2023-03-31T06:51:21.2865620Z test_decomp.py::TestDecompCPU::test_comprehensive_randn_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2865784Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2865947Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_bool PASSED [ 51%] 2023-03-31T06:51:21.2866119Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex128 PASSED [ 51%] 2023-03-31T06:51:21.2866287Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex32 PASSED [ 51%] 2023-03-31T06:51:21.2866448Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:21.2866600Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:21.2866758Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int16 PASSED [ 51%] 2023-03-31T06:51:21.2866916Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int32 PASSED [ 51%] 2023-03-31T06:51:21.2867075Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_int8 PASSED [ 51%] 2023-03-31T06:51:21.2867235Z test_decomp.py::TestDecompCPU::test_comprehensive_ravel_cpu_uint8 PASSED [ 51%] 2023-03-31T06:51:21.2867399Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2867557Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_bool PASSED [ 51%] 2023-03-31T06:51:21.2867722Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex128 PASSED [ 51%] 2023-03-31T06:51:21.2867875Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:21.2868039Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float16 PASSED [ 51%] 2023-03-31T06:51:21.2868200Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:21.2868356Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_int64 PASSED [ 51%] 2023-03-31T06:51:21.2868513Z test_decomp.py::TestDecompCPU::test_comprehensive_real_cpu_uint8 PASSED [ 51%] 2023-03-31T06:51:21.2868687Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2868857Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_bool PASSED [ 51%] 2023-03-31T06:51:21.2869031Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:21.2869186Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int16 PASSED [ 51%] 2023-03-31T06:51:21.2869353Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_int32 PASSED [ 51%] 2023-03-31T06:51:21.2869553Z test_decomp.py::TestDecompCPU::test_comprehensive_reciprocal_cpu_uint8 PASSED [ 51%] 2023-03-31T06:51:21.2869722Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2869891Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:21.2870054Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int32 PASSED [ 51%] 2023-03-31T06:51:21.2870216Z test_decomp.py::TestDecompCPU::test_comprehensive_remainder_cpu_int64 PASSED [ 51%] 2023-03-31T06:51:21.2870399Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2870595Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_complex64 PASSED [ 51%] 2023-03-31T06:51:21.2870744Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:21.2870932Z test_decomp.py::TestDecompCPU::test_comprehensive_renorm_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:21.2871099Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2871258Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float32 PASSED [ 51%] 2023-03-31T06:51:21.2871416Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_float64 PASSED [ 51%] 2023-03-31T06:51:21.2871580Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_cpu_int16 PASSED [ 51%] 2023-03-31T06:51:21.2871761Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_bfloat16 PASSED [ 51%] 2023-03-31T06:51:21.2871941Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_bool PASSED [ 51%] 2023-03-31T06:51:21.2872115Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:21.2872297Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_complex64 PASSED [ 52%] 2023-03-31T06:51:21.2872480Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float16 PASSED [ 52%] 2023-03-31T06:51:21.2872658Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:21.2872836Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int16 PASSED [ 52%] 2023-03-31T06:51:21.2873010Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int32 PASSED [ 52%] 2023-03-31T06:51:21.2873180Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:21.2873355Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_int8 PASSED [ 52%] 2023-03-31T06:51:21.2873517Z test_decomp.py::TestDecompCPU::test_comprehensive_repeat_interleave_cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:21.2873693Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_complex64 PASSED [ 52%] 2023-03-31T06:51:21.2873865Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_float16 PASSED [ 52%] 2023-03-31T06:51:21.2874035Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int16 PASSED [ 52%] 2023-03-31T06:51:21.2874202Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int32 PASSED [ 52%] 2023-03-31T06:51:21.2874367Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_as_cpu_int8 PASSED [ 52%] 2023-03-31T06:51:21.2874533Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_bool PASSED [ 52%] 2023-03-31T06:51:21.2874704Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:21.2874871Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_float64 PASSED [ 52%] 2023-03-31T06:51:21.2875023Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int64 PASSED [ 52%] 2023-03-31T06:51:21.2875184Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_int8 PASSED [ 52%] 2023-03-31T06:51:21.2875349Z test_decomp.py::TestDecompCPU::test_comprehensive_reshape_cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:21.2875543Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_bool PASSED [ 52%] 2023-03-31T06:51:21.2875712Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_float64 PASSED [ 52%] 2023-03-31T06:51:21.2875875Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int64 PASSED [ 52%] 2023-03-31T06:51:21.2876036Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_int8 PASSED [ 52%] 2023-03-31T06:51:21.2876196Z test_decomp.py::TestDecompCPU::test_comprehensive_resize__cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:21.2876381Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:21.2876570Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_bool PASSED [ 52%] 2023-03-31T06:51:21.2876737Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_float64 PASSED [ 52%] 2023-03-31T06:51:21.2876979Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int16 PASSED [ 52%] 2023-03-31T06:51:21.2877143Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int64 PASSED [ 52%] 2023-03-31T06:51:21.2877305Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_int8 PASSED [ 52%] 2023-03-31T06:51:21.2877469Z test_decomp.py::TestDecompCPU::test_comprehensive_resize_as__cpu_uint8 PASSED [ 52%] 2023-03-31T06:51:21.2877642Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_bfloat16 PASSED [ 52%] 2023-03-31T06:51:21.2877796Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_bool PASSED [ 52%] 2023-03-31T06:51:21.2877974Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_complex128 PASSED [ 52%] 2023-03-31T06:51:21.2878150Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_complex64 PASSED [ 52%] 2023-03-31T06:51:21.2878321Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float16 PASSED [ 52%] 2023-03-31T06:51:21.2878493Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_float32 PASSED [ 52%] 2023-03-31T06:51:21.2878663Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:21.2878832Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_conj_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:21.2878999Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_bool PASSED [ 53%] 2023-03-31T06:51:21.2879172Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_complex128 PASSED [ 53%] 2023-03-31T06:51:21.2879327Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:21.2879499Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:21.2879668Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:21.2879836Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:21.2880005Z test_decomp.py::TestDecompCPU::test_comprehensive_resolve_neg_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:21.2880170Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_bool PASSED [ 53%] 2023-03-31T06:51:21.2880337Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex32 PASSED [ 53%] 2023-03-31T06:51:21.2880497Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_complex64 PASSED [ 53%] 2023-03-31T06:51:21.2880647Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:21.2880806Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:21.2880964Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:21.2881126Z test_decomp.py::TestDecompCPU::test_comprehensive_roll_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:21.2881284Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_bool PASSED [ 53%] 2023-03-31T06:51:21.2881442Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:21.2881623Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:21.2881780Z test_decomp.py::TestDecompCPU::test_comprehensive_rot90_cpu_uint8 PASSED [ 53%] 2023-03-31T06:51:21.2881928Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:21.2882083Z test_decomp.py::TestDecompCPU::test_comprehensive_round_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:21.2882262Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:21.2882464Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_0_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:21.2882667Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:21.2882841Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:21.2883243Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_3_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:21.2883428Z test_decomp.py::TestDecompCPU::test_comprehensive_round_decimals_neg_3_cpu_float64 PASSED [ 53%] 2023-03-31T06:51:21.2883582Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_complex64 PASSED [ 53%] 2023-03-31T06:51:21.2883746Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_float32 PASSED [ 53%] 2023-03-31T06:51:21.2883906Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:21.2884063Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:21.2884225Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int64 PASSED [ 53%] 2023-03-31T06:51:21.2884386Z test_decomp.py::TestDecompCPU::test_comprehensive_rsqrt_cpu_int8 PASSED [ 53%] 2023-03-31T06:51:21.2884548Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_bfloat16 PASSED [ 53%] 2023-03-31T06:51:21.2884712Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_float16 PASSED [ 53%] 2023-03-31T06:51:21.2884869Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int16 PASSED [ 53%] 2023-03-31T06:51:21.2885011Z test_decomp.py::TestDecompCPU::test_comprehensive_rsub_cpu_int32 PASSED [ 53%] 2023-03-31T06:51:21.2885189Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex128 PASSED [ 53%] 2023-03-31T06:51:21.2885363Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex32 PASSED [ 54%] 2023-03-31T06:51:21.2885538Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_complex64 PASSED [ 54%] 2023-03-31T06:51:21.2885711Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:21.2885885Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:21.2886055Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int32 PASSED [ 54%] 2023-03-31T06:51:21.2886227Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2886384Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:21.2886552Z test_decomp.py::TestDecompCPU::test_comprehensive_scalar_tensor_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:21.2886719Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_bool PASSED [ 54%] 2023-03-31T06:51:21.2886891Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_complex64 PASSED [ 54%] 2023-03-31T06:51:21.2887063Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_float64 PASSED [ 54%] 2023-03-31T06:51:21.2887232Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:21.2887398Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2887560Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_add_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:21.2887714Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_bool PASSED [ 54%] 2023-03-31T06:51:21.2887944Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:21.2888106Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int32 PASSED [ 54%] 2023-03-31T06:51:21.2888267Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2888429Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:21.2888589Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:21.2888804Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_bfloat16 PASSED [ 54%] 2023-03-31T06:51:21.2889013Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_bool PASSED [ 54%] 2023-03-31T06:51:21.2889196Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_float16 PASSED [ 54%] 2023-03-31T06:51:21.2889395Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:21.2889573Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int32 PASSED [ 54%] 2023-03-31T06:51:21.2889746Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2889923Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:21.2890095Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amax_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:21.2890279Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float16 PASSED [ 54%] 2023-03-31T06:51:21.2890459Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_float32 PASSED [ 54%] 2023-03-31T06:51:21.2890634Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int16 PASSED [ 54%] 2023-03-31T06:51:21.2890799Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2890977Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_amin_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:21.2891155Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_float16 PASSED [ 54%] 2023-03-31T06:51:21.2891330Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int64 PASSED [ 54%] 2023-03-31T06:51:21.2891504Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_int8 PASSED [ 54%] 2023-03-31T06:51:21.2891679Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_mean_cpu_uint8 PASSED [ 54%] 2023-03-31T06:51:21.2891852Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_bool PASSED [ 54%] 2023-03-31T06:51:21.2892032Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:21.2892212Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:21.2892376Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:21.2892554Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_int8 PASSED [ 55%] 2023-03-31T06:51:21.2892727Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_prod_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:21.2892901Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_bool PASSED [ 55%] 2023-03-31T06:51:21.2893076Z test_decomp.py::TestDecompCPU::test_comprehensive_scatter_reduce_sum_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2893251Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_bfloat16 PASSED [ 55%] 2023-03-31T06:51:21.2893425Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:21.2893594Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2893749Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:21.2893948Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:21.2894118Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_int8 PASSED [ 55%] 2023-03-31T06:51:21.2894288Z test_decomp.py::TestDecompCPU::test_comprehensive_searchsorted_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:21.2894456Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex32 PASSED [ 55%] 2023-03-31T06:51:21.2894653Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_complex64 PASSED [ 55%] 2023-03-31T06:51:21.2894843Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float16 PASSED [ 55%] 2023-03-31T06:51:21.2895002Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:21.2895165Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2895340Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:21.2895505Z test_decomp.py::TestDecompCPU::test_comprehensive_select_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:21.2895679Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float16 PASSED [ 55%] 2023-03-31T06:51:21.2895853Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:21.2896023Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2896194Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:21.2896366Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int64 PASSED [ 55%] 2023-03-31T06:51:21.2896534Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_int8 PASSED [ 55%] 2023-03-31T06:51:21.2896695Z test_decomp.py::TestDecompCPU::test_comprehensive_select_scatter_cpu_uint8 PASSED [ 55%] 2023-03-31T06:51:21.2896859Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex128 PASSED [ 55%] 2023-03-31T06:51:21.2897024Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex32 PASSED [ 55%] 2023-03-31T06:51:21.2897187Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_complex64 PASSED [ 55%] 2023-03-31T06:51:21.2897346Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_float16 PASSED [ 55%] 2023-03-31T06:51:21.2897503Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2897664Z test_decomp.py::TestDecompCPU::test_comprehensive_sgn_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:21.2897833Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_complex128 PASSED [ 55%] 2023-03-31T06:51:21.2897981Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float32 PASSED [ 55%] 2023-03-31T06:51:21.2898144Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_float64 PASSED [ 55%] 2023-03-31T06:51:21.2898303Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int16 PASSED [ 55%] 2023-03-31T06:51:21.2898463Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int32 PASSED [ 55%] 2023-03-31T06:51:21.2898622Z test_decomp.py::TestDecompCPU::test_comprehensive_short_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:21.2898788Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_bool PASSED [ 56%] 2023-03-31T06:51:21.2898948Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:21.2899108Z test_decomp.py::TestDecompCPU::test_comprehensive_sigmoid_cpu_uint8 PASSED [ 56%] 2023-03-31T06:51:21.2899257Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_bfloat16 PASSED [ 56%] 2023-03-31T06:51:21.2899416Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_bool PASSED [ 56%] 2023-03-31T06:51:21.2899576Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float16 PASSED [ 56%] 2023-03-31T06:51:21.2899739Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2899925Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2900083Z test_decomp.py::TestDecompCPU::test_comprehensive_sign_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:21.2900271Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_bartlett_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2900459Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_bartlett_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2900667Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_blackman_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2900860Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_cosine_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2901050Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_gaussian_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2901273Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_general_hamming_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2901460Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hamming_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2901638Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hamming_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2901817Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_hann_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2901999Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_kaiser_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2902184Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_kaiser_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2902365Z test_decomp.py::TestDecompCPU::test_comprehensive_signal_windows_nuttall_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2902519Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float16 PASSED [ 56%] 2023-03-31T06:51:21.2902681Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2902841Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2903007Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int32 PASSED [ 56%] 2023-03-31T06:51:21.2903172Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:21.2903419Z test_decomp.py::TestDecompCPU::test_comprehensive_signbit_cpu_uint8 PASSED [ 56%] 2023-03-31T06:51:21.2903618Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2903780Z test_decomp.py::TestDecompCPU::test_comprehensive_sin_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:21.2903930Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2904087Z test_decomp.py::TestDecompCPU::test_comprehensive_sinc_cpu_int8 PASSED [ 56%] 2023-03-31T06:51:21.2904251Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_complex128 PASSED [ 56%] 2023-03-31T06:51:21.2904412Z test_decomp.py::TestDecompCPU::test_comprehensive_sinh_cpu_float32 PASSED [ 56%] 2023-03-31T06:51:21.2904582Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_complex128 PASSED [ 56%] 2023-03-31T06:51:21.2904744Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_float64 PASSED [ 56%] 2023-03-31T06:51:21.2904901Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int32 PASSED [ 56%] 2023-03-31T06:51:21.2905059Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int64 PASSED [ 56%] 2023-03-31T06:51:21.2905204Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:21.2905361Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:21.2905531Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_bool PASSED [ 57%] 2023-03-31T06:51:21.2905703Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float16 PASSED [ 57%] 2023-03-31T06:51:21.2905878Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:21.2906092Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:21.2906261Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:21.2906431Z test_decomp.py::TestDecompCPU::test_comprehensive_slice_scatter_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:21.2906598Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_bfloat16 PASSED [ 57%] 2023-03-31T06:51:21.2906748Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:21.2906967Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_complex128 PASSED [ 57%] 2023-03-31T06:51:21.2907173Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:21.2907350Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int16 PASSED [ 57%] 2023-03-31T06:51:21.2907553Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:21.2907725Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:21.2907896Z test_decomp.py::TestDecompCPU::test_comprehensive_softmax_with_dtype_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:21.2908054Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_bool PASSED [ 57%] 2023-03-31T06:51:21.2908202Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_float16 PASSED [ 57%] 2023-03-31T06:51:21.2908363Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:21.2908524Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:21.2908680Z test_decomp.py::TestDecompCPU::test_comprehensive_sort_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:21.2908856Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_mm_reduce_cpu_bfloat16 PASSED [ 57%] 2023-03-31T06:51:21.2909048Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_complex128 PASSED [ 57%] 2023-03-31T06:51:21.2909233Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_complex64 PASSED [ 57%] 2023-03-31T06:51:21.2909416Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:21.2909598Z test_decomp.py::TestDecompCPU::test_comprehensive_sparse_sampled_addmm_cpu_float64 PASSED [ 57%] 2023-03-31T06:51:21.2909757Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_bool PASSED [ 57%] 2023-03-31T06:51:21.2909928Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:21.2910101Z test_decomp.py::TestDecompCPU::test_comprehensive_special_airy_ai_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:21.2910277Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_float32 PASSED [ 57%] 2023-03-31T06:51:21.2910451Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_int64 PASSED [ 57%] 2023-03-31T06:51:21.2910624Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j0_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:21.2910797Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_bool PASSED [ 57%] 2023-03-31T06:51:21.2910971Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_float64 PASSED [ 57%] 2023-03-31T06:51:21.2911129Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int16 PASSED [ 57%] 2023-03-31T06:51:21.2911303Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int32 PASSED [ 57%] 2023-03-31T06:51:21.2911475Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_int8 PASSED [ 57%] 2023-03-31T06:51:21.2911646Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_j1_cpu_uint8 PASSED [ 57%] 2023-03-31T06:51:21.2911820Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y0_cpu_int16 PASSED [ 57%] 2023-03-31T06:51:21.2912019Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_bool PASSED [ 58%] 2023-03-31T06:51:21.2912188Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int32 PASSED [ 58%] 2023-03-31T06:51:21.2912356Z test_decomp.py::TestDecompCPU::test_comprehensive_special_bessel_y1_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:21.2912551Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_bool PASSED [ 58%] 2023-03-31T06:51:21.2912769Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:21.2912990Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2913182Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:21.2913402Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_t_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:21.2913594Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_bool PASSED [ 58%] 2023-03-31T06:51:21.2913791Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:21.2913980Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_u_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2914501Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2914882Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2915253Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2915639Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2916018Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_v_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2916400Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2916778Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2917157Z test_decomp.py::TestDecompCPU::test_comprehensive_special_chebyshev_polynomial_w_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 58%] 2023-03-31T06:51:21.2917333Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:21.2917505Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:21.2917676Z test_decomp.py::TestDecompCPU::test_comprehensive_special_entr_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:21.2917849Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:21.2918020Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2918180Z test_decomp.py::TestDecompCPU::test_comprehensive_special_erfcx_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:21.2918373Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2918564Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:21.2918756Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_h_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:21.2918989Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_float32 PASSED [ 58%] 2023-03-31T06:51:21.2919184Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_float64 PASSED [ 58%] 2023-03-31T06:51:21.2919376Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2919565Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int64 PASSED [ 58%] 2023-03-31T06:51:21.2919783Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:21.2919994Z test_decomp.py::TestDecompCPU::test_comprehensive_special_hermite_polynomial_he_cpu_uint8 PASSED [ 58%] 2023-03-31T06:51:21.2920163Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_bool PASSED [ 58%] 2023-03-31T06:51:21.2920357Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int16 PASSED [ 58%] 2023-03-31T06:51:21.2920527Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i0e_cpu_int8 PASSED [ 58%] 2023-03-31T06:51:21.2920696Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_float32 PASSED [ 58%] 2023-03-31T06:51:21.2920860Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:21.2921027Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:21.2921188Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2921337Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1_cpu_uint8 PASSED [ 59%] 2023-03-31T06:51:21.2921508Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2921674Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:21.2921837Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2922001Z test_decomp.py::TestDecompCPU::test_comprehensive_special_i1e_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:21.2922192Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_bool PASSED [ 59%] 2023-03-31T06:51:21.2922389Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2922579Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:21.2922774Z test_decomp.py::TestDecompCPU::test_comprehensive_special_laguerre_polynomial_l_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:21.2923372Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:21.2923757Z test_decomp.py::TestDecompCPU::test_comprehensive_special_legendre_polynomial_p_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 59%] 2023-03-31T06:51:21.2923940Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2924111Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2924282Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:21.2924450Z test_decomp.py::TestDecompCPU::test_comprehensive_special_log_ndtr_cpu_uint8 PASSED [ 59%] 2023-03-31T06:51:21.2924640Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2924826Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:21.2925016Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:21.2925188Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2925436Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:21.2925625Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i0_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:21.2925810Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_bool PASSED [ 59%] 2023-03-31T06:51:21.2925995Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2926214Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int16 PASSED [ 59%] 2023-03-31T06:51:21.2926467Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2926647Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:21.2926877Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_i1_cpu_uint8 PASSED [ 59%] 2023-03-31T06:51:21.2927052Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_float32 PASSED [ 59%] 2023-03-31T06:51:21.2927242Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_float64 PASSED [ 59%] 2023-03-31T06:51:21.2927427Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:21.2927610Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k0_cpu_int8 PASSED [ 59%] 2023-03-31T06:51:21.2927794Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_bool PASSED [ 59%] 2023-03-31T06:51:21.2927973Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int32 PASSED [ 59%] 2023-03-31T06:51:21.2928151Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_int64 PASSED [ 59%] 2023-03-31T06:51:21.2928331Z test_decomp.py::TestDecompCPU::test_comprehensive_special_modified_bessel_k1_cpu_uint8 PASSED [ 59%] 2023-03-31T06:51:21.2928564Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_float64 SKIPPED (special.ndtr in torch.float64 not supported) [ 59%] 2023-03-31T06:51:21.2928775Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtr_cpu_int8 SKIPPED (special.ndtr in torch.int8 not supported) [ 60%] 2023-03-31T06:51:21.2928950Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:21.2929121Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:21.2929292Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:21.2929463Z test_decomp.py::TestDecompCPU::test_comprehensive_special_ndtri_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:21.2929673Z test_decomp.py::TestDecompCPU::test_comprehensive_special_polygamma_special_polygamma_n_0_cpu_bfloat16 PASSED [ 60%] 2023-03-31T06:51:21.2929875Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:21.2930070Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_float64 PASSED [ 60%] 2023-03-31T06:51:21.2930265Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int16 PASSED [ 60%] 2023-03-31T06:51:21.2930451Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:21.2930646Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_int8 PASSED [ 60%] 2023-03-31T06:51:21.2930840Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k0_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:21.2931030Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_bool PASSED [ 60%] 2023-03-31T06:51:21.2931229Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_float32 PASSED [ 60%] 2023-03-31T06:51:21.2931448Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_int32 PASSED [ 60%] 2023-03-31T06:51:21.2931645Z test_decomp.py::TestDecompCPU::test_comprehensive_special_scaled_modified_bessel_k1_cpu_uint8 PASSED [ 60%] 2023-03-31T06:51:21.2932052Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2932471Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2932873Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2933296Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_t_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2933691Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2934081Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2934470Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2934864Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2935253Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2935642Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_u_cpu_uint8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2936029Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2936421Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2936809Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2937201Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2937589Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_v_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2937974Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_bool SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2938356Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2938744Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int16 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2939159Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2939547Z test_decomp.py::TestDecompCPU::test_comprehensive_special_shifted_chebyshev_polynomial_w_cpu_int8 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 60%] 2023-03-31T06:51:21.2939737Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_bool PASSED [ 60%] 2023-03-31T06:51:21.2939929Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_float64 PASSED [ 60%] 2023-03-31T06:51:21.2940143Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:21.2940350Z test_decomp.py::TestDecompCPU::test_comprehensive_special_spherical_bessel_j0_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:21.2940526Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:21.2940728Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:21.2940890Z test_decomp.py::TestDecompCPU::test_comprehensive_special_xlog1py_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:21.2941060Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:21.2941230Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:21.2941399Z test_decomp.py::TestDecompCPU::test_comprehensive_special_zeta_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:21.2941565Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_bool PASSED [ 61%] 2023-03-31T06:51:21.2941735Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:21.2941901Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:21.2942063Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:21.2942215Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:21.2942378Z test_decomp.py::TestDecompCPU::test_comprehensive_split_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:21.2942555Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:21.2942726Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_bool PASSED [ 61%] 2023-03-31T06:51:21.2942907Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:21.2943085Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:21.2943260Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:21.2943526Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int16 PASSED [ 61%] 2023-03-31T06:51:21.2943701Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_int64 PASSED [ 61%] 2023-03-31T06:51:21.2943861Z test_decomp.py::TestDecompCPU::test_comprehensive_split_list_args_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:21.2944038Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:21.2944219Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex128 PASSED [ 61%] 2023-03-31T06:51:21.2944399Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_complex64 PASSED [ 61%] 2023-03-31T06:51:21.2944576Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float16 PASSED [ 61%] 2023-03-31T06:51:21.2944752Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:21.2944927Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:21.2945099Z test_decomp.py::TestDecompCPU::test_comprehensive_split_with_sizes_cpu_int8 PASSED [ 61%] 2023-03-31T06:51:21.2945251Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:21.2945460Z test_decomp.py::TestDecompCPU::test_comprehensive_sqrt_cpu_bool PASSED [ 61%] 2023-03-31T06:51:21.2945631Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:21.2945796Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:21.2945955Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:21.2946117Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_int32 PASSED [ 61%] 2023-03-31T06:51:21.2946304Z test_decomp.py::TestDecompCPU::test_comprehensive_square_cpu_uint8 PASSED [ 61%] 2023-03-31T06:51:21.2946496Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float32 PASSED [ 61%] 2023-03-31T06:51:21.2946646Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_cpu_float64 PASSED [ 61%] 2023-03-31T06:51:21.2946854Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_bfloat16 PASSED [ 61%] 2023-03-31T06:51:21.2947039Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:21.2947220Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_complex32 PASSED [ 62%] 2023-03-31T06:51:21.2947394Z test_decomp.py::TestDecompCPU::test_comprehensive_squeeze_multiple_cpu_int8 PASSED [ 62%] 2023-03-31T06:51:21.2947560Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:21.2947729Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:21.2947894Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2948056Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int32 PASSED [ 62%] 2023-03-31T06:51:21.2948202Z test_decomp.py::TestDecompCPU::test_comprehensive_stack_cpu_int64 PASSED [ 62%] 2023-03-31T06:51:21.2948367Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:21.2948529Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:21.2948687Z test_decomp.py::TestDecompCPU::test_comprehensive_std_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2948862Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:21.2949027Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2949208Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:21.2949393Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:21.2949559Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:21.2949734Z test_decomp.py::TestDecompCPU::test_comprehensive_std_mean_unbiased_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2949914Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:21.2950090Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:21.2950260Z test_decomp.py::TestDecompCPU::test_comprehensive_std_unbiased_cpu_float32 PASSED [ 62%] 2023-03-31T06:51:21.2950427Z test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:21.2950593Z test_decomp.py::TestDecompCPU::test_comprehensive_stft_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:21.2950756Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_bfloat16 PASSED [ 62%] 2023-03-31T06:51:21.2950910Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:21.2951072Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex32 PASSED [ 62%] 2023-03-31T06:51:21.2951233Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_complex64 PASSED [ 62%] 2023-03-31T06:51:21.2951395Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:21.2951582Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2951740Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int16 PASSED [ 62%] 2023-03-31T06:51:21.2951895Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int32 PASSED [ 62%] 2023-03-31T06:51:21.2952049Z test_decomp.py::TestDecompCPU::test_comprehensive_sub_cpu_int8 PASSED [ 62%] 2023-03-31T06:51:21.2952240Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float16 PASSED [ 62%] 2023-03-31T06:51:21.2952425Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_float64 PASSED [ 62%] 2023-03-31T06:51:21.2952578Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int32 PASSED [ 62%] 2023-03-31T06:51:21.2952732Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_int8 PASSED [ 62%] 2023-03-31T06:51:21.2952916Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_cpu_uint8 PASSED [ 62%] 2023-03-31T06:51:21.2953088Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_bool PASSED [ 62%] 2023-03-31T06:51:21.2953262Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:21.2953428Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:21.2953589Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:21.2953741Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:21.2953906Z test_decomp.py::TestDecompCPU::test_comprehensive_sum_to_size_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:21.2954071Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:21.2954229Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:21.2954390Z test_decomp.py::TestDecompCPU::test_comprehensive_svd_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:21.2954557Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:21.2954713Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_float64 PASSED [ 63%] 2023-03-31T06:51:21.2954865Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:21.2955007Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:21.2955160Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:21.2955312Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:21.2955466Z test_decomp.py::TestDecompCPU::test_comprehensive_t_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:21.2955644Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:21.2955822Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:21.2955997Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float16 PASSED [ 63%] 2023-03-31T06:51:21.2956169Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:21.2956326Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:21.2956492Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:21.2956660Z test_decomp.py::TestDecompCPU::test_comprehensive_take_along_dim_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:21.2956828Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:21.2956987Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:21.2957143Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:21.2957303Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:21.2957490Z test_decomp.py::TestDecompCPU::test_comprehensive_take_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:21.2957640Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:21.2957804Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:21.2957963Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:21.2958115Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int32 PASSED [ 63%] 2023-03-31T06:51:21.2958299Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int64 PASSED [ 63%] 2023-03-31T06:51:21.2958478Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_int8 PASSED [ 63%] 2023-03-31T06:51:21.2958629Z test_decomp.py::TestDecompCPU::test_comprehensive_tan_cpu_uint8 PASSED [ 63%] 2023-03-31T06:51:21.2958823Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_bfloat16 PASSED [ 63%] 2023-03-31T06:51:21.2958984Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_bool PASSED [ 63%] 2023-03-31T06:51:21.2959137Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_complex128 PASSED [ 63%] 2023-03-31T06:51:21.2959300Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_complex64 PASSED [ 63%] 2023-03-31T06:51:21.2959461Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_float32 PASSED [ 63%] 2023-03-31T06:51:21.2959619Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int16 PASSED [ 63%] 2023-03-31T06:51:21.2959777Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:21.2959934Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:21.2960090Z test_decomp.py::TestDecompCPU::test_comprehensive_tanh_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:21.2960258Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_bool PASSED [ 64%] 2023-03-31T06:51:21.2960419Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float16 PASSED [ 64%] 2023-03-31T06:51:21.2960594Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2960763Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:21.2960931Z test_decomp.py::TestDecompCPU::test_comprehensive_tensor_split_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:21.2961102Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:21.2961270Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2961436Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:21.2961596Z test_decomp.py::TestDecompCPU::test_comprehensive_tensordot_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:21.2961747Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:21.2961905Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_bool PASSED [ 64%] 2023-03-31T06:51:21.2962068Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float16 PASSED [ 64%] 2023-03-31T06:51:21.2962226Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2962387Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:21.2962543Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int32 PASSED [ 64%] 2023-03-31T06:51:21.2962702Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:21.2962860Z test_decomp.py::TestDecompCPU::test_comprehensive_tile_cpu_uint8 PASSED [ 64%] 2023-03-31T06:51:21.2963003Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:21.2963349Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_bool PASSED [ 64%] 2023-03-31T06:51:21.2963514Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_complex128 PASSED [ 64%] 2023-03-31T06:51:21.2963730Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_complex64 PASSED [ 64%] 2023-03-31T06:51:21.2963889Z test_decomp.py::TestDecompCPU::test_comprehensive_to_cpu_float16 PASSED [ 64%] 2023-03-31T06:51:21.2964055Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float16 PASSED [ 64%] 2023-03-31T06:51:21.2964217Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2964384Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_float64 PASSED [ 64%] 2023-03-31T06:51:21.2964569Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:21.2964769Z test_decomp.py::TestDecompCPU::test_comprehensive_to_sparse_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:21.2964929Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_bfloat16 PASSED [ 64%] 2023-03-31T06:51:21.2965128Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2965286Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:21.2965440Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int64 PASSED [ 64%] 2023-03-31T06:51:21.2965596Z test_decomp.py::TestDecompCPU::test_comprehensive_topk_cpu_int8 PASSED [ 64%] 2023-03-31T06:51:21.2965762Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_complex64 PASSED [ 64%] 2023-03-31T06:51:21.2965926Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_float32 PASSED [ 64%] 2023-03-31T06:51:21.2966072Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int16 PASSED [ 64%] 2023-03-31T06:51:21.2966231Z test_decomp.py::TestDecompCPU::test_comprehensive_trace_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:21.2966397Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_bool PASSED [ 65%] 2023-03-31T06:51:21.2966574Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex32 PASSED [ 65%] 2023-03-31T06:51:21.2966750Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:21.2966917Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:21.2967081Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:21.2967245Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_int16 PASSED [ 65%] 2023-03-31T06:51:21.2967394Z test_decomp.py::TestDecompCPU::test_comprehensive_transpose_cpu_uint8 PASSED [ 65%] 2023-03-31T06:51:21.2967567Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:21.2967742Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:21.2967913Z test_decomp.py::TestDecompCPU::test_comprehensive_trapezoid_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:21.2968079Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:21.2968243Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:21.2968401Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:21.2968558Z test_decomp.py::TestDecompCPU::test_comprehensive_trapz_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:21.2968703Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_bool PASSED [ 65%] 2023-03-31T06:51:21.2968871Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:21.2969028Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_int16 PASSED [ 65%] 2023-03-31T06:51:21.2969186Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_cpu_uint8 PASSED [ 65%] 2023-03-31T06:51:21.2969355Z test_decomp.py::TestDecompCPU::test_comprehensive_tril_indices_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:21.2969519Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:21.2969715Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:21.2969881Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_complex64 PASSED [ 65%] 2023-03-31T06:51:21.2970030Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:21.2970188Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_cpu_uint8 PASSED [ 65%] 2023-03-31T06:51:21.2970357Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_indices_cpu_int32 PASSED [ 65%] 2023-03-31T06:51:21.2970556Z test_decomp.py::TestDecompCPU::test_comprehensive_triu_indices_cpu_int64 PASSED [ 65%] 2023-03-31T06:51:21.2970753Z test_decomp.py::TestDecompCPU::test_comprehensive_true_divide_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:21.2970915Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:21.2971112Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:21.2971272Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int16 PASSED [ 65%] 2023-03-31T06:51:21.2971419Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_int8 PASSED [ 65%] 2023-03-31T06:51:21.2971579Z test_decomp.py::TestDecompCPU::test_comprehensive_trunc_cpu_uint8 PASSED [ 65%] 2023-03-31T06:51:21.2971742Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_bfloat16 PASSED [ 65%] 2023-03-31T06:51:21.2971901Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_bool PASSED [ 65%] 2023-03-31T06:51:21.2972069Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_complex128 PASSED [ 65%] 2023-03-31T06:51:21.2972233Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float16 PASSED [ 65%] 2023-03-31T06:51:21.2972393Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float32 PASSED [ 65%] 2023-03-31T06:51:21.2972551Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_float64 PASSED [ 65%] 2023-03-31T06:51:21.2972717Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:21.2972866Z test_decomp.py::TestDecompCPU::test_comprehensive_unbind_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:21.2973038Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_complex64 PASSED [ 66%] 2023-03-31T06:51:21.2973207Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:21.2973370Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2973536Z test_decomp.py::TestDecompCPU::test_comprehensive_unflatten_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:21.2973706Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_bool PASSED [ 66%] 2023-03-31T06:51:21.2973880Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex32 PASSED [ 66%] 2023-03-31T06:51:21.2974048Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_complex64 PASSED [ 66%] 2023-03-31T06:51:21.2974207Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2974378Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:21.2974544Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:21.2974713Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:21.2974881Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_copy_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:21.2975047Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:21.2975208Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_bool PASSED [ 66%] 2023-03-31T06:51:21.2975376Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex128 PASSED [ 66%] 2023-03-31T06:51:21.2975528Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_complex32 PASSED [ 66%] 2023-03-31T06:51:21.2975695Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float16 PASSED [ 66%] 2023-03-31T06:51:21.2975889Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2976050Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:21.2976214Z test_decomp.py::TestDecompCPU::test_comprehensive_unfold_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:21.2976384Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_complex128 PASSED [ 66%] 2023-03-31T06:51:21.2976613Z test_decomp.py::TestDecompCPU::test_comprehensive_uniform_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2976825Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:21.2977004Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2977199Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:21.2977380Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:21.2977555Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int64 PASSED [ 66%] 2023-03-31T06:51:21.2977731Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_consecutive_cpu_int8 PASSED [ 66%] 2023-03-31T06:51:21.2977897Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_bfloat16 PASSED [ 66%] 2023-03-31T06:51:21.2978059Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float32 PASSED [ 66%] 2023-03-31T06:51:21.2978219Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_float64 PASSED [ 66%] 2023-03-31T06:51:21.2978382Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:21.2978530Z test_decomp.py::TestDecompCPU::test_comprehensive_unique_cpu_int32 PASSED [ 66%] 2023-03-31T06:51:21.2978699Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_int16 PASSED [ 66%] 2023-03-31T06:51:21.2978866Z test_decomp.py::TestDecompCPU::test_comprehensive_unsafe_split_cpu_uint8 PASSED [ 66%] 2023-03-31T06:51:21.2979033Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_bool PASSED [ 66%] 2023-03-31T06:51:21.2979204Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex32 PASSED [ 67%] 2023-03-31T06:51:21.2979373Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:21.2979540Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:21.2979710Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:21.2979862Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int32 PASSED [ 67%] 2023-03-31T06:51:21.2980021Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_int64 PASSED [ 67%] 2023-03-31T06:51:21.2980180Z test_decomp.py::TestDecompCPU::test_comprehensive_unsqueeze_cpu_uint8 PASSED [ 67%] 2023-03-31T06:51:21.2980346Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2980504Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:21.2980662Z test_decomp.py::TestDecompCPU::test_comprehensive_var_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:21.2980825Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:21.2980992Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2981161Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:21.2981313Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:21.2981472Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:21.2981656Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2981862Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:21.2982042Z test_decomp.py::TestDecompCPU::test_comprehensive_var_mean_unbiased_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:21.2982216Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2982391Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:21.2982586Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:21.2982767Z test_decomp.py::TestDecompCPU::test_comprehensive_var_unbiased_cpu_float64 PASSED [ 67%] 2023-03-31T06:51:21.2982933Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2983093Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int16 PASSED [ 67%] 2023-03-31T06:51:21.2983276Z test_decomp.py::TestDecompCPU::test_comprehensive_vdot_cpu_int64 PASSED [ 67%] 2023-03-31T06:51:21.2983544Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2983718Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:21.2983880Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:21.2984041Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_int64 PASSED [ 67%] 2023-03-31T06:51:21.2984190Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_cpu_uint8 PASSED [ 67%] 2023-03-31T06:51:21.2984366Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_real_cpu_complex128 PASSED [ 67%] 2023-03-31T06:51:21.2984539Z test_decomp.py::TestDecompCPU::test_comprehensive_view_as_real_cpu_complex64 PASSED [ 67%] 2023-03-31T06:51:21.2984705Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:21.2984872Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_bool PASSED [ 67%] 2023-03-31T06:51:21.2985041Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_float32 PASSED [ 67%] 2023-03-31T06:51:21.2985205Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_int8 PASSED [ 67%] 2023-03-31T06:51:21.2985368Z test_decomp.py::TestDecompCPU::test_comprehensive_view_copy_cpu_uint8 PASSED [ 67%] 2023-03-31T06:51:21.2985529Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_bfloat16 PASSED [ 67%] 2023-03-31T06:51:21.2985678Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float16 PASSED [ 67%] 2023-03-31T06:51:21.2985839Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_float64 PASSED [ 68%] 2023-03-31T06:51:21.2985995Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:21.2986153Z test_decomp.py::TestDecompCPU::test_comprehensive_view_cpu_uint8 PASSED [ 68%] 2023-03-31T06:51:21.2986319Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:51:21.2986492Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_complex64 PASSED [ 68%] 2023-03-31T06:51:21.2986657Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:21.2986818Z test_decomp.py::TestDecompCPU::test_comprehensive_vsplit_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:21.2986967Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:51:21.2987126Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_bool PASSED [ 68%] 2023-03-31T06:51:21.2987297Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex128 PASSED [ 68%] 2023-03-31T06:51:21.2987467Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_complex32 PASSED [ 68%] 2023-03-31T06:51:21.2987631Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float16 PASSED [ 68%] 2023-03-31T06:51:21.2987793Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:21.2987985Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_float64 PASSED [ 68%] 2023-03-31T06:51:21.2988150Z test_decomp.py::TestDecompCPU::test_comprehensive_vstack_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:21.2988300Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float16 PASSED [ 68%] 2023-03-31T06:51:21.2988462Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_float64 PASSED [ 68%] 2023-03-31T06:51:21.2988621Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:21.2988808Z test_decomp.py::TestDecompCPU::test_comprehensive_where_cpu_uint8 PASSED [ 68%] 2023-03-31T06:51:21.2988995Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:51:21.2989156Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:21.2989343Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_float64 PASSED [ 68%] 2023-03-31T06:51:21.2989507Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:21.2989653Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_int8 PASSED [ 68%] 2023-03-31T06:51:21.2989811Z test_decomp.py::TestDecompCPU::test_comprehensive_xlogy_cpu_uint8 PASSED [ 68%] 2023-03-31T06:51:21.2990030Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_complex64 SKIPPED (zero_ in torch.complex64 not supported) [ 68%] 2023-03-31T06:51:21.2990237Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int16 SKIPPED (zero_ in torch.int16 not supported) [ 68%] 2023-03-31T06:51:21.2990434Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int32 SKIPPED (zero_ in torch.int32 not supported) [ 68%] 2023-03-31T06:51:21.2990636Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int64 SKIPPED (zero_ in torch.int64 not supported) [ 68%] 2023-03-31T06:51:21.2990838Z test_decomp.py::TestDecompCPU::test_comprehensive_zero__cpu_int8 SKIPPED (zero_ in torch.int8 not supported) [ 68%] 2023-03-31T06:51:21.2991005Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_bfloat16 PASSED [ 68%] 2023-03-31T06:51:21.2991164Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_bool PASSED [ 68%] 2023-03-31T06:51:21.2991322Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex128 PASSED [ 68%] 2023-03-31T06:51:21.2991489Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex32 PASSED [ 68%] 2023-03-31T06:51:21.2991651Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_complex64 PASSED [ 68%] 2023-03-31T06:51:21.2991815Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_float32 PASSED [ 68%] 2023-03-31T06:51:21.2991977Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int16 PASSED [ 68%] 2023-03-31T06:51:21.2992134Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int32 PASSED [ 68%] 2023-03-31T06:51:21.2992293Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:21.2992466Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:21.2992625Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex32 PASSED [ 69%] 2023-03-31T06:51:21.2992799Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:21.2992966Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float16 PASSED [ 69%] 2023-03-31T06:51:21.2993135Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:21.2993301Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:21.2993469Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:21.2993632Z test_decomp.py::TestDecompCPU::test_comprehensive_zeros_like_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:21.2993805Z test_decomp.py::TestDecompCPU::test_quick__native_batch_norm_legit_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:21.2994077Z test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_bfloat16 SKIPPED (_softmax_backward_data in torch.bfloat16 not supported) [ 69%] 2023-03-31T06:51:21.2994239Z test_decomp.py::TestDecompCPU::test_quick__softmax_backward_data_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:21.2994396Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex32 PASSED [ 69%] 2023-03-31T06:51:21.2994547Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:21.2994734Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:21.2994908Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int16 PASSED [ 69%] 2023-03-31T06:51:21.2995054Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int32 PASSED [ 69%] 2023-03-31T06:51:21.2995229Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:21.2995374Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:21.2995509Z test_decomp.py::TestDecompCPU::test_quick_abs_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:21.2995659Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:21.2995809Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:21.2995956Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int16 PASSED [ 69%] 2023-03-31T06:51:21.2996104Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:21.2996250Z test_decomp.py::TestDecompCPU::test_quick_acos_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:21.2996398Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:21.2996544Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:21.2996679Z test_decomp.py::TestDecompCPU::test_quick_acosh_cpu_int8 PASSED [ 69%] 2023-03-31T06:51:21.2996828Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:21.2996973Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_bool PASSED [ 69%] 2023-03-31T06:51:21.2997120Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_float16 PASSED [ 69%] 2023-03-31T06:51:21.2997266Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_float32 PASSED [ 69%] 2023-03-31T06:51:21.2997411Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_float64 PASSED [ 69%] 2023-03-31T06:51:21.2997558Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_int32 PASSED [ 69%] 2023-03-31T06:51:21.2997703Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_int64 PASSED [ 69%] 2023-03-31T06:51:21.2997834Z test_decomp.py::TestDecompCPU::test_quick_add_cpu_uint8 PASSED [ 69%] 2023-03-31T06:51:21.2997985Z test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_bfloat16 PASSED [ 69%] 2023-03-31T06:51:21.2998138Z test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_complex128 PASSED [ 69%] 2023-03-31T06:51:21.2998289Z test_decomp.py::TestDecompCPU::test_quick_addcdiv_cpu_complex64 PASSED [ 69%] 2023-03-31T06:51:21.2998441Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:21.2998592Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:21.2998739Z test_decomp.py::TestDecompCPU::test_quick_addcmul_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.2998891Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:21.2999029Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:21.2999174Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.2999318Z test_decomp.py::TestDecompCPU::test_quick_addmm_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:21.2999488Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:21.2999684Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:21.2999848Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:21.3000012Z test_decomp.py::TestDecompCPU::test_quick_addmm_decomposed_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.3000168Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:21.3000343Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:21.3000504Z test_decomp.py::TestDecompCPU::test_quick_addmv_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:21.3000653Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:21.3000804Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_complex128 PASSED [ 70%] 2023-03-31T06:51:21.3000983Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:21.3001135Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:21.3001280Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.3001429Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_int8 PASSED [ 70%] 2023-03-31T06:51:21.3001575Z test_decomp.py::TestDecompCPU::test_quick_addr_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:21.3001708Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_bool PASSED [ 70%] 2023-03-31T06:51:21.3001858Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_complex64 PASSED [ 70%] 2023-03-31T06:51:21.3002006Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_float16 PASSED [ 70%] 2023-03-31T06:51:21.3002153Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:21.3002300Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.3002444Z test_decomp.py::TestDecompCPU::test_quick_all_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:21.3002591Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_bool PASSED [ 70%] 2023-03-31T06:51:21.3002737Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float32 PASSED [ 70%] 2023-03-31T06:51:21.3002872Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:21.3003173Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:21.3003324Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.3003472Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_int64 PASSED [ 70%] 2023-03-31T06:51:21.3003616Z test_decomp.py::TestDecompCPU::test_quick_amax_cpu_uint8 PASSED [ 70%] 2023-03-31T06:51:21.3003764Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_bfloat16 PASSED [ 70%] 2023-03-31T06:51:21.3003913Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_float64 PASSED [ 70%] 2023-03-31T06:51:21.3004060Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int16 PASSED [ 70%] 2023-03-31T06:51:21.3004192Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int32 PASSED [ 70%] 2023-03-31T06:51:21.3004336Z test_decomp.py::TestDecompCPU::test_quick_amin_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3004485Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3004634Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:21.3004784Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:21.3004932Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:21.3005078Z test_decomp.py::TestDecompCPU::test_quick_aminmax_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3005225Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:21.3005420Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3005574Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_complex128 PASSED [ 71%] 2023-03-31T06:51:21.3005719Z test_decomp.py::TestDecompCPU::test_quick_any_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:21.3005863Z test_decomp.py::TestDecompCPU::test_quick_arange_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3006024Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3006232Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex128 PASSED [ 71%] 2023-03-31T06:51:21.3006438Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex32 PASSED [ 71%] 2023-03-31T06:51:21.3006607Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_complex64 PASSED [ 71%] 2023-03-31T06:51:21.3006801Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:21.3006968Z test_decomp.py::TestDecompCPU::test_quick_as_strided_scatter_cpu_int8 PASSED [ 71%] 2023-03-31T06:51:21.3007118Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3007270Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_complex128 PASSED [ 71%] 2023-03-31T06:51:21.3007416Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:21.3007562Z test_decomp.py::TestDecompCPU::test_quick_asin_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3007711Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:21.3007860Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_float64 PASSED [ 71%] 2023-03-31T06:51:21.3007992Z test_decomp.py::TestDecompCPU::test_quick_asinh_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3008142Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:21.3008287Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3008438Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:21.3008582Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int16 PASSED [ 71%] 2023-03-31T06:51:21.3008726Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:21.3008870Z test_decomp.py::TestDecompCPU::test_quick_atan2_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3009023Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_complex128 PASSED [ 71%] 2023-03-31T06:51:21.3009159Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:21.3009305Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:21.3009449Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_int64 PASSED [ 71%] 2023-03-31T06:51:21.3009595Z test_decomp.py::TestDecompCPU::test_quick_atan_cpu_uint8 PASSED [ 71%] 2023-03-31T06:51:21.3009745Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_bfloat16 PASSED [ 71%] 2023-03-31T06:51:21.3009892Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_bool PASSED [ 71%] 2023-03-31T06:51:21.3010040Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_float32 PASSED [ 71%] 2023-03-31T06:51:21.3010185Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int32 PASSED [ 71%] 2023-03-31T06:51:21.3010330Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3010465Z test_decomp.py::TestDecompCPU::test_quick_atanh_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:21.3010617Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_bool PASSED [ 72%] 2023-03-31T06:51:21.3010766Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3010917Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3011094Z test_decomp.py::TestDecompCPU::test_quick_bitwise_and_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:21.3011264Z test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:21.3011425Z test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:21.3011589Z test_decomp.py::TestDecompCPU::test_quick_bitwise_left_shift_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:21.3011726Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_bool PASSED [ 72%] 2023-03-31T06:51:21.3011902Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3012076Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3012226Z test_decomp.py::TestDecompCPU::test_quick_bitwise_not_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:21.3012427Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_bool PASSED [ 72%] 2023-03-31T06:51:21.3012575Z test_decomp.py::TestDecompCPU::test_quick_bitwise_or_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3012746Z test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3012906Z test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:21.3013050Z test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3013213Z test_decomp.py::TestDecompCPU::test_quick_bitwise_right_shift_cpu_int8 PASSED [ 72%] 2023-03-31T06:51:21.3013369Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_bool PASSED [ 72%] 2023-03-31T06:51:21.3013521Z test_decomp.py::TestDecompCPU::test_quick_bitwise_xor_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3013673Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_bfloat16 PASSED [ 72%] 2023-03-31T06:51:21.3013826Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_float16 PASSED [ 72%] 2023-03-31T06:51:21.3013978Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3014132Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:21.3014267Z test_decomp.py::TestDecompCPU::test_quick_bucketize_cpu_int64 PASSED [ 72%] 2023-03-31T06:51:21.3014411Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_bool PASSED [ 72%] 2023-03-31T06:51:21.3014561Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex128 PASSED [ 72%] 2023-03-31T06:51:21.3014707Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_complex64 PASSED [ 72%] 2023-03-31T06:51:21.3014855Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int16 PASSED [ 72%] 2023-03-31T06:51:21.3015000Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:21.3015144Z test_decomp.py::TestDecompCPU::test_quick_cat_cpu_uint8 PASSED [ 72%] 2023-03-31T06:51:21.3015293Z test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float16 XFAIL [ 72%] 2023-03-31T06:51:21.3015428Z test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float32 XFAIL [ 72%] 2023-03-31T06:51:21.3015574Z test_decomp.py::TestDecompCPU::test_quick_cauchy_cpu_float64 XFAIL [ 72%] 2023-03-31T06:51:21.3015719Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_float32 PASSED [ 72%] 2023-03-31T06:51:21.3015865Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_float64 PASSED [ 72%] 2023-03-31T06:51:21.3016012Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int32 PASSED [ 72%] 2023-03-31T06:51:21.3016158Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:21.3016303Z test_decomp.py::TestDecompCPU::test_quick_ceil_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:21.3016449Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:21.3016583Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_int16 PASSED [ 73%] 2023-03-31T06:51:21.3016730Z test_decomp.py::TestDecompCPU::test_quick_clamp_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:21.3016910Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3017062Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:21.3017209Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:21.3017356Z test_decomp.py::TestDecompCPU::test_quick_clamp_max_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:21.3017533Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:21.3017714Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3017848Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_int32 PASSED [ 73%] 2023-03-31T06:51:21.3017994Z test_decomp.py::TestDecompCPU::test_quick_clamp_min_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:21.3018171Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:21.3018316Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_bool PASSED [ 73%] 2023-03-31T06:51:21.3018466Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_complex128 PASSED [ 73%] 2023-03-31T06:51:21.3018614Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3018762Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:21.3018907Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:21.3019039Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:21.3019186Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:21.3019331Z test_decomp.py::TestDecompCPU::test_quick_clone_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:21.3026109Z test_decomp.py::TestDecompCPU::test_quick_complex_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3026338Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_bfloat16 PASSED [ 73%] 2023-03-31T06:51:21.3026517Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex128 PASSED [ 73%] 2023-03-31T06:51:21.3026686Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_complex64 PASSED [ 73%] 2023-03-31T06:51:21.3026850Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3026998Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:21.3027457Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 73%] 2023-03-31T06:51:21.3027805Z test_decomp.py::TestDecompCPU::test_quick_conj_physical_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 73%] 2023-03-31T06:51:21.3027978Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_complex64 PASSED [ 73%] 2023-03-31T06:51:21.3028146Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float16 PASSED [ 73%] 2023-03-31T06:51:21.3028307Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:21.3028463Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_float64 PASSED [ 73%] 2023-03-31T06:51:21.3028623Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int64 PASSED [ 73%] 2023-03-31T06:51:21.3028785Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:21.3028936Z test_decomp.py::TestDecompCPU::test_quick_constant_pad_nd_cpu_uint8 PASSED [ 73%] 2023-03-31T06:51:21.3029096Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_float32 PASSED [ 73%] 2023-03-31T06:51:21.3029250Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_int8 PASSED [ 73%] 2023-03-31T06:51:21.3029403Z test_decomp.py::TestDecompCPU::test_quick_copysign_cpu_uint8 PASSED [ 74%] 2023-03-31T06:51:21.3029554Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_bool PASSED [ 74%] 2023-03-31T06:51:21.3029855Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_complex64 PASSED [ 74%] 2023-03-31T06:51:21.3030007Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_float64 PASSED [ 74%] 2023-03-31T06:51:21.3030156Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:21.3030287Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:21.3030471Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int64 PASSED [ 74%] 2023-03-31T06:51:21.3030647Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_int8 PASSED [ 74%] 2023-03-31T06:51:21.3030792Z test_decomp.py::TestDecompCPU::test_quick_cos_cpu_uint8 PASSED [ 74%] 2023-03-31T06:51:21.3030941Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:21.3031122Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_bool PASSED [ 74%] 2023-03-31T06:51:21.3031277Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:21.3031427Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_complex64 PASSED [ 74%] 2023-03-31T06:51:21.3031560Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int16 PASSED [ 74%] 2023-03-31T06:51:21.3031705Z test_decomp.py::TestDecompCPU::test_quick_cosh_cpu_int64 PASSED [ 74%] 2023-03-31T06:51:21.3031855Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:21.3032009Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:21.3032163Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_complex64 PASSED [ 74%] 2023-03-31T06:51:21.3032312Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_float32 PASSED [ 74%] 2023-03-31T06:51:21.3032460Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int64 PASSED [ 74%] 2023-03-31T06:51:21.3032606Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_int8 PASSED [ 74%] 2023-03-31T06:51:21.3032740Z test_decomp.py::TestDecompCPU::test_quick_cumsum_cpu_uint8 PASSED [ 74%] 2023-03-31T06:51:21.3032935Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_bool SKIPPED (diag in torch.bool not supported) [ 74%] 2023-03-31T06:51:21.3033146Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_complex128 SKIPPED (diag in torch.complex128 not supported) [ 74%] 2023-03-31T06:51:21.3033356Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_complex64 SKIPPED (diag in torch.complex64 not supported) [ 74%] 2023-03-31T06:51:21.3033559Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_float32 SKIPPED (diag in torch.float32 not supported) [ 74%] 2023-03-31T06:51:21.3033753Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int16 SKIPPED (diag in torch.int16 not supported) [ 74%] 2023-03-31T06:51:21.3033948Z test_decomp.py::TestDecompCPU::test_quick_diag_cpu_int32 SKIPPED (diag in torch.int32 not supported) [ 74%] 2023-03-31T06:51:21.3034107Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_bfloat16 PASSED [ 74%] 2023-03-31T06:51:21.3034270Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex32 PASSED [ 74%] 2023-03-31T06:51:21.3034417Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_complex64 PASSED [ 74%] 2023-03-31T06:51:21.3034571Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_float16 PASSED [ 74%] 2023-03-31T06:51:21.3034721Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int32 PASSED [ 74%] 2023-03-31T06:51:21.3034875Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_int8 PASSED [ 74%] 2023-03-31T06:51:21.3035029Z test_decomp.py::TestDecompCPU::test_quick_diag_embed_cpu_uint8 PASSED [ 74%] 2023-03-31T06:51:21.3035182Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_bool PASSED [ 74%] 2023-03-31T06:51:21.3035352Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex128 PASSED [ 74%] 2023-03-31T06:51:21.3035543Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex32 PASSED [ 74%] 2023-03-31T06:51:21.3035691Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:21.3035853Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:21.3036007Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:21.3036158Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:21.3036341Z test_decomp.py::TestDecompCPU::test_quick_diagonal_copy_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:21.3036525Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:21.3036679Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex128 PASSED [ 75%] 2023-03-31T06:51:21.3036830Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_complex32 PASSED [ 75%] 2023-03-31T06:51:21.3036994Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:21.3037147Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:21.3037294Z test_decomp.py::TestDecompCPU::test_quick_diagonal_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:21.3037465Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:21.3037632Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float16 PASSED [ 75%] 2023-03-31T06:51:21.3037797Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:21.3037961Z test_decomp.py::TestDecompCPU::test_quick_diagonal_scatter_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:21.3038115Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:21.3038251Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_bool PASSED [ 75%] 2023-03-31T06:51:21.3038405Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:21.3038556Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:21.3038706Z test_decomp.py::TestDecompCPU::test_quick_digamma_cpu_int64 PASSED [ 75%] 2023-03-31T06:51:21.3038873Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float16 PASSED [ 75%] 2023-03-31T06:51:21.3039039Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:21.3039203Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:21.3039365Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int32 PASSED [ 75%] 2023-03-31T06:51:21.3039528Z test_decomp.py::TestDecompCPU::test_quick_div_floor_rounding_cpu_int8 PASSED [ 75%] 2023-03-31T06:51:21.3039681Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_bool PASSED [ 75%] 2023-03-31T06:51:21.3039856Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_complex128 PASSED [ 75%] 2023-03-31T06:51:21.3040024Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:21.3040191Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:21.3040353Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:21.3040512Z test_decomp.py::TestDecompCPU::test_quick_div_no_rounding_mode_cpu_uint8 PASSED [ 75%] 2023-03-31T06:51:21.3040682Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:21.3040850Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:21.3040999Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_float64 PASSED [ 75%] 2023-03-31T06:51:21.3041162Z test_decomp.py::TestDecompCPU::test_quick_div_trunc_rounding_cpu_int16 PASSED [ 75%] 2023-03-31T06:51:21.3041316Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_bfloat16 PASSED [ 75%] 2023-03-31T06:51:21.3041495Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_complex128 PASSED [ 75%] 2023-03-31T06:51:21.3041651Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:21.3041799Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_float32 PASSED [ 75%] 2023-03-31T06:51:21.3041945Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:21.3042090Z test_decomp.py::TestDecompCPU::test_quick_dot_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:21.3042322Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_bool SKIPPED (empty_like in torch.bool not supported) [ 76%] 2023-03-31T06:51:21.3042569Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_complex128 SKIPPED (empty_like in torch.complex128 not supported) [ 76%] 2023-03-31T06:51:21.3042816Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float16 SKIPPED (empty_like in torch.float16 not supported) [ 76%] 2023-03-31T06:51:21.3043373Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float32 SKIPPED (empty_like in torch.float32 not supported) [ 76%] 2023-03-31T06:51:21.3043589Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_float64 SKIPPED (empty_like in torch.float64 not supported) [ 76%] 2023-03-31T06:51:21.3043791Z test_decomp.py::TestDecompCPU::test_quick_empty_like_cpu_uint8 SKIPPED (empty_like in torch.uint8 not supported) [ 76%] 2023-03-31T06:51:21.3043945Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_bool PASSED [ 76%] 2023-03-31T06:51:21.3044102Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_complex128 PASSED [ 76%] 2023-03-31T06:51:21.3044255Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:21.3044391Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:21.3044543Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:21.3044692Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:21.3044838Z test_decomp.py::TestDecompCPU::test_quick_eq_cpu_int64 PASSED [ 76%] 2023-03-31T06:51:21.3044987Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:21.3045135Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_float64 PASSED [ 76%] 2023-03-31T06:51:21.3045281Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:21.3045428Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int64 PASSED [ 76%] 2023-03-31T06:51:21.3045561Z test_decomp.py::TestDecompCPU::test_quick_erf_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:21.3045712Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:21.3045861Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_bool PASSED [ 76%] 2023-03-31T06:51:21.3046007Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:21.3046153Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:21.3046299Z test_decomp.py::TestDecompCPU::test_quick_erfc_cpu_uint8 PASSED [ 76%] 2023-03-31T06:51:21.3046450Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_bfloat16 PASSED [ 76%] 2023-03-31T06:51:21.3046599Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_float32 PASSED [ 76%] 2023-03-31T06:51:21.3046734Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int64 PASSED [ 76%] 2023-03-31T06:51:21.3046882Z test_decomp.py::TestDecompCPU::test_quick_erfinv_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:21.3047028Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_bool PASSED [ 76%] 2023-03-31T06:51:21.3047176Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_complex64 PASSED [ 76%] 2023-03-31T06:51:21.3047324Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int16 PASSED [ 76%] 2023-03-31T06:51:21.3047551Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int32 PASSED [ 76%] 2023-03-31T06:51:21.3047699Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int64 PASSED [ 76%] 2023-03-31T06:51:21.3047845Z test_decomp.py::TestDecompCPU::test_quick_exp2_cpu_int8 PASSED [ 76%] 2023-03-31T06:51:21.3047990Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_bool PASSED [ 76%] 2023-03-31T06:51:21.3048127Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_complex128 PASSED [ 76%] 2023-03-31T06:51:21.3048318Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_complex64 PASSED [ 76%] 2023-03-31T06:51:21.3048495Z test_decomp.py::TestDecompCPU::test_quick_exp_cpu_int64 PASSED [ 77%] 2023-03-31T06:51:21.3048647Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_bfloat16 PASSED [ 77%] 2023-03-31T06:51:21.3048829Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_bool PASSED [ 77%] 2023-03-31T06:51:21.3048980Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float16 PASSED [ 77%] 2023-03-31T06:51:21.3049129Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:21.3049277Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:21.3049411Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int64 PASSED [ 77%] 2023-03-31T06:51:21.3049559Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_int8 PASSED [ 77%] 2023-03-31T06:51:21.3049709Z test_decomp.py::TestDecompCPU::test_quick_expand_cpu_uint8 PASSED [ 77%] 2023-03-31T06:51:21.3049862Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_bfloat16 PASSED [ 77%] 2023-03-31T06:51:21.3050008Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_bool PASSED [ 77%] 2023-03-31T06:51:21.3050160Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:21.3050311Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:21.3050460Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:21.3050591Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int16 PASSED [ 77%] 2023-03-31T06:51:21.3050736Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_int32 PASSED [ 77%] 2023-03-31T06:51:21.3050881Z test_decomp.py::TestDecompCPU::test_quick_expm1_cpu_uint8 PASSED [ 77%] 2023-03-31T06:51:21.3051027Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_float16 PASSED [ 77%] 2023-03-31T06:51:21.3051174Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int16 PASSED [ 77%] 2023-03-31T06:51:21.3051319Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_int64 PASSED [ 77%] 2023-03-31T06:51:21.3051464Z test_decomp.py::TestDecompCPU::test_quick_eye_cpu_uint8 PASSED [ 77%] 2023-03-31T06:51:21.3051617Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:21.3051753Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:21.3052136Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3052471Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3052805Z test_decomp.py::TestDecompCPU::test_quick_fft_fft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3053131Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3053287Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:21.3053445Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:21.3053594Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:21.3053782Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:21.3054096Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3054423Z test_decomp.py::TestDecompCPU::test_quick_fft_fft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3054607Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_complex64 PASSED [ 77%] 2023-03-31T06:51:21.3054791Z test_decomp.py::TestDecompCPU::test_quick_fft_fftn_cpu_float32 PASSED [ 77%] 2023-03-31T06:51:21.3055117Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3055301Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_float64 PASSED [ 77%] 2023-03-31T06:51:21.3055631Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 77%] 2023-03-31T06:51:21.3055788Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_complex128 PASSED [ 77%] 2023-03-31T06:51:21.3055940Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3056251Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3056575Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3056902Z test_decomp.py::TestDecompCPU::test_quick_fft_hfft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3057063Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_complex128 PASSED [ 78%] 2023-03-31T06:51:21.3057219Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3057380Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:21.3057711Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3058036Z test_decomp.py::TestDecompCPU::test_quick_fft_hfftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3058191Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3058501Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3058825Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3058981Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:21.3059136Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3059289Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:21.3059611Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3059926Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3060257Z test_decomp.py::TestDecompCPU::test_quick_fft_ifft_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3060420Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_complex128 PASSED [ 78%] 2023-03-31T06:51:21.3060733Z test_decomp.py::TestDecompCPU::test_quick_fft_ifftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3061099Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3061259Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3061585Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3061942Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3062296Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3062624Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3062995Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3063321Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfft_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3063582Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:21.3063904Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3064268Z test_decomp.py::TestDecompCPU::test_quick_fft_ihfftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3064432Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_complex128 PASSED [ 78%] 2023-03-31T06:51:21.3064593Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:21.3064921Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3065251Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 78%] 2023-03-31T06:51:21.3065409Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_complex64 PASSED [ 78%] 2023-03-31T06:51:21.3065562Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_float32 PASSED [ 78%] 2023-03-31T06:51:21.3065717Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_float64 PASSED [ 78%] 2023-03-31T06:51:21.3066043Z test_decomp.py::TestDecompCPU::test_quick_fft_irfft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3066193Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_complex64 PASSED [ 79%] 2023-03-31T06:51:21.3066346Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:21.3066503Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:21.3066828Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3067155Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3067479Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3067809Z test_decomp.py::TestDecompCPU::test_quick_fft_irfftn_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3068134Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3068465Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3068814Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3069137Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft2_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3069293Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:21.3069447Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:21.3069799Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3070160Z test_decomp.py::TestDecompCPU::test_quick_fft_rfft_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3070509Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3070666Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:21.3070986Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3071309Z test_decomp.py::TestDecompCPU::test_quick_fft_rfftn_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 79%] 2023-03-31T06:51:21.3071447Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_bool PASSED [ 79%] 2023-03-31T06:51:21.3071602Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex32 PASSED [ 79%] 2023-03-31T06:51:21.3071758Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_complex64 PASSED [ 79%] 2023-03-31T06:51:21.3071909Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float16 PASSED [ 79%] 2023-03-31T06:51:21.3072062Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:21.3072209Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_int8 PASSED [ 79%] 2023-03-31T06:51:21.3072356Z test_decomp.py::TestDecompCPU::test_quick_fill_cpu_uint8 PASSED [ 79%] 2023-03-31T06:51:21.3072505Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_bfloat16 PASSED [ 79%] 2023-03-31T06:51:21.3072639Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_bool PASSED [ 79%] 2023-03-31T06:51:21.3072789Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float32 PASSED [ 79%] 2023-03-31T06:51:21.3072940Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:21.3073087Z test_decomp.py::TestDecompCPU::test_quick_flip_cpu_int64 PASSED [ 79%] 2023-03-31T06:51:21.3073237Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_bfloat16 PASSED [ 79%] 2023-03-31T06:51:21.3073384Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int32 PASSED [ 79%] 2023-03-31T06:51:21.3073529Z test_decomp.py::TestDecompCPU::test_quick_floor_cpu_int64 PASSED [ 79%] 2023-03-31T06:51:21.3073692Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_bfloat16 PASSED [ 79%] 2023-03-31T06:51:21.3073838Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_float64 PASSED [ 79%] 2023-03-31T06:51:21.3073988Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int16 PASSED [ 79%] 2023-03-31T06:51:21.3074140Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int32 PASSED [ 79%] 2023-03-31T06:51:21.3074295Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:21.3074449Z test_decomp.py::TestDecompCPU::test_quick_floor_divide_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:21.3074598Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:21.3074745Z test_decomp.py::TestDecompCPU::test_quick_fmax_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:21.3074896Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_float32 PASSED [ 80%] 2023-03-31T06:51:21.3075063Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:21.3075210Z test_decomp.py::TestDecompCPU::test_quick_fmin_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:21.3075358Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:21.3075503Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:21.3075647Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:21.3075818Z test_decomp.py::TestDecompCPU::test_quick_fmod_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:21.3075993Z test_decomp.py::TestDecompCPU::test_quick_frac_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:21.3076138Z test_decomp.py::TestDecompCPU::test_quick_frac_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:21.3076337Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_bool PASSED [ 80%] 2023-03-31T06:51:21.3076492Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex128 PASSED [ 80%] 2023-03-31T06:51:21.3076640Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_complex32 PASSED [ 80%] 2023-03-31T06:51:21.3076785Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:21.3076930Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:21.3077072Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:21.3077216Z test_decomp.py::TestDecompCPU::test_quick_full_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:21.3077363Z test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int64 PASSED [ 80%] 2023-03-31T06:51:21.3077492Z test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_int8 PASSED [ 80%] 2023-03-31T06:51:21.3077636Z test_decomp.py::TestDecompCPU::test_quick_gcd_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:21.3077780Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:21.3077925Z test_decomp.py::TestDecompCPU::test_quick_ge_cpu_uint8 PASSED [ 80%] 2023-03-31T06:51:21.3078075Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_bfloat16 XFAIL [ 80%] 2023-03-31T06:51:21.3078226Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float16 XFAIL [ 80%] 2023-03-31T06:51:21.3078375Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_float32 XFAIL [ 80%] 2023-03-31T06:51:21.3078522Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int16 XFAIL [ 80%] 2023-03-31T06:51:21.3078656Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int32 XFAIL [ 80%] 2023-03-31T06:51:21.3078807Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_int8 XFAIL [ 80%] 2023-03-31T06:51:21.3078954Z test_decomp.py::TestDecompCPU::test_quick_geometric_cpu_uint8 XFAIL [ 80%] 2023-03-31T06:51:21.3079119Z test_decomp.py::TestDecompCPU::test_quick_grid_sampler_2d_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:21.3079270Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_bfloat16 PASSED [ 80%] 2023-03-31T06:51:21.3079414Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float16 PASSED [ 80%] 2023-03-31T06:51:21.3079558Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_float64 PASSED [ 80%] 2023-03-31T06:51:21.3079701Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int16 PASSED [ 80%] 2023-03-31T06:51:21.3079844Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int32 PASSED [ 80%] 2023-03-31T06:51:21.3079976Z test_decomp.py::TestDecompCPU::test_quick_gt_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:21.3080129Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3080278Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_bool PASSED [ 81%] 2023-03-31T06:51:21.3080430Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:21.3080579Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:21.3080769Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:21.3080921Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int64 PASSED [ 81%] 2023-03-31T06:51:21.3081069Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:21.3081206Z test_decomp.py::TestDecompCPU::test_quick_heaviside_cpu_uint8 PASSED [ 81%] 2023-03-31T06:51:21.3081381Z test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3081553Z test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:21.3081700Z test_decomp.py::TestDecompCPU::test_quick_igamma_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:21.3081850Z test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3082025Z test_decomp.py::TestDecompCPU::test_quick_igammac_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:21.3082176Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_bool PASSED [ 81%] 2023-03-31T06:51:21.3082337Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex128 PASSED [ 81%] 2023-03-31T06:51:21.3082475Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_complex32 PASSED [ 81%] 2023-03-31T06:51:21.3082625Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:21.3082773Z test_decomp.py::TestDecompCPU::test_quick_index_add_cpu_int32 PASSED [ 81%] 2023-03-31T06:51:21.3082930Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3083269Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_bool PASSED [ 81%] 2023-03-31T06:51:21.3083430Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex32 PASSED [ 81%] 2023-03-31T06:51:21.3083592Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_complex64 PASSED [ 81%] 2023-03-31T06:51:21.3083746Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_float16 PASSED [ 81%] 2023-03-31T06:51:21.3083881Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:21.3084030Z test_decomp.py::TestDecompCPU::test_quick_index_copy_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:21.3084183Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3084332Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_bool PASSED [ 81%] 2023-03-31T06:51:21.3084495Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_complex128 PASSED [ 81%] 2023-03-31T06:51:21.3084649Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float32 PASSED [ 81%] 2023-03-31T06:51:21.3084799Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_float64 PASSED [ 81%] 2023-03-31T06:51:21.3084947Z test_decomp.py::TestDecompCPU::test_quick_index_fill_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:21.3085094Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3085262Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_complex32 PASSED [ 81%] 2023-03-31T06:51:21.3085413Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int16 PASSED [ 81%] 2023-03-31T06:51:21.3085563Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int64 PASSED [ 81%] 2023-03-31T06:51:21.3085715Z test_decomp.py::TestDecompCPU::test_quick_index_select_cpu_int8 PASSED [ 81%] 2023-03-31T06:51:21.3085868Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_bfloat16 PASSED [ 81%] 2023-03-31T06:51:21.3086021Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex128 PASSED [ 81%] 2023-03-31T06:51:21.3086171Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_complex64 PASSED [ 82%] 2023-03-31T06:51:21.3086304Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:21.3086455Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:21.3086667Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int16 PASSED [ 82%] 2023-03-31T06:51:21.3086813Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:21.3086958Z test_decomp.py::TestDecompCPU::test_quick_isinf_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:21.3087106Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_bfloat16 PASSED [ 82%] 2023-03-31T06:51:21.3087252Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_bool PASSED [ 82%] 2023-03-31T06:51:21.3087447Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_complex128 PASSED [ 82%] 2023-03-31T06:51:21.3087623Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:21.3087768Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:21.3087949Z test_decomp.py::TestDecompCPU::test_quick_isnan_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:21.3088101Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_bool PASSED [ 82%] 2023-03-31T06:51:21.3088250Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_float16 PASSED [ 82%] 2023-03-31T06:51:21.3088397Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:21.3088546Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:21.3088694Z test_decomp.py::TestDecompCPU::test_quick_isneginf_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:21.3088834Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:21.3088987Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int16 PASSED [ 82%] 2023-03-31T06:51:21.3089134Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:21.3089281Z test_decomp.py::TestDecompCPU::test_quick_isposinf_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:21.3089430Z test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:21.3089576Z test_decomp.py::TestDecompCPU::test_quick_lcm_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:21.3089724Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_bfloat16 PASSED [ 82%] 2023-03-31T06:51:21.3089867Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_bool PASSED [ 82%] 2023-03-31T06:51:21.3089999Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:21.3090143Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_int64 PASSED [ 82%] 2023-03-31T06:51:21.3090287Z test_decomp.py::TestDecompCPU::test_quick_le_cpu_uint8 PASSED [ 82%] 2023-03-31T06:51:21.3090439Z test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_complex64 PASSED [ 82%] 2023-03-31T06:51:21.3090585Z test_decomp.py::TestDecompCPU::test_quick_lerp_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:21.3090733Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_bool PASSED [ 82%] 2023-03-31T06:51:21.3090883Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_float64 PASSED [ 82%] 2023-03-31T06:51:21.3091029Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int32 PASSED [ 82%] 2023-03-31T06:51:21.3091176Z test_decomp.py::TestDecompCPU::test_quick_lgamma_cpu_int8 PASSED [ 82%] 2023-03-31T06:51:21.3091328Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_bfloat16 PASSED [ 82%] 2023-03-31T06:51:21.3091698Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_bool SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 82%] 2023-03-31T06:51:21.3091863Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_float32 PASSED [ 82%] 2023-03-31T06:51:21.3092203Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 82%] 2023-03-31T06:51:21.3092546Z test_decomp.py::TestDecompCPU::test_quick_linalg_diagonal_cpu_uint8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 83%] 2023-03-31T06:51:21.3092751Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3092923Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3093091Z test_decomp.py::TestDecompCPU::test_quick_linalg_vector_norm_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:21.3093250Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3093416Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:21.3093597Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int32 PASSED [ 83%] 2023-03-31T06:51:21.3093749Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:21.3093898Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_int8 PASSED [ 83%] 2023-03-31T06:51:21.3094079Z test_decomp.py::TestDecompCPU::test_quick_linspace_cpu_uint8 PASSED [ 83%] 2023-03-31T06:51:21.3094231Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:21.3094383Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3094531Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3094664Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:21.3094811Z test_decomp.py::TestDecompCPU::test_quick_log10_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:21.3094958Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_bool PASSED [ 83%] 2023-03-31T06:51:21.3095110Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3095258Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:21.3095406Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:21.3095554Z test_decomp.py::TestDecompCPU::test_quick_log1p_cpu_uint8 PASSED [ 83%] 2023-03-31T06:51:21.3095699Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_bool PASSED [ 83%] 2023-03-31T06:51:21.3095834Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3095982Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3096126Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int16 PASSED [ 83%] 2023-03-31T06:51:21.3096273Z test_decomp.py::TestDecompCPU::test_quick_log2_cpu_int64 PASSED [ 83%] 2023-03-31T06:51:21.3096419Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_bool PASSED [ 83%] 2023-03-31T06:51:21.3096567Z test_decomp.py::TestDecompCPU::test_quick_log_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3096718Z test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_bfloat16 XFAIL [ 83%] 2023-03-31T06:51:21.3096869Z test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float32 XFAIL [ 83%] 2023-03-31T06:51:21.3097006Z test_decomp.py::TestDecompCPU::test_quick_log_normal_cpu_float64 XFAIL [ 83%] 2023-03-31T06:51:21.3097164Z test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:21.3097315Z test_decomp.py::TestDecompCPU::test_quick_log_softmax_cpu_float64 PASSED [ 83%] 2023-03-31T06:51:21.3097468Z test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:21.3097627Z test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3097780Z test_decomp.py::TestDecompCPU::test_quick_logaddexp_cpu_float32 PASSED [ 83%] 2023-03-31T06:51:21.3097938Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_bfloat16 PASSED [ 83%] 2023-03-31T06:51:21.3098103Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_complex128 PASSED [ 83%] 2023-03-31T06:51:21.3098251Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_complex64 PASSED [ 83%] 2023-03-31T06:51:21.3098433Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float16 PASSED [ 83%] 2023-03-31T06:51:21.3098587Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3098738Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:21.3098889Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:21.3099040Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:21.3099221Z test_decomp.py::TestDecompCPU::test_quick_logical_and_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:21.3099407Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:21.3099544Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_bool PASSED [ 84%] 2023-03-31T06:51:21.3099722Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3099876Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:21.3100027Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:21.3100175Z test_decomp.py::TestDecompCPU::test_quick_logical_not_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:21.3100325Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_bool PASSED [ 84%] 2023-03-31T06:51:21.3100487Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_complex128 PASSED [ 84%] 2023-03-31T06:51:21.3100637Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float16 PASSED [ 84%] 2023-03-31T06:51:21.3100776Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3100928Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:21.3101077Z test_decomp.py::TestDecompCPU::test_quick_logical_or_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:21.3101230Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3101381Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:21.3101530Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:21.3101680Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:21.3101829Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_int8 PASSED [ 84%] 2023-03-31T06:51:21.3101965Z test_decomp.py::TestDecompCPU::test_quick_logical_xor_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:21.3102116Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:21.3102264Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_bool PASSED [ 84%] 2023-03-31T06:51:21.3102407Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:21.3102555Z test_decomp.py::TestDecompCPU::test_quick_logit_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:21.3102707Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3102856Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:21.3103004Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_int32 PASSED [ 84%] 2023-03-31T06:51:21.3103139Z test_decomp.py::TestDecompCPU::test_quick_logspace_cpu_uint8 PASSED [ 84%] 2023-03-31T06:51:21.3103290Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3103547Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int16 PASSED [ 84%] 2023-03-31T06:51:21.3103703Z test_decomp.py::TestDecompCPU::test_quick_logsumexp_cpu_int64 PASSED [ 84%] 2023-03-31T06:51:21.3103849Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_bfloat16 PASSED [ 84%] 2023-03-31T06:51:21.3103995Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float16 PASSED [ 84%] 2023-03-31T06:51:21.3104141Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float32 PASSED [ 84%] 2023-03-31T06:51:21.3104322Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_float64 PASSED [ 84%] 2023-03-31T06:51:21.3104467Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int16 PASSED [ 85%] 2023-03-31T06:51:21.3104596Z test_decomp.py::TestDecompCPU::test_quick_lt_cpu_int32 PASSED [ 85%] 2023-03-31T06:51:21.3104756Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:21.3104945Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex128 PASSED [ 85%] 2023-03-31T06:51:21.3105131Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_complex32 PASSED [ 85%] 2023-03-31T06:51:21.3105284Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:21.3105435Z test_decomp.py::TestDecompCPU::test_quick_masked_fill_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:21.3105630Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:21.3105781Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_bool PASSED [ 85%] 2023-03-31T06:51:21.3105916Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float16 PASSED [ 85%] 2023-03-31T06:51:21.3106065Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:21.3106213Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int32 PASSED [ 85%] 2023-03-31T06:51:21.3106359Z test_decomp.py::TestDecompCPU::test_quick_maximum_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:21.3106509Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:21.3106659Z test_decomp.py::TestDecompCPU::test_quick_mean_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:21.3106888Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_bfloat16 SKIPPED (meshgrid in torch.bfloat16 not supported) [ 85%] 2023-03-31T06:51:21.3107111Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_bool SKIPPED (meshgrid in torch.bool not supported) [ 85%] 2023-03-31T06:51:21.3107328Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_complex64 SKIPPED (meshgrid in torch.complex64 not supported) [ 85%] 2023-03-31T06:51:21.3107552Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float32 SKIPPED (meshgrid in torch.float32 not supported) [ 85%] 2023-03-31T06:51:21.3107767Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_list_of_tensors_cpu_float64 SKIPPED (meshgrid in torch.float64 not supported) [ 85%] 2023-03-31T06:51:21.3108006Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_complex64 SKIPPED (meshgrid in torch.complex64 not supported) [ 85%] 2023-03-31T06:51:21.3108234Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_float16 SKIPPED (meshgrid in torch.float16 not supported) [ 85%] 2023-03-31T06:51:21.3108457Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int16 SKIPPED (meshgrid in torch.int16 not supported) [ 85%] 2023-03-31T06:51:21.3108676Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int32 SKIPPED (meshgrid in torch.int32 not supported) [ 85%] 2023-03-31T06:51:21.3108890Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int64 SKIPPED (meshgrid in torch.int64 not supported) [ 85%] 2023-03-31T06:51:21.3109107Z test_decomp.py::TestDecompCPU::test_quick_meshgrid_variadic_tensors_cpu_int8 SKIPPED (meshgrid in torch.int8 not supported) [ 85%] 2023-03-31T06:51:21.3109262Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_bool PASSED [ 85%] 2023-03-31T06:51:21.3109405Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float16 PASSED [ 85%] 2023-03-31T06:51:21.3109558Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_float64 PASSED [ 85%] 2023-03-31T06:51:21.3109711Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:21.3109861Z test_decomp.py::TestDecompCPU::test_quick_minimum_cpu_int8 PASSED [ 85%] 2023-03-31T06:51:21.3110045Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_bfloat16 PASSED [ 85%] 2023-03-31T06:51:21.3110195Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_bool PASSED [ 85%] 2023-03-31T06:51:21.3110346Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex128 PASSED [ 85%] 2023-03-31T06:51:21.3110493Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_complex32 PASSED [ 85%] 2023-03-31T06:51:21.3110652Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:21.3110823Z test_decomp.py::TestDecompCPU::test_quick_mul_cpu_int64 PASSED [ 85%] 2023-03-31T06:51:21.3110970Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_complex64 PASSED [ 85%] 2023-03-31T06:51:21.3111117Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_float32 PASSED [ 85%] 2023-03-31T06:51:21.3111288Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:21.3111433Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3111576Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_int8 PASSED [ 86%] 2023-03-31T06:51:21.3111720Z test_decomp.py::TestDecompCPU::test_quick_mv_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:21.3111890Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3112044Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_1_cpu_int16 PASSED [ 86%] 2023-03-31T06:51:21.3112215Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:21.3112384Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3112552Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_3_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3112722Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:21.3112889Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:21.3113054Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:21.3113218Z test_decomp.py::TestDecompCPU::test_quick_mvlgamma_mvlgamma_p_5_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3113356Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_bool PASSED [ 86%] 2023-03-31T06:51:21.3113511Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float16 PASSED [ 86%] 2023-03-31T06:51:21.3113665Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:21.3113817Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3113966Z test_decomp.py::TestDecompCPU::test_quick_nan_to_num_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3114119Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:21.3114272Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:21.3114421Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3114554Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:21.3114702Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3114849Z test_decomp.py::TestDecompCPU::test_quick_nansum_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:21.3115002Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float16 PASSED [ 86%] 2023-03-31T06:51:21.3115159Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_float32 PASSED [ 86%] 2023-03-31T06:51:21.3115310Z test_decomp.py::TestDecompCPU::test_quick_narrow_copy_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3115536Z test_decomp.py::TestDecompCPU::test_quick_native_batch_norm_cpu_float32 SKIPPED (native_batch_norm in torch.float32 not supported) [ 86%] 2023-03-31T06:51:21.3115736Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_bfloat16 PASSED [ 86%] 2023-03-31T06:51:21.3115910Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float16 PASSED [ 86%] 2023-03-31T06:51:21.3116066Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3116235Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int16 PASSED [ 86%] 2023-03-31T06:51:21.3116430Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int32 PASSED [ 86%] 2023-03-31T06:51:21.3116621Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int64 PASSED [ 86%] 2023-03-31T06:51:21.3116790Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_int8 PASSED [ 86%] 2023-03-31T06:51:21.3116986Z test_decomp.py::TestDecompCPU::test_quick_native_dropout_backward_cpu_uint8 PASSED [ 86%] 2023-03-31T06:51:21.3117152Z test_decomp.py::TestDecompCPU::test_quick_native_layer_norm_cpu_float64 PASSED [ 86%] 2023-03-31T06:51:21.3117305Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_complex64 PASSED [ 86%] 2023-03-31T06:51:21.3117438Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float16 PASSED [ 87%] 2023-03-31T06:51:21.3117586Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float32 PASSED [ 87%] 2023-03-31T06:51:21.3117732Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_float64 PASSED [ 87%] 2023-03-31T06:51:21.3117879Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:21.3118025Z test_decomp.py::TestDecompCPU::test_quick_ne_cpu_int32 PASSED [ 87%] 2023-03-31T06:51:21.3118172Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:21.3118322Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_complex32 PASSED [ 87%] 2023-03-31T06:51:21.3118471Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_float32 PASSED [ 87%] 2023-03-31T06:51:21.3118604Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:21.3118751Z test_decomp.py::TestDecompCPU::test_quick_neg_cpu_int64 PASSED [ 87%] 2023-03-31T06:51:21.3118965Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_bfloat16 SKIPPED (new_empty in torch.bfloat16 not supported) [ 87%] 2023-03-31T06:51:21.3119170Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_bool SKIPPED (new_empty in torch.bool not supported) [ 87%] 2023-03-31T06:51:21.3119389Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex128 SKIPPED (new_empty in torch.complex128 not supported) [ 87%] 2023-03-31T06:51:21.3119606Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex32 SKIPPED (new_empty in torch.complex32 not supported) [ 87%] 2023-03-31T06:51:21.3119823Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_complex64 SKIPPED (new_empty in torch.complex64 not supported) [ 87%] 2023-03-31T06:51:21.3120031Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float32 SKIPPED (new_empty in torch.float32 not supported) [ 87%] 2023-03-31T06:51:21.3120239Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_float64 SKIPPED (new_empty in torch.float64 not supported) [ 87%] 2023-03-31T06:51:21.3120425Z test_decomp.py::TestDecompCPU::test_quick_new_empty_cpu_int8 SKIPPED (new_empty in torch.int8 not supported) [ 87%] 2023-03-31T06:51:21.3120653Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_bfloat16 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3120881Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_complex32 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3121104Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_float64 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3121356Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int16 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3121572Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int32 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3121792Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_int8 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3122011Z test_decomp.py::TestDecompCPU::test_quick_new_empty_strided_cpu_uint8 SKIPPED (Expected: new_empty_strided is not comparable) [ 87%] 2023-03-31T06:51:21.3122192Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_bfloat16 PASSED [ 87%] 2023-03-31T06:51:21.3122415Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_bool PASSED [ 87%] 2023-03-31T06:51:21.3122559Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex128 PASSED [ 87%] 2023-03-31T06:51:21.3122740Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex32 PASSED [ 87%] 2023-03-31T06:51:21.3122898Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_complex64 PASSED [ 87%] 2023-03-31T06:51:21.3123227Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_float64 PASSED [ 87%] 2023-03-31T06:51:21.3123383Z test_decomp.py::TestDecompCPU::test_quick_new_full_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:21.3123532Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_bool PASSED [ 87%] 2023-03-31T06:51:21.3123685Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex128 PASSED [ 87%] 2023-03-31T06:51:21.3123841Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_complex64 PASSED [ 87%] 2023-03-31T06:51:21.3123977Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_float64 PASSED [ 87%] 2023-03-31T06:51:21.3124126Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int16 PASSED [ 87%] 2023-03-31T06:51:21.3124275Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int64 PASSED [ 87%] 2023-03-31T06:51:21.3124425Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_int8 PASSED [ 87%] 2023-03-31T06:51:21.3124575Z test_decomp.py::TestDecompCPU::test_quick_new_ones_cpu_uint8 PASSED [ 88%] 2023-03-31T06:51:21.3124723Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_bool PASSED [ 88%] 2023-03-31T06:51:21.3124876Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_complex64 PASSED [ 88%] 2023-03-31T06:51:21.3125027Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3125164Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3125314Z test_decomp.py::TestDecompCPU::test_quick_new_zeros_cpu_int64 PASSED [ 88%] 2023-03-31T06:51:21.3125470Z test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3125620Z test_decomp.py::TestDecompCPU::test_quick_nextafter_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3125828Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3126031Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3126225Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_binary_cross_entropy_with_logits_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3126395Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_elu_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3126567Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:21.3126728Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_embedding_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3126897Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3127063Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_gelu_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3127230Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_glu_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3127492Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardshrink_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3127670Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardsigmoid_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3127846Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardswish_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3128018Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3128219Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int16 PASSED [ 88%] 2023-03-31T06:51:21.3128420Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_hardtanh_cpu_int32 PASSED [ 88%] 2023-03-31T06:51:21.3128592Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_huber_loss_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3128771Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_bfloat16 PASSED [ 88%] 2023-03-31T06:51:21.3128996Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_leaky_relu_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3129168Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_logsigmoid_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3129334Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3129501Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mish_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3129670Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_mse_loss_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3129832Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_bool PASSED [ 88%] 2023-03-31T06:51:21.3130014Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_complex128 PASSED [ 88%] 2023-03-31T06:51:21.3130193Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float16 PASSED [ 88%] 2023-03-31T06:51:21.3130371Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float32 PASSED [ 88%] 2023-03-31T06:51:21.3130548Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_float64 PASSED [ 88%] 2023-03-31T06:51:21.3130719Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int16 PASSED [ 88%] 2023-03-31T06:51:21.3130891Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_int64 PASSED [ 88%] 2023-03-31T06:51:21.3131062Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_pad_constant_cpu_uint8 PASSED [ 88%] 2023-03-31T06:51:21.3131221Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:21.3131394Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_prelu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:21.3131633Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_bfloat16 SKIPPED (nn.functional.relu6 in torch.bfloat16 not supported) [ 89%] 2023-03-31T06:51:21.3131873Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_float32 SKIPPED (nn.functional.relu6 in torch.float32 not supported) [ 89%] 2023-03-31T06:51:21.3132102Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int16 SKIPPED (nn.functional.relu6 in torch.int16 not supported) [ 89%] 2023-03-31T06:51:21.3132327Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int32 SKIPPED (nn.functional.relu6 in torch.int32 not supported) [ 89%] 2023-03-31T06:51:21.3132553Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu6_cpu_int64 SKIPPED (nn.functional.relu6 in torch.int64 not supported) [ 89%] 2023-03-31T06:51:21.3132723Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:21.3132891Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int16 PASSED [ 89%] 2023-03-31T06:51:21.3133051Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int32 PASSED [ 89%] 2023-03-31T06:51:21.3133197Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int64 PASSED [ 89%] 2023-03-31T06:51:21.3133390Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_relu_cpu_int8 PASSED [ 89%] 2023-03-31T06:51:21.3133559Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:21.3133725Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_rrelu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:21.3133889Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:21.3134054Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_silu_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:21.3134254Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:21.3134456Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softplus_cpu_float32 PASSED [ 89%] 2023-03-31T06:51:21.3134618Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_softshrink_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:21.3134826Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_bfloat16 PASSED [ 89%] 2023-03-31T06:51:21.3135000Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_complex128 PASSED [ 89%] 2023-03-31T06:51:21.3135167Z test_decomp.py::TestDecompCPU::test_quick_nn_functional_unfold_cpu_float64 PASSED [ 89%] 2023-03-31T06:51:21.3135367Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 89%] 2023-03-31T06:51:21.3135571Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 89%] 2023-03-31T06:51:21.3135776Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 89%] 2023-03-31T06:51:21.3135972Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 89%] 2023-03-31T06:51:21.3136170Z test_decomp.py::TestDecompCPU::test_quick_norm_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 89%] 2023-03-31T06:51:21.3136366Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_bfloat16 SKIPPED (norm in torch.bfloat16 not supported) [ 89%] 2023-03-31T06:51:21.3136573Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_complex64 SKIPPED (norm in torch.complex64 not supported) [ 89%] 2023-03-31T06:51:21.3136771Z test_decomp.py::TestDecompCPU::test_quick_norm_fro_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 89%] 2023-03-31T06:51:21.3136965Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float16 SKIPPED (norm in torch.float16 not supported) [ 89%] 2023-03-31T06:51:21.3137158Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 89%] 2023-03-31T06:51:21.3137355Z test_decomp.py::TestDecompCPU::test_quick_norm_inf_cpu_float64 SKIPPED (norm in torch.float64 not supported) [ 89%] 2023-03-31T06:51:21.3137560Z test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_complex128 SKIPPED (norm in torch.complex128 not supported) [ 89%] 2023-03-31T06:51:21.3137755Z test_decomp.py::TestDecompCPU::test_quick_norm_nuc_cpu_float32 SKIPPED (norm in torch.float32 not supported) [ 89%] 2023-03-31T06:51:21.3137923Z test_decomp.py::TestDecompCPU::test_quick_normal_cpu_bfloat16 SKIPPED (Skipped!) [ 89%] 2023-03-31T06:51:21.3138089Z test_decomp.py::TestDecompCPU::test_quick_normal_cpu_float16 SKIPPED (Skipped!) [ 89%] 2023-03-31T06:51:21.3138239Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_bfloat16 XFAIL [ 89%] 2023-03-31T06:51:21.3138402Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_complex128 XFAIL [ 89%] 2023-03-31T06:51:21.3138565Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float16 XFAIL [ 90%] 2023-03-31T06:51:21.3138725Z test_decomp.py::TestDecompCPU::test_quick_normal_in_place_cpu_float32 XFAIL [ 90%] 2023-03-31T06:51:21.3138911Z test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_bfloat16 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:21.3139097Z test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float32 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:21.3139305Z test_decomp.py::TestDecompCPU::test_quick_normal_number_mean_cpu_float64 SKIPPED (Skipped!) [ 90%] 2023-03-31T06:51:21.3139459Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_bool PASSED [ 90%] 2023-03-31T06:51:21.3139599Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:21.3139749Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int16 PASSED [ 90%] 2023-03-31T06:51:21.3139922Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int32 PASSED [ 90%] 2023-03-31T06:51:21.3140096Z test_decomp.py::TestDecompCPU::test_quick_ones_cpu_int8 PASSED [ 90%] 2023-03-31T06:51:21.3140257Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_complex128 PASSED [ 90%] 2023-03-31T06:51:21.3140408Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_int32 PASSED [ 90%] 2023-03-31T06:51:21.3140589Z test_decomp.py::TestDecompCPU::test_quick_ones_like_cpu_uint8 PASSED [ 90%] 2023-03-31T06:51:21.3140744Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex128 PASSED [ 90%] 2023-03-31T06:51:21.3140884Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_complex32 PASSED [ 90%] 2023-03-31T06:51:21.3141034Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float16 PASSED [ 90%] 2023-03-31T06:51:21.3141184Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:21.3141336Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int16 PASSED [ 90%] 2023-03-31T06:51:21.3141486Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int32 PASSED [ 90%] 2023-03-31T06:51:21.3141636Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_int8 PASSED [ 90%] 2023-03-31T06:51:21.3141785Z test_decomp.py::TestDecompCPU::test_quick_permute_cpu_uint8 PASSED [ 90%] 2023-03-31T06:51:21.3141936Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_bfloat16 PASSED [ 90%] 2023-03-31T06:51:21.3142075Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_complex64 PASSED [ 90%] 2023-03-31T06:51:21.3142222Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float16 PASSED [ 90%] 2023-03-31T06:51:21.3142368Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_float32 PASSED [ 90%] 2023-03-31T06:51:21.3142513Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int16 PASSED [ 90%] 2023-03-31T06:51:21.3142656Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int32 PASSED [ 90%] 2023-03-31T06:51:21.3142804Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int64 PASSED [ 90%] 2023-03-31T06:51:21.3142950Z test_decomp.py::TestDecompCPU::test_quick_pow_cpu_int8 PASSED [ 90%] 2023-03-31T06:51:21.3143102Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_complex128 PASSED [ 90%] 2023-03-31T06:51:21.3143238Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_complex64 PASSED [ 90%] 2023-03-31T06:51:21.3143491Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_float64 PASSED [ 90%] 2023-03-31T06:51:21.3143647Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int16 PASSED [ 90%] 2023-03-31T06:51:21.3143792Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int32 PASSED [ 90%] 2023-03-31T06:51:21.3143936Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_int64 PASSED [ 90%] 2023-03-31T06:51:21.3144082Z test_decomp.py::TestDecompCPU::test_quick_prod_cpu_uint8 PASSED [ 90%] 2023-03-31T06:51:21.3144230Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_bfloat16 XFAIL [ 90%] 2023-03-31T06:51:21.3144381Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex128 XFAIL [ 90%] 2023-03-31T06:51:21.3144516Z test_decomp.py::TestDecompCPU::test_quick_randn_cpu_complex64 XFAIL [ 91%] 2023-03-31T06:51:21.3144670Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:21.3144835Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:21.3145041Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:21.3145195Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:21.3145349Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int32 PASSED [ 91%] 2023-03-31T06:51:21.3145499Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:21.3145649Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_int8 PASSED [ 91%] 2023-03-31T06:51:21.3145832Z test_decomp.py::TestDecompCPU::test_quick_reciprocal_cpu_uint8 PASSED [ 91%] 2023-03-31T06:51:21.3145997Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:21.3146148Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:21.3146333Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:21.3146482Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int16 PASSED [ 91%] 2023-03-31T06:51:21.3146634Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:21.3146784Z test_decomp.py::TestDecompCPU::test_quick_remainder_cpu_int8 PASSED [ 91%] 2023-03-31T06:51:21.3146933Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_bool PASSED [ 91%] 2023-03-31T06:51:21.3147085Z test_decomp.py::TestDecompCPU::test_quick_repeat_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:21.3147221Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:21.3147368Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_bool PASSED [ 91%] 2023-03-31T06:51:21.3147517Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex128 PASSED [ 91%] 2023-03-31T06:51:21.3147665Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:21.3147813Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:21.3147961Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:21.3148108Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:21.3148251Z test_decomp.py::TestDecompCPU::test_quick_roll_cpu_int8 PASSED [ 91%] 2023-03-31T06:51:21.3148384Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:21.3148528Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_bool PASSED [ 91%] 2023-03-31T06:51:21.3148679Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:21.3148826Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:21.3148972Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_float64 PASSED [ 91%] 2023-03-31T06:51:21.3149118Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int32 PASSED [ 91%] 2023-03-31T06:51:21.3149264Z test_decomp.py::TestDecompCPU::test_quick_rot90_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:21.3149412Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_bfloat16 PASSED [ 91%] 2023-03-31T06:51:21.3149543Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_bool PASSED [ 91%] 2023-03-31T06:51:21.3149693Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_float32 PASSED [ 91%] 2023-03-31T06:51:21.3149837Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int16 PASSED [ 91%] 2023-03-31T06:51:21.3149983Z test_decomp.py::TestDecompCPU::test_quick_rsqrt_cpu_int64 PASSED [ 91%] 2023-03-31T06:51:21.3150135Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_complex64 PASSED [ 91%] 2023-03-31T06:51:21.3150281Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float16 PASSED [ 91%] 2023-03-31T06:51:21.3150428Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:21.3150577Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3150741Z test_decomp.py::TestDecompCPU::test_quick_rsub_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:21.3151157Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_complex32 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:21.3151488Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_int16 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:21.3151850Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_int64 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:21.3152207Z test_decomp.py::TestDecompCPU::test_quick_select_cpu_int8 SKIPPED (only backwards is decomposed, but dtype doesn't support AD) [ 92%] 2023-03-31T06:51:21.3152357Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_bool PASSED [ 92%] 2023-03-31T06:51:21.3152541Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:21.3152694Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_complex64 PASSED [ 92%] 2023-03-31T06:51:21.3152843Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float16 PASSED [ 92%] 2023-03-31T06:51:21.3152978Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3153125Z test_decomp.py::TestDecompCPU::test_quick_sgn_cpu_uint8 PASSED [ 92%] 2023-03-31T06:51:21.3153277Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_bfloat16 PASSED [ 92%] 2023-03-31T06:51:21.3153432Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:21.3153584Z test_decomp.py::TestDecompCPU::test_quick_sigmoid_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3153732Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float16 PASSED [ 92%] 2023-03-31T06:51:21.3153882Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:21.3154028Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3154162Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:21.3154308Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:21.3154450Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:21.3154597Z test_decomp.py::TestDecompCPU::test_quick_sign_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:21.3154747Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_bool PASSED [ 92%] 2023-03-31T06:51:21.3154898Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float16 PASSED [ 92%] 2023-03-31T06:51:21.3155046Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:21.3155194Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3155330Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:21.3155478Z test_decomp.py::TestDecompCPU::test_quick_signbit_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:21.3155623Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_bool PASSED [ 92%] 2023-03-31T06:51:21.3155772Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_complex128 PASSED [ 92%] 2023-03-31T06:51:21.3155918Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_complex64 PASSED [ 92%] 2023-03-31T06:51:21.3156063Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_float32 PASSED [ 92%] 2023-03-31T06:51:21.3156210Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_float64 PASSED [ 92%] 2023-03-31T06:51:21.3156357Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int16 PASSED [ 92%] 2023-03-31T06:51:21.3156488Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int32 PASSED [ 92%] 2023-03-31T06:51:21.3156634Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int64 PASSED [ 92%] 2023-03-31T06:51:21.3156816Z test_decomp.py::TestDecompCPU::test_quick_sin_cpu_int8 PASSED [ 92%] 2023-03-31T06:51:21.3156966Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:21.3157109Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_bool PASSED [ 93%] 2023-03-31T06:51:21.3157260Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_complex128 PASSED [ 93%] 2023-03-31T06:51:21.3157406Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:21.3157577Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:21.3157735Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:21.3157881Z test_decomp.py::TestDecompCPU::test_quick_sinc_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:21.3158028Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:21.3158207Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:21.3158353Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3158499Z test_decomp.py::TestDecompCPU::test_quick_sinh_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:21.3158653Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_complex128 PASSED [ 93%] 2023-03-31T06:51:21.3158799Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:21.3158947Z test_decomp.py::TestDecompCPU::test_quick_slice_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3159087Z test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:21.3159236Z test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:21.3159386Z test_decomp.py::TestDecompCPU::test_quick_softmax_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:21.3159548Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:21.3159700Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_bool PASSED [ 93%] 2023-03-31T06:51:21.3159852Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:21.3160003Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3160155Z test_decomp.py::TestDecompCPU::test_quick_special_entr_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:21.3160293Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3160446Z test_decomp.py::TestDecompCPU::test_quick_special_erfcx_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:21.3160608Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_bfloat16 PASSED [ 93%] 2023-03-31T06:51:21.3160759Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_bool PASSED [ 93%] 2023-03-31T06:51:21.3160913Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:21.3161066Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:21.3161217Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:21.3161365Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3161502Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int64 PASSED [ 93%] 2023-03-31T06:51:21.3161651Z test_decomp.py::TestDecompCPU::test_quick_special_i0e_cpu_int8 PASSED [ 93%] 2023-03-31T06:51:21.3161802Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:21.3161954Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_float64 PASSED [ 93%] 2023-03-31T06:51:21.3162104Z test_decomp.py::TestDecompCPU::test_quick_special_i1_cpu_uint8 PASSED [ 93%] 2023-03-31T06:51:21.3162255Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_bool PASSED [ 93%] 2023-03-31T06:51:21.3162409Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_float32 PASSED [ 93%] 2023-03-31T06:51:21.3162587Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int16 PASSED [ 93%] 2023-03-31T06:51:21.3162725Z test_decomp.py::TestDecompCPU::test_quick_special_i1e_cpu_int32 PASSED [ 93%] 2023-03-31T06:51:21.3162885Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_bool PASSED [ 94%] 2023-03-31T06:51:21.3163272Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:21.3163438Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:21.3163667Z test_decomp.py::TestDecompCPU::test_quick_special_log_ndtr_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:21.3163932Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_bfloat16 SKIPPED (special.ndtr in torch.bfloat16 not supported) [ 94%] 2023-03-31T06:51:21.3164152Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_float32 SKIPPED (special.ndtr in torch.float32 not supported) [ 94%] 2023-03-31T06:51:21.3164402Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int16 SKIPPED (special.ndtr in torch.int16 not supported) [ 94%] 2023-03-31T06:51:21.3164612Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int32 SKIPPED (special.ndtr in torch.int32 not supported) [ 94%] 2023-03-31T06:51:21.3164809Z test_decomp.py::TestDecompCPU::test_quick_special_ndtr_cpu_int8 SKIPPED (special.ndtr in torch.int8 not supported) [ 94%] 2023-03-31T06:51:21.3164970Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:21.3165132Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:21.3165292Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:21.3165448Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:21.3165605Z test_decomp.py::TestDecompCPU::test_quick_special_ndtri_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:21.3165766Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:21.3165926Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:21.3166072Z test_decomp.py::TestDecompCPU::test_quick_special_xlog1py_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:21.3166226Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:21.3166380Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:21.3166535Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:21.3166690Z test_decomp.py::TestDecompCPU::test_quick_special_zeta_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:21.3166844Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_complex128 PASSED [ 94%] 2023-03-31T06:51:21.3166994Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_float64 PASSED [ 94%] 2023-03-31T06:51:21.3167144Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:21.3167279Z test_decomp.py::TestDecompCPU::test_quick_split_cpu_uint8 PASSED [ 94%] 2023-03-31T06:51:21.3167447Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_bfloat16 PASSED [ 94%] 2023-03-31T06:51:21.3167608Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_bool PASSED [ 94%] 2023-03-31T06:51:21.3167776Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_complex128 PASSED [ 94%] 2023-03-31T06:51:21.3167939Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float16 PASSED [ 94%] 2023-03-31T06:51:21.3168101Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_float32 PASSED [ 94%] 2023-03-31T06:51:21.3168257Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int32 PASSED [ 94%] 2023-03-31T06:51:21.3168415Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int64 PASSED [ 94%] 2023-03-31T06:51:21.3168562Z test_decomp.py::TestDecompCPU::test_quick_split_list_args_cpu_int8 PASSED [ 94%] 2023-03-31T06:51:21.3168766Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_bfloat16 PASSED [ 94%] 2023-03-31T06:51:21.3168926Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_bool PASSED [ 94%] 2023-03-31T06:51:21.3169093Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex32 PASSED [ 94%] 2023-03-31T06:51:21.3169253Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_complex64 PASSED [ 94%] 2023-03-31T06:51:21.3169504Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int16 PASSED [ 94%] 2023-03-31T06:51:21.3169691Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:21.3169850Z test_decomp.py::TestDecompCPU::test_quick_split_with_sizes_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:21.3170004Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:21.3170174Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3170327Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:21.3170476Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:21.3170624Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int16 PASSED [ 95%] 2023-03-31T06:51:21.3170774Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:21.3170920Z test_decomp.py::TestDecompCPU::test_quick_sqrt_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:21.3171074Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:21.3171234Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:21.3171373Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex32 PASSED [ 95%] 2023-03-31T06:51:21.3171526Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3171680Z test_decomp.py::TestDecompCPU::test_quick_squeeze_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:21.3171848Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex32 PASSED [ 95%] 2023-03-31T06:51:21.3172015Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3172182Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:21.3172343Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:21.3172504Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:21.3172656Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int16 PASSED [ 95%] 2023-03-31T06:51:21.3172814Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int32 PASSED [ 95%] 2023-03-31T06:51:21.3172972Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_int64 PASSED [ 95%] 2023-03-31T06:51:21.3173135Z test_decomp.py::TestDecompCPU::test_quick_squeeze_multiple_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:21.3173287Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:21.3173438Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_bool PASSED [ 95%] 2023-03-31T06:51:21.3173590Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3173742Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:21.3173877Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int16 PASSED [ 95%] 2023-03-31T06:51:21.3174025Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_int8 PASSED [ 95%] 2023-03-31T06:51:21.3174173Z test_decomp.py::TestDecompCPU::test_quick_stack_cpu_uint8 PASSED [ 95%] 2023-03-31T06:51:21.3174323Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:21.3174477Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:21.3174654Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_float16 PASSED [ 95%] 2023-03-31T06:51:21.3174804Z test_decomp.py::TestDecompCPU::test_quick_std_cpu_float64 PASSED [ 95%] 2023-03-31T06:51:21.3174954Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_bfloat16 PASSED [ 95%] 2023-03-31T06:51:21.3175092Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_complex128 PASSED [ 95%] 2023-03-31T06:51:21.3175245Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3175420Z test_decomp.py::TestDecompCPU::test_quick_std_mean_cpu_float32 PASSED [ 95%] 2023-03-31T06:51:21.3175609Z test_decomp.py::TestDecompCPU::test_quick_std_unbiased_cpu_complex64 PASSED [ 95%] 2023-03-31T06:51:21.3175757Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:21.3175933Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex32 PASSED [ 96%] 2023-03-31T06:51:21.3176084Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:21.3176236Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3176367Z test_decomp.py::TestDecompCPU::test_quick_sub_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:21.3176514Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:21.3176663Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:21.3176812Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:21.3176959Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float16 PASSED [ 96%] 2023-03-31T06:51:21.3177104Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3177252Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:21.3177396Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:21.3177529Z test_decomp.py::TestDecompCPU::test_quick_sum_cpu_uint8 PASSED [ 96%] 2023-03-31T06:51:21.3177676Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:21.3177821Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_float16 PASSED [ 96%] 2023-03-31T06:51:21.3177963Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3178110Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:21.3178255Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:21.3178398Z test_decomp.py::TestDecompCPU::test_quick_t_cpu_uint8 PASSED [ 96%] 2023-03-31T06:51:21.3178543Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_bfloat16 PASSED [ 96%] 2023-03-31T06:51:21.3178690Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3178823Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:21.3178967Z test_decomp.py::TestDecompCPU::test_quick_tan_cpu_int8 PASSED [ 96%] 2023-03-31T06:51:21.3179113Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_bool PASSED [ 96%] 2023-03-31T06:51:21.3179262Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:21.3179412Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:21.3179561Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3179709Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:21.3179854Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int32 PASSED [ 96%] 2023-03-31T06:51:21.3179989Z test_decomp.py::TestDecompCPU::test_quick_tanh_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:21.3180166Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:21.3180320Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_complex64 PASSED [ 96%] 2023-03-31T06:51:21.3180467Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_float32 PASSED [ 96%] 2023-03-31T06:51:21.3180615Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_float64 PASSED [ 96%] 2023-03-31T06:51:21.3180765Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int16 PASSED [ 96%] 2023-03-31T06:51:21.3180935Z test_decomp.py::TestDecompCPU::test_quick_trace_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:21.3181123Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_complex128 PASSED [ 96%] 2023-03-31T06:51:21.3181260Z test_decomp.py::TestDecompCPU::test_quick_transpose_cpu_int64 PASSED [ 96%] 2023-03-31T06:51:21.3181407Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:21.3181584Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_complex64 PASSED [ 97%] 2023-03-31T06:51:21.3181733Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:21.3181881Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:21.3182025Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:21.3182170Z test_decomp.py::TestDecompCPU::test_quick_tril_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:21.3182320Z test_decomp.py::TestDecompCPU::test_quick_tril_indices_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:21.3182456Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:21.3182606Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float16 PASSED [ 97%] 2023-03-31T06:51:21.3182751Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:21.3182898Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:21.3183045Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:21.3183189Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:21.3183429Z test_decomp.py::TestDecompCPU::test_quick_triu_cpu_uint8 PASSED [ 97%] 2023-03-31T06:51:21.3183588Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:21.3183723Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:21.3183874Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:21.3184020Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:21.3184165Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:21.3184314Z test_decomp.py::TestDecompCPU::test_quick_trunc_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:21.3184466Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex128 PASSED [ 97%] 2023-03-31T06:51:21.3184619Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_complex64 PASSED [ 97%] 2023-03-31T06:51:21.3184770Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float16 PASSED [ 97%] 2023-03-31T06:51:21.3184903Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:21.3185052Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:21.3185202Z test_decomp.py::TestDecompCPU::test_quick_unbind_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:21.3185366Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_complex32 PASSED [ 97%] 2023-03-31T06:51:21.3185519Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float16 PASSED [ 97%] 2023-03-31T06:51:21.3185672Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:21.3185824Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_float64 PASSED [ 97%] 2023-03-31T06:51:21.3186010Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:21.3186147Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int64 PASSED [ 97%] 2023-03-31T06:51:21.3186299Z test_decomp.py::TestDecompCPU::test_quick_unfold_copy_cpu_int8 PASSED [ 97%] 2023-03-31T06:51:21.3186451Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_bfloat16 PASSED [ 97%] 2023-03-31T06:51:21.3186601Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_complex128 PASSED [ 97%] 2023-03-31T06:51:21.3186775Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_float32 PASSED [ 97%] 2023-03-31T06:51:21.3186949Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int16 PASSED [ 97%] 2023-03-31T06:51:21.3187095Z test_decomp.py::TestDecompCPU::test_quick_unfold_cpu_int32 PASSED [ 97%] 2023-03-31T06:51:21.3187286Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_complex128 XFAIL [ 97%] 2023-03-31T06:51:21.3187426Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_complex64 XFAIL [ 98%] 2023-03-31T06:51:21.3187576Z test_decomp.py::TestDecompCPU::test_quick_uniform_cpu_float64 XFAIL [ 98%] 2023-03-31T06:51:21.3187735Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:21.3187898Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:21.3188060Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_complex32 PASSED [ 98%] 2023-03-31T06:51:21.3188215Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int16 PASSED [ 98%] 2023-03-31T06:51:21.3188368Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int64 PASSED [ 98%] 2023-03-31T06:51:21.3188518Z test_decomp.py::TestDecompCPU::test_quick_unsafe_split_cpu_int8 PASSED [ 98%] 2023-03-31T06:51:21.3188658Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:21.3188812Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_complex32 PASSED [ 98%] 2023-03-31T06:51:21.3188964Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:21.3189115Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int16 PASSED [ 98%] 2023-03-31T06:51:21.3189264Z test_decomp.py::TestDecompCPU::test_quick_unsqueeze_cpu_int32 PASSED [ 98%] 2023-03-31T06:51:21.3189410Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:21.3189559Z test_decomp.py::TestDecompCPU::test_quick_var_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:21.3189714Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:21.3189863Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:21.3189998Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:21.3190150Z test_decomp.py::TestDecompCPU::test_quick_var_mean_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:21.3190320Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:21.3190483Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:21.3190644Z test_decomp.py::TestDecompCPU::test_quick_var_mean_unbiased_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:21.3190804Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:21.3190961Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:21.3191122Z test_decomp.py::TestDecompCPU::test_quick_var_unbiased_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:21.3191261Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_bfloat16 PASSED [ 98%] 2023-03-31T06:51:21.3191411Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex128 PASSED [ 98%] 2023-03-31T06:51:21.3191564Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:21.3191744Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_float16 PASSED [ 98%] 2023-03-31T06:51:21.3191894Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:21.3192041Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_float64 PASSED [ 98%] 2023-03-31T06:51:21.3192187Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_int16 PASSED [ 98%] 2023-03-31T06:51:21.3192333Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_int32 PASSED [ 98%] 2023-03-31T06:51:21.3192494Z test_decomp.py::TestDecompCPU::test_quick_view_cpu_int8 PASSED [ 98%] 2023-03-31T06:51:21.3192666Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_bool PASSED [ 98%] 2023-03-31T06:51:21.3192818Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex32 PASSED [ 98%] 2023-03-31T06:51:21.3192998Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_complex64 PASSED [ 98%] 2023-03-31T06:51:21.3193146Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_float32 PASSED [ 98%] 2023-03-31T06:51:21.3193295Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3193441Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_int8 PASSED [ 99%] 2023-03-31T06:51:21.3193587Z test_decomp.py::TestDecompCPU::test_quick_where_cpu_uint8 PASSED [ 99%] 2023-03-31T06:51:21.3193721Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_bool PASSED [ 99%] 2023-03-31T06:51:21.3193870Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float16 PASSED [ 99%] 2023-03-31T06:51:21.3194020Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3194167Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3194311Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int32 PASSED [ 99%] 2023-03-31T06:51:21.3194459Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_int8 PASSED [ 99%] 2023-03-31T06:51:21.3194608Z test_decomp.py::TestDecompCPU::test_quick_xlogy_cpu_uint8 PASSED [ 99%] 2023-03-31T06:51:21.3194812Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_bfloat16 SKIPPED (zero_ in torch.bfloat16 not supported) [ 99%] 2023-03-31T06:51:21.3194990Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_bool SKIPPED (zero_ in torch.bool not supported) [ 99%] 2023-03-31T06:51:21.3195188Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float16 SKIPPED (zero_ in torch.float16 not supported) [ 99%] 2023-03-31T06:51:21.3195386Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_float64 SKIPPED (zero_ in torch.float64 not supported) [ 99%] 2023-03-31T06:51:21.3195579Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int16 SKIPPED (zero_ in torch.int16 not supported) [ 99%] 2023-03-31T06:51:21.3195766Z test_decomp.py::TestDecompCPU::test_quick_zero__cpu_int64 SKIPPED (zero_ in torch.int64 not supported) [ 99%] 2023-03-31T06:51:21.3195922Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex128 PASSED [ 99%] 2023-03-31T06:51:21.3196077Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_complex64 PASSED [ 99%] 2023-03-31T06:51:21.3196227Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3196376Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int16 PASSED [ 99%] 2023-03-31T06:51:21.3196509Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int32 PASSED [ 99%] 2023-03-31T06:51:21.3196659Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_int64 PASSED [ 99%] 2023-03-31T06:51:21.3196809Z test_decomp.py::TestDecompCPU::test_quick_zeros_cpu_uint8 PASSED [ 99%] 2023-03-31T06:51:21.3196960Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3197110Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int16 PASSED [ 99%] 2023-03-31T06:51:21.3197261Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int32 PASSED [ 99%] 2023-03-31T06:51:21.3197440Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_int8 PASSED [ 99%] 2023-03-31T06:51:21.3197591Z test_decomp.py::TestDecompCPU::test_quick_zeros_like_cpu_uint8 PASSED [ 99%] 2023-03-31T06:51:21.3197758Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_GRU_train_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3197939Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_eval_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3198150Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_eval_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3198358Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_train_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3198533Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_LSTM_train_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3198739Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_eval_mode_cpu_float32 PASSED [ 99%] 2023-03-31T06:51:21.3198921Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_eval_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3199101Z test_decomp.py::TestDecompCPU::test_rnn_decomp_module_nn_RNN_train_mode_cpu_float64 PASSED [ 99%] 2023-03-31T06:51:21.3199248Z test_decomp.py::TestDecompCPU::test_uniform_cpu PASSED [ 99%] 2023-03-31T06:51:21.3199415Z test_decomp.py::DecompContiguousTestsCPU::test_contiguous_log_softmax_cpu PASSED [ 99%] 2023-03-31T06:51:21.3199608Z test_decomp.py::DecompAmpTestsCPU::test_amp_batch_norm_backward_cpu SKIPPED (Only runs on cuda) [100%] 2023-03-31T06:51:21.3199624Z 2023-03-31T06:51:21.3200017Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-9412a89c8914e16a.xml - 2023-03-31T06:51:21.3200161Z === 3546 passed, 280 skipped, 4 deselected, 23 xfailed in 1587.75s (0:26:27) === 2023-03-31T06:51:21.3200346Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:51:21.3200353Z 2023-03-31T06:51:21.3200735Z ##[endgroup] 2023-03-31T06:51:21.3201038Z FINISHED PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_99_jfdlp.log) 2023-03-31T06:51:21.3201045Z 2023-03-31T06:51:21.5750616Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_decomp.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '-k=_linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:51:21.574615] 2023-03-31T06:51:31.5478514Z 2023-03-31T06:51:31.5478985Z Expand the folded group to see the log file of test_decomp 2023-03-31T06:51:31.5480055Z ##[group]PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_795ihe11.log) 2023-03-31T06:51:31.5480837Z Test results will be stored in test-reports/python-pytest/test_decomp/test_decomp-7f63c915f8ee55fa.xml 2023-03-31T06:51:31.5481402Z ============================= test session starts ============================== 2023-03-31T06:51:31.5482137Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T06:51:31.5482408Z cachedir: .pytest_cache 2023-03-31T06:51:31.5482854Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T06:51:31.5483471Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T06:51:31.5483928Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T06:51:31.5484285Z collecting ... collected 7785 items / 7777 deselected / 8 selected 2023-03-31T06:51:31.5485523Z Running 8 items in this shard: test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_float64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_complex128, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_complex64, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_float32, test/test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_float64 2023-03-31T06:51:31.5486388Z 2023-03-31T06:51:31.5486629Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_complex128 PASSED [ 12%] 2023-03-31T06:51:31.5487059Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_complex64 PASSED [ 25%] 2023-03-31T06:51:31.5487456Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_float32 PASSED [ 37%] 2023-03-31T06:51:31.5487875Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_cpu_float64 PASSED [ 50%] 2023-03-31T06:51:31.5488245Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_complex128 PASSED [ 62%] 2023-03-31T06:51:31.5488636Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_complex64 PASSED [ 75%] 2023-03-31T06:51:31.5489069Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_float32 PASSED [ 87%] 2023-03-31T06:51:31.5489572Z test_decomp.py::TestDecompCPU::test_comprehensive_linalg_cholesky_ex_cpu_float64 PASSED [100%] 2023-03-31T06:51:31.5489779Z 2023-03-31T06:51:31.5490158Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-7f63c915f8ee55fa.xml - 2023-03-31T06:51:31.5490634Z ====================== 8 passed, 7777 deselected in 6.41s ====================== 2023-03-31T06:51:31.5490962Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T06:51:31.5491170Z 2023-03-31T06:51:31.5491423Z ##[endgroup] 2023-03-31T06:51:31.5491828Z FINISHED PRINTING LOG FILE of test_decomp (/var/lib/jenkins/workspace/test/test-reports/test_decomp_795ihe11.log) 2023-03-31T06:51:31.5492054Z 2023-03-31T06:51:31.5492229Z Running test_ops_jit ... [2023-03-31 06:51:31.548815] 2023-03-31T06:51:33.2773112Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:51:33.2801600Z Ignoring disabled issues: [] 2023-03-31T06:51:33.2985639Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T06:51:33.2988165Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_jit.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=0', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:51:33.298541] 2023-03-31T06:51:33.3015394Z Ignoring disabled issues: [] 2023-03-31T06:51:33.3204765Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_jit.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=1', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 06:51:33.320023] 2023-03-31T07:00:29.5138343Z 2023-03-31T07:00:29.5138911Z Expand the folded group to see the log file of test_ops_jit 2023-03-31T07:00:29.5139652Z ##[group]PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_npw9qolt.log) 2023-03-31T07:00:29.5140359Z Test results will be stored in test-reports/python-pytest/test_ops_jit/test_ops_jit-ad78045603f8ca6d.xml 2023-03-31T07:00:29.5140706Z ============================= test session starts ============================== 2023-03-31T07:00:29.5144204Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:00:29.5144663Z cachedir: .pytest_cache 2023-03-31T07:00:29.5145726Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:00:29.5146544Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:00:29.5147518Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:00:29.5148273Z collecting ... collected 1079 items / 3 deselected / 1076 selected 2023-03-31T07:00:29.5285094Z Running 504 items in this shard: test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_abs_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_acosh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_asinh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atan_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atanh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_cat_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_digamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_floor_rounding_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_no_rounding_mode_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erfc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_gt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_le_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_householder_product_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_inv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_matrix_power_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log1p_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log_softmax_with_dtype_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_logit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_logsumexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mH_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_max_binary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mul_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_5_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_ne_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_neg_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_transpose_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_trunc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_vstack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_T_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___getitem___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___getitem___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rdiv___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmatmul___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmul___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmul___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__native_batch_norm_legit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__segment_reduce_lengths_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_abs_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acos_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acos_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acosh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_add_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addbmm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcdiv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcmul_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_all_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_amax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_aminmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_any_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argsort_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argwhere_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_scatter_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asin_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asinh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atanh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atanh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_2d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_3d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_baddbmm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bernoulli_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bfloat16_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bfloat16_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_block_diag_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_shapes_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_byte_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_byte_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cartesian_prod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cat_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cauchy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdist_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdouble_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chalf_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_char_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_char_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_min_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clone_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clone_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_column_stack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_combinations_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_physical_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_constant_pad_nd_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cos_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cosh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_count_nonzero_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cross_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cross_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cummin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumprod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumsum_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumsum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumulative_trapezoid_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumulative_trapezoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_deg2rad_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagflat_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_copy_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_scatter_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diff_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dist_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_floor_rounding_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_no_rounding_mode_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_no_rounding_mode_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dot_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_double_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dsplit_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dstack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_einsum_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_permuted_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_permuted_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eq_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_equal_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_equal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expm1_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exponential_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftshift_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfftn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftshift_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftshift_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfftn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fill_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fill_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flatten_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fliplr_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fliplr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flipud_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_power_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_power_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_frac_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_frexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geqrf_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geqrf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gradient_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gradient_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_heaviside_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histogram_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hsplit_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hsplit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_igammac_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_copy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_fill_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_fill_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_put_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_put_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_select_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_inner_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_int_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isclose_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isfinite_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isfinite_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isinf_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isinf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isnan_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_2inputs_2outputs_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_4inputs_with_extra_args_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kron_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kthvalue_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lerp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lerp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lgamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cond_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cond_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_singular_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_diagonal_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_diagonal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eig_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvalsh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_householder_product_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_householder_product_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_ex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_grad_oriented_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_ex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_norm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_hermitian_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_hermitian_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_multi_dot_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_subgradients_at_zero_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_hermitian_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_hermitian_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_singular_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_slogdet_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_slogdet_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_ex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_triangular_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_triangular_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svdvals_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorinv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorsolve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vecdot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vector_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linspace_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log10_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log10_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log1p_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_normal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_with_dtype_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logdet_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_not_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logspace_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logsumexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_long_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_unpack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mT_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_amin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_argmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_argmin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumprod_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumsum_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumsum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_fill_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_log_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_logsumexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_normalize_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_normalize_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_scatter_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_select_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_std_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_sum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_var_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_binary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_pool2d_with_indices_backward_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_reduction_with_dim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_list_of_tensors_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_list_of_tensors_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_variadic_tensors_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_reduction_no_dim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_minimum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mode_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_movedim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_msort_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mul_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mul_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nan_to_num_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nansum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_copy_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_copy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_dropout_backward_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ne_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_neg_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_strided_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_strided_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_full_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_zeros_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_batch_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv2d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cosine_similarity_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cross_entropy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_elu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_embedding_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_with_train_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_fractional_max_pool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_gaussian_nll_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_grid_sample_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_group_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardswish_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardtanh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_huber_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_instance_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_area_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_bicubic_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_nearest_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_trilinear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_l1_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_layer_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_leaky_relu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_logsigmoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool1d_grad_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multi_margin_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multilabel_margin_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_normalize_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_circular_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_reflect_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_replicate_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pairwise_distance_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pdist_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_shuffle_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_unshuffle_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_poisson_nll_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_prelu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_relu6_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_relu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_selu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_silu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_soft_margin_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softsign_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_threshold_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nonzero_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nonzero_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_fro_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_nuc_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_nuc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_in_place_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_number_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_like_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ormqr_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_outer_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_permute_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_permute_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_4_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_positive_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_positive_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pow_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_prod_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_qr_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_qr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rad2deg_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rand_like_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rand_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_like_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ravel_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_real_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reciprocal_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_remainder_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_renorm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_interleave_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_interleave_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize_as__cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize_as__cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_conj_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_neg_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_neg_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_roll_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rot90_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsqrt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsub_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scalar_tensor_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_sum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_searchsorted_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sgn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sgn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_short_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sign_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_bartlett_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_blackman_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_general_hamming_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_hamming_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_hann_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_kaiser_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_sampled_addmm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_sampled_addmm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_j0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_j1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_y1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_hermite_polynomial_h_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i0e_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i1e_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_legendre_polynomial_p_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_log_ndtr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_k0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_k1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_ndtri_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_polygamma_special_polygamma_n_0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_scaled_modified_bessel_k0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_scaled_modified_bessel_k1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_u_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_w_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_zeta_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_with_sizes_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sqrt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_square_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_multiple_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_unbiased_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_unbiased_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_unbiased_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sub_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_to_size_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_to_size_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_t_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_along_dim_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_along_dim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tan_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensor_split_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensordot_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensordot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_sparse_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_topk_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trace_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapezoid_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapezoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapz_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triangular_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tril_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_true_divide_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_true_divide_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unflatten_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unflatten_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_uniform_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unique_consecutive_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsqueeze_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_unbiased_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vdot_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vdot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_complex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vsplit_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vstack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vstack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_where_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_xlogy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_like_cpu_complex64 2023-03-31T07:00:29.5350478Z 2023-03-31T07:00:29.5350752Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_abs_cpu_float32 PASSED [ 0%] 2023-03-31T07:00:29.5351664Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_acosh_cpu_float32 PASSED [ 0%] 2023-03-31T07:00:29.5352340Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_asinh_cpu_float32 PASSED [ 0%] 2023-03-31T07:00:29.5352859Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atan_cpu_float32 PASSED [ 0%] 2023-03-31T07:00:29.5353351Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atanh_cpu_float32 PASSED [ 0%] 2023-03-31T07:00:29.5353826Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_cat_cpu_float32 XFAIL [ 1%] 2023-03-31T07:00:29.5354317Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_digamma_cpu_float32 PASSED [ 1%] 2023-03-31T07:00:29.5354869Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_floor_rounding_cpu_float32 PASSED [ 1%] 2023-03-31T07:00:29.5355416Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_no_rounding_mode_cpu_float32 PASSED [ 1%] 2023-03-31T07:00:29.5355935Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erf_cpu_float32 PASSED [ 1%] 2023-03-31T07:00:29.5356441Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erfc_cpu_float32 PASSED [ 2%] 2023-03-31T07:00:29.5356943Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_gt_cpu_float32 PASSED [ 2%] 2023-03-31T07:00:29.5357441Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_le_cpu_float32 PASSED [ 2%] 2023-03-31T07:00:29.5357960Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_householder_product_cpu_float32 PASSED [ 2%] 2023-03-31T07:00:29.5358497Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_inv_cpu_float32 PASSED [ 2%] 2023-03-31T07:00:29.5359037Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_matrix_power_cpu_float32 PASSED [ 3%] 2023-03-31T07:00:29.5359554Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log1p_cpu_float32 PASSED [ 3%] 2023-03-31T07:00:29.5360068Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log_softmax_cpu_float32 PASSED [ 3%] 2023-03-31T07:00:29.5360605Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_log_softmax_with_dtype_cpu_float32 PASSED [ 3%] 2023-03-31T07:00:29.5361132Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_logit_cpu_float32 PASSED [ 3%] 2023-03-31T07:00:29.5361639Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_logsumexp_cpu_float32 PASSED [ 4%] 2023-03-31T07:00:29.5362133Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mH_cpu_float32 PASSED [ 4%] 2023-03-31T07:00:29.5362633Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_max_binary_cpu_float32 XFAIL [ 4%] 2023-03-31T07:00:29.5363272Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mul_cpu_float32 PASSED [ 4%] 2023-03-31T07:00:29.5363830Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_5_cpu_float32 PASSED [ 4%] 2023-03-31T07:00:29.5364358Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_ne_cpu_float32 PASSED [ 5%] 2023-03-31T07:00:29.5364893Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv1d_cpu_float32 PASSED [ 5%] 2023-03-31T07:00:29.5365439Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv2d_cpu_float32 PASSED [ 5%] 2023-03-31T07:00:29.5366154Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose3d_cpu_float32 PASSED [ 5%] 2023-03-31T07:00:29.5366752Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_neg_3_cpu_float32 SKIPPED (Skipped!) [ 5%] 2023-03-31T07:00:29.5367311Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_softmax_cpu_float32 PASSED [ 6%] 2023-03-31T07:00:29.5367826Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_transpose_cpu_float32 PASSED [ 6%] 2023-03-31T07:00:29.5368418Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_trunc_cpu_float32 PASSED [ 6%] 2023-03-31T07:00:29.5368976Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_vstack_cpu_float32 XFAIL [ 6%] 2023-03-31T07:00:29.5369485Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_T_cpu_complex64 XFAIL [ 6%] 2023-03-31T07:00:29.5369998Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___getitem___cpu_complex64 PASSED [ 7%] 2023-03-31T07:00:29.5370598Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___getitem___cpu_float32 PASSED [ 7%] 2023-03-31T07:00:29.5371134Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rdiv___cpu_complex64 XFAIL [ 7%] 2023-03-31T07:00:29.5371659Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmatmul___cpu_complex64 XFAIL [ 7%] 2023-03-31T07:00:29.5372167Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmul___cpu_complex64 XFAIL [ 7%] 2023-03-31T07:00:29.5372669Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmul___cpu_float32 XFAIL [ 8%] 2023-03-31T07:00:29.5373210Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__native_batch_norm_legit_cpu_float32 XFAIL [ 8%] 2023-03-31T07:00:29.5373764Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__segment_reduce_lengths_cpu_float32 PASSED [ 8%] 2023-03-31T07:00:29.5374277Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_abs_cpu_float32 PASSED [ 8%] 2023-03-31T07:00:29.5374797Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acos_cpu_complex64 PASSED [ 8%] 2023-03-31T07:00:29.5375314Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acos_cpu_float32 PASSED [ 9%] 2023-03-31T07:00:29.5375852Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acosh_cpu_float32 PASSED [ 9%] 2023-03-31T07:00:29.5376386Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_add_cpu_complex64 PASSED [ 9%] 2023-03-31T07:00:29.5376922Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addbmm_cpu_complex64 PASSED [ 9%] 2023-03-31T07:00:29.5377454Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcdiv_cpu_complex64 PASSED [ 9%] 2023-03-31T07:00:29.5377970Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcmul_cpu_float32 PASSED [ 10%] 2023-03-31T07:00:29.5378481Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_cpu_complex64 PASSED [ 10%] 2023-03-31T07:00:29.5379005Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmv_cpu_float32 PASSED [ 10%] 2023-03-31T07:00:29.5379525Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_all_cpu_complex64 PASSED [ 10%] 2023-03-31T07:00:29.5380039Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_amax_cpu_float32 PASSED [ 10%] 2023-03-31T07:00:29.5380549Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_aminmax_cpu_float32 PASSED [ 11%] 2023-03-31T07:00:29.5381052Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_any_cpu_complex64 PASSED [ 11%] 2023-03-31T07:00:29.5381581Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argsort_cpu_float32 SKIPPED (Skipped!) [ 11%] 2023-03-31T07:00:29.5382112Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argwhere_cpu_complex64 PASSED [ 11%] 2023-03-31T07:00:29.5382741Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_cpu_complex64 PASSED [ 11%] 2023-03-31T07:00:29.5383288Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_cpu_float32 PASSED [ 12%] 2023-03-31T07:00:29.5383824Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_scatter_cpu_complex64 PASSED [ 12%] 2023-03-31T07:00:29.5384465Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_scatter_cpu_float32 PASSED [ 12%] 2023-03-31T07:00:29.5385031Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asin_cpu_complex64 PASSED [ 12%] 2023-03-31T07:00:29.5385576Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asinh_cpu_complex64 PASSED [ 12%] 2023-03-31T07:00:29.5386082Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atanh_cpu_complex64 PASSED [ 13%] 2023-03-31T07:00:29.5386697Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atanh_cpu_float32 PASSED [ 13%] 2023-03-31T07:00:29.5387265Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_2d_cpu_complex64 PASSED [ 13%] 2023-03-31T07:00:29.5387818Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_3d_cpu_complex64 PASSED [ 13%] 2023-03-31T07:00:29.5388425Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_3d_cpu_float32 XFAIL [ 13%] 2023-03-31T07:00:29.5388953Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_baddbmm_cpu_complex64 PASSED [ 14%] 2023-03-31T07:00:29.5389493Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bernoulli_cpu_float32 XFAIL [ 14%] 2023-03-31T07:00:29.5390021Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bfloat16_cpu_complex64 XFAIL [ 14%] 2023-03-31T07:00:29.5390560Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bfloat16_cpu_float32 XFAIL [ 14%] 2023-03-31T07:00:29.5391123Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_block_diag_cpu_complex64 PASSED [ 14%] 2023-03-31T07:00:29.5391708Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_shapes_cpu_float32 SKIPPED (Skipped!) [ 15%] 2023-03-31T07:00:29.5392279Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_byte_cpu_complex64 XFAIL [ 15%] 2023-03-31T07:00:29.5392792Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_byte_cpu_float32 XFAIL [ 15%] 2023-03-31T07:00:29.5393336Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cartesian_prod_cpu_float32 XFAIL [ 15%] 2023-03-31T07:00:29.5393861Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cat_cpu_complex64 PASSED [ 15%] 2023-03-31T07:00:29.5394376Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cauchy_cpu_float32 XFAIL [ 16%] 2023-03-31T07:00:29.5394907Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdist_cpu_float32 PASSED [ 16%] 2023-03-31T07:00:29.5395458Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdouble_cpu_complex64 XFAIL [ 16%] 2023-03-31T07:00:29.5395983Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chalf_cpu_complex64 XFAIL [ 16%] 2023-03-31T07:00:29.5396487Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_char_cpu_complex64 XFAIL [ 16%] 2023-03-31T07:00:29.5396990Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_char_cpu_float32 XFAIL [ 17%] 2023-03-31T07:00:29.5397538Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_solve_cpu_float32 PASSED [ 17%] 2023-03-31T07:00:29.5398079Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_min_cpu_float32 PASSED [ 17%] 2023-03-31T07:00:29.5398587Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clone_cpu_complex64 PASSED [ 17%] 2023-03-31T07:00:29.5399088Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clone_cpu_float32 PASSED [ 17%] 2023-03-31T07:00:29.5399603Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_column_stack_cpu_float32 PASSED [ 18%] 2023-03-31T07:00:29.5400128Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_combinations_cpu_float32 PASSED [ 18%] 2023-03-31T07:00:29.5400651Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_cpu_float32 PASSED [ 18%] 2023-03-31T07:00:29.5401172Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_physical_cpu_complex64 PASSED [ 18%] 2023-03-31T07:00:29.5401712Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_constant_pad_nd_cpu_float32 PASSED [ 18%] 2023-03-31T07:00:29.5402327Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cos_cpu_complex64 PASSED [ 19%] 2023-03-31T07:00:29.5402853Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cosh_cpu_complex64 PASSED [ 19%] 2023-03-31T07:00:29.5403533Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_count_nonzero_cpu_complex64 PASSED [ 19%] 2023-03-31T07:00:29.5404066Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cross_cpu_complex64 PASSED [ 19%] 2023-03-31T07:00:29.5404692Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cross_cpu_float32 PASSED [ 19%] 2023-03-31T07:00:29.5405253Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cummin_cpu_float32 PASSED [ 20%] 2023-03-31T07:00:29.5405774Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumprod_cpu_float32 PASSED [ 20%] 2023-03-31T07:00:29.5406330Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumsum_cpu_complex64 PASSED [ 20%] 2023-03-31T07:00:29.5406844Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumsum_cpu_float32 PASSED [ 20%] 2023-03-31T07:00:29.5407378Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumulative_trapezoid_cpu_complex64 PASSED [ 20%] 2023-03-31T07:00:29.5407932Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumulative_trapezoid_cpu_float32 PASSED [ 21%] 2023-03-31T07:00:29.5408449Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_deg2rad_cpu_float32 PASSED [ 21%] 2023-03-31T07:00:29.5408962Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagflat_cpu_complex64 PASSED [ 21%] 2023-03-31T07:00:29.5409493Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_copy_cpu_complex64 PASSED [ 21%] 2023-03-31T07:00:29.5410018Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_cpu_float32 PASSED [ 21%] 2023-03-31T07:00:29.5410545Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_scatter_cpu_complex64 PASSED [ 22%] 2023-03-31T07:00:29.5411072Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diff_cpu_complex64 PASSED [ 22%] 2023-03-31T07:00:29.5411574Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dist_cpu_float32 PASSED [ 22%] 2023-03-31T07:00:29.5412079Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_floor_rounding_cpu_float32 PASSED [ 22%] 2023-03-31T07:00:29.5412628Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_no_rounding_mode_cpu_complex64 PASSED [ 22%] 2023-03-31T07:00:29.5413180Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_no_rounding_mode_cpu_float32 PASSED [ 23%] 2023-03-31T07:00:29.5413701Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dot_cpu_complex64 PASSED [ 23%] 2023-03-31T07:00:29.5414190Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_double_cpu_complex64 XFAIL [ 23%] 2023-03-31T07:00:29.5414701Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dsplit_cpu_complex64 PASSED [ 23%] 2023-03-31T07:00:29.5415219Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dstack_cpu_complex64 PASSED [ 23%] 2023-03-31T07:00:29.5415737Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_einsum_cpu_complex64 SKIPPED (Skipped!) [ 24%] 2023-03-31T07:00:29.5416276Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_cpu_float32 SKIPPED (Skipped!) [ 24%] 2023-03-31T07:00:29.5416831Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_permuted_cpu_complex64 SKIPPED (Skipped!) [ 24%] 2023-03-31T07:00:29.5417406Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_permuted_cpu_float32 SKIPPED (Skipped!) [ 24%] 2023-03-31T07:00:29.5417909Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eq_cpu_float32 PASSED [ 24%] 2023-03-31T07:00:29.5418406Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_equal_cpu_complex64 PASSED [ 25%] 2023-03-31T07:00:29.5418906Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_equal_cpu_float32 PASSED [ 25%] 2023-03-31T07:00:29.5419399Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erf_cpu_float32 PASSED [ 25%] 2023-03-31T07:00:29.5419957Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp2_cpu_float32 PASSED [ 25%] 2023-03-31T07:00:29.5420456Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_cpu_complex64 PASSED [ 25%] 2023-03-31T07:00:29.5420962Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_cpu_float32 PASSED [ 25%] 2023-03-31T07:00:29.5421443Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expm1_cpu_complex64 PASSED [ 26%] 2023-03-31T07:00:29.5421987Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exponential_cpu_float32 XFAIL [ 26%] 2023-03-31T07:00:29.5422538Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft_cpu_complex64 PASSED [ 26%] 2023-03-31T07:00:29.5423130Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft_cpu_float32 PASSED [ 26%] 2023-03-31T07:00:29.5423675Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftn_cpu_complex64 PASSED [ 26%] 2023-03-31T07:00:29.5424182Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftn_cpu_float32 PASSED [ 27%] 2023-03-31T07:00:29.5424696Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftshift_cpu_complex64 PASSED [ 27%] 2023-03-31T07:00:29.5425224Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfftn_cpu_complex64 PASSED [ 27%] 2023-03-31T07:00:29.5425711Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfftn_cpu_float32 PASSED [ 27%] 2023-03-31T07:00:29.5426221Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft_cpu_complex64 PASSED [ 27%] 2023-03-31T07:00:29.5426722Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft_cpu_float32 PASSED [ 28%] 2023-03-31T07:00:29.5427224Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftshift_cpu_complex64 PASSED [ 28%] 2023-03-31T07:00:29.5427751Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftshift_cpu_float32 PASSED [ 28%] 2023-03-31T07:00:29.5428266Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfft_cpu_float32 PASSED [ 28%] 2023-03-31T07:00:29.5428773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft2_cpu_complex64 PASSED [ 28%] 2023-03-31T07:00:29.5429270Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft2_cpu_float32 PASSED [ 29%] 2023-03-31T07:00:29.5429785Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfftn_cpu_complex64 PASSED [ 29%] 2023-03-31T07:00:29.5430292Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fill_cpu_complex64 XFAIL [ 29%] 2023-03-31T07:00:29.5430778Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fill_cpu_float32 XFAIL [ 29%] 2023-03-31T07:00:29.5431274Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flatten_cpu_float32 PASSED [ 29%] 2023-03-31T07:00:29.5431777Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fliplr_cpu_complex64 PASSED [ 30%] 2023-03-31T07:00:29.5432287Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fliplr_cpu_float32 PASSED [ 30%] 2023-03-31T07:00:29.5432775Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flipud_cpu_float32 PASSED [ 30%] 2023-03-31T07:00:29.5433268Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_cpu_complex64 XFAIL [ 30%] 2023-03-31T07:00:29.5433763Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_cpu_float32 XFAIL [ 30%] 2023-03-31T07:00:29.5434284Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_power_cpu_complex64 SKIPPED (Skipped!) [ 31%] 2023-03-31T07:00:29.5434831Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_float_power_cpu_float32 SKIPPED (Skipped!) [ 31%] 2023-03-31T07:00:29.5435347Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmin_cpu_float32 PASSED [ 31%] 2023-03-31T07:00:29.5435833Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmod_cpu_float32 PASSED [ 31%] 2023-03-31T07:00:29.5436307Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_frac_cpu_float32 PASSED [ 31%] 2023-03-31T07:00:29.5436871Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_frexp_cpu_float32 PASSED [ 32%] 2023-03-31T07:00:29.5437382Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_cpu_float32 SKIPPED (Skipped!) [ 32%] 2023-03-31T07:00:29.5437903Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_like_cpu_float32 PASSED [ 32%] 2023-03-31T07:00:29.5438395Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geqrf_cpu_complex64 PASSED [ 32%] 2023-03-31T07:00:29.5438931Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geqrf_cpu_float32 PASSED [ 32%] 2023-03-31T07:00:29.5439484Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gradient_cpu_complex64 SKIPPED (Skipped!) [ 33%] 2023-03-31T07:00:29.5440034Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gradient_cpu_float32 SKIPPED (Skipped!) [ 33%] 2023-03-31T07:00:29.5440588Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_heaviside_cpu_float32 PASSED [ 33%] 2023-03-31T07:00:29.5441095Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histogram_cpu_float32 XFAIL [ 33%] 2023-03-31T07:00:29.5441605Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hsplit_cpu_complex64 PASSED [ 33%] 2023-03-31T07:00:29.5442094Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hsplit_cpu_float32 PASSED [ 34%] 2023-03-31T07:00:29.5442587Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_igammac_cpu_float32 PASSED [ 34%] 2023-03-31T07:00:29.5443241Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_copy_cpu_float32 PASSED [ 34%] 2023-03-31T07:00:29.5443766Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_fill_cpu_complex64 PASSED [ 34%] 2023-03-31T07:00:29.5444263Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_fill_cpu_float32 PASSED [ 34%] 2023-03-31T07:00:29.5444782Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_put_cpu_complex64 PASSED [ 35%] 2023-03-31T07:00:29.5445292Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_put_cpu_float32 PASSED [ 35%] 2023-03-31T07:00:29.5445793Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_select_cpu_float32 PASSED [ 35%] 2023-03-31T07:00:29.5446296Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_inner_cpu_complex64 PASSED [ 35%] 2023-03-31T07:00:29.5446791Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_int_cpu_complex64 XFAIL [ 35%] 2023-03-31T07:00:29.5447284Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isclose_cpu_float32 PASSED [ 36%] 2023-03-31T07:00:29.5447788Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isfinite_cpu_complex64 PASSED [ 36%] 2023-03-31T07:00:29.5448299Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isfinite_cpu_float32 PASSED [ 36%] 2023-03-31T07:00:29.5448795Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isinf_cpu_complex64 PASSED [ 36%] 2023-03-31T07:00:29.5449293Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isinf_cpu_float32 PASSED [ 36%] 2023-03-31T07:00:29.5449778Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isnan_cpu_complex64 PASSED [ 37%] 2023-03-31T07:00:29.5450341Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_2inputs_2outputs_cpu_complex64 SKIPPED (Only runs on cuda) [ 37%] 2023-03-31T07:00:29.5450974Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_4inputs_with_extra_args_cpu_float32 SKIPPED (Only runs on cuda) [ 37%] 2023-03-31T07:00:29.5451584Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_cpu_float32 SKIPPED (Only runs on cuda) [ 37%] 2023-03-31T07:00:29.5452111Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kron_cpu_complex64 PASSED [ 37%] 2023-03-31T07:00:29.5452603Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kthvalue_cpu_float32 PASSED [ 38%] 2023-03-31T07:00:29.5453107Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lerp_cpu_complex64 PASSED [ 38%] 2023-03-31T07:00:29.5453678Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lerp_cpu_float32 PASSED [ 38%] 2023-03-31T07:00:29.5454183Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lgamma_cpu_float32 PASSED [ 38%] 2023-03-31T07:00:29.5454694Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cond_cpu_complex64 PASSED [ 38%] 2023-03-31T07:00:29.5455217Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cond_cpu_float32 PASSED [ 39%] 2023-03-31T07:00:29.5455820Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_singular_cpu_complex64 PASSED [ 39%] 2023-03-31T07:00:29.5456407Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_diagonal_cpu_complex64 PASSED [ 39%] 2023-03-31T07:00:29.5456932Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_diagonal_cpu_float32 PASSED [ 39%] 2023-03-31T07:00:29.5457494Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eig_cpu_complex64 PASSED [ 39%] 2023-03-31T07:00:29.5458026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvalsh_cpu_complex64 PASSED [ 40%] 2023-03-31T07:00:29.5458579Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_householder_product_cpu_complex64 PASSED [ 40%] 2023-03-31T07:00:29.5459146Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_householder_product_cpu_float32 PASSED [ 40%] 2023-03-31T07:00:29.5459671Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_cpu_complex64 PASSED [ 40%] 2023-03-31T07:00:29.5460203Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_cpu_float32 PASSED [ 40%] 2023-03-31T07:00:29.5460739Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_ex_cpu_float32 PASSED [ 41%] 2023-03-31T07:00:29.5461261Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_solve_cpu_float32 PASSED [ 41%] 2023-03-31T07:00:29.5461789Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_grad_oriented_cpu_float32 XFAIL [ 41%] 2023-03-31T07:00:29.5462335Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_cpu_complex64 PASSED [ 41%] 2023-03-31T07:00:29.5462955Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_ex_cpu_float32 PASSED [ 41%] 2023-03-31T07:00:29.5463527Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_solve_cpu_float32 PASSED [ 42%] 2023-03-31T07:00:29.5464044Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_norm_cpu_complex64 PASSED [ 42%] 2023-03-31T07:00:29.5464589Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_norm_cpu_float32 PASSED [ 42%] 2023-03-31T07:00:29.5465146Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_cpu_float32 SKIPPED (Skipped!) [ 42%] 2023-03-31T07:00:29.5465704Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_hermitian_cpu_complex64 PASSED [ 42%] 2023-03-31T07:00:29.5466276Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_hermitian_cpu_float32 PASSED [ 43%] 2023-03-31T07:00:29.5466826Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_multi_dot_cpu_complex64 PASSED [ 43%] 2023-03-31T07:00:29.5467346Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_cpu_complex64 PASSED [ 43%] 2023-03-31T07:00:29.5467861Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_cpu_float32 PASSED [ 43%] 2023-03-31T07:00:29.5468418Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_subgradients_at_zero_cpu_complex64 PASSED [ 43%] 2023-03-31T07:00:29.5468981Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_cpu_float32 PASSED [ 44%] 2023-03-31T07:00:29.5469518Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_hermitian_cpu_complex64 PASSED [ 44%] 2023-03-31T07:00:29.5470055Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_hermitian_cpu_float32 PASSED [ 44%] 2023-03-31T07:00:29.5470798Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_singular_cpu_complex64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 44%] 2023-03-31T07:00:29.5471423Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_slogdet_cpu_complex64 PASSED [ 44%] 2023-03-31T07:00:29.5471950Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_slogdet_cpu_float32 PASSED [ 45%] 2023-03-31T07:00:29.5472467Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_cpu_complex64 PASSED [ 45%] 2023-03-31T07:00:29.5473043Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_ex_cpu_float32 PASSED [ 45%] 2023-03-31T07:00:29.5473618Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_triangular_cpu_complex64 PASSED [ 45%] 2023-03-31T07:00:29.5474164Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_triangular_cpu_float32 PASSED [ 45%] 2023-03-31T07:00:29.5474754Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svdvals_cpu_complex64 PASSED [ 46%] 2023-03-31T07:00:29.5475288Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorinv_cpu_complex64 PASSED [ 46%] 2023-03-31T07:00:29.5475829Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorsolve_cpu_complex64 PASSED [ 46%] 2023-03-31T07:00:29.5476349Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vecdot_cpu_float32 PASSED [ 46%] 2023-03-31T07:00:29.5476876Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vector_norm_cpu_float32 PASSED [ 46%] 2023-03-31T07:00:29.5477389Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linspace_cpu_float32 XFAIL [ 47%] 2023-03-31T07:00:29.5477896Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log10_cpu_complex64 PASSED [ 47%] 2023-03-31T07:00:29.5478374Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log10_cpu_float32 PASSED [ 47%] 2023-03-31T07:00:29.5478870Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log1p_cpu_float32 PASSED [ 47%] 2023-03-31T07:00:29.5479363Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log2_cpu_float32 PASSED [ 47%] 2023-03-31T07:00:29.5479834Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_cpu_complex64 PASSED [ 48%] 2023-03-31T07:00:29.5480322Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_cpu_float32 PASSED [ 48%] 2023-03-31T07:00:29.5480812Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_normal_cpu_float32 XFAIL [ 48%] 2023-03-31T07:00:29.5481354Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_with_dtype_cpu_float32 PASSED [ 48%] 2023-03-31T07:00:29.5482940Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logdet_cpu_complex64 SKIPPED (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/96979 for platform(s) linux, mac, macos, rocm, win. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 48%] 2023-03-31T07:00:29.5484082Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_not_cpu_complex64 PASSED [ 49%] 2023-03-31T07:00:29.5484588Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logspace_cpu_float32 XFAIL [ 49%] 2023-03-31T07:00:29.5485102Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logsumexp_cpu_float32 PASSED [ 49%] 2023-03-31T07:00:29.5485597Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_long_cpu_float32 XFAIL [ 49%] 2023-03-31T07:00:29.5486079Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_cpu_complex64 XFAIL [ 49%] 2023-03-31T07:00:29.5486584Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_solve_cpu_complex64 PASSED [ 50%] 2023-03-31T07:00:29.5487106Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_unpack_cpu_complex64 PASSED [ 50%] 2023-03-31T07:00:29.5487615Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mT_cpu_complex64 XFAIL [ 50%] 2023-03-31T07:00:29.5488238Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_amin_cpu_float32 XFAIL [ 50%] 2023-03-31T07:00:29.5488752Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_argmax_cpu_float32 XFAIL [ 50%] 2023-03-31T07:00:29.5489261Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_argmin_cpu_float32 XFAIL [ 50%] 2023-03-31T07:00:29.5489796Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumprod_cpu_complex64 SKIPPED (Skipped!) [ 51%] 2023-03-31T07:00:29.5490423Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumsum_cpu_complex64 SKIPPED (Skipped!) [ 51%] 2023-03-31T07:00:29.5491026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumsum_cpu_float32 SKIPPED (Skipped!) [ 51%] 2023-03-31T07:00:29.5491563Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_fill_cpu_float32 PASSED [ 51%] 2023-03-31T07:00:29.5492120Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_log_softmax_cpu_float32 XFAIL [ 51%] 2023-03-31T07:00:29.5492661Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_logsumexp_cpu_float32 SKIPPED (Skipped!) [ 52%] 2023-03-31T07:00:29.5493197Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_mean_cpu_float32 XFAIL [ 52%] 2023-03-31T07:00:29.5493695Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_norm_cpu_float32 XFAIL [ 52%] 2023-03-31T07:00:29.5494199Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_normalize_cpu_complex64 XFAIL [ 52%] 2023-03-31T07:00:29.5494718Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_normalize_cpu_float32 XFAIL [ 52%] 2023-03-31T07:00:29.5495251Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_scatter_cpu_complex64 PASSED [ 53%] 2023-03-31T07:00:29.5495770Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_scatter_cpu_float32 PASSED [ 53%] 2023-03-31T07:00:29.5496307Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_select_cpu_float32 PASSED [ 53%] 2023-03-31T07:00:29.5496823Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_softmax_cpu_float32 XFAIL [ 53%] 2023-03-31T07:00:29.5497325Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_std_cpu_float32 XFAIL [ 53%] 2023-03-31T07:00:29.5497811Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_sum_cpu_float32 XFAIL [ 54%] 2023-03-31T07:00:29.5498309Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_var_cpu_complex64 XFAIL [ 54%] 2023-03-31T07:00:29.5498819Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_binary_cpu_float32 PASSED [ 54%] 2023-03-31T07:00:29.5499358Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_pool2d_with_indices_backward_cpu_float32 XFAIL [ 54%] 2023-03-31T07:00:29.5499899Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_reduction_with_dim_cpu_float32 PASSED [ 54%] 2023-03-31T07:00:29.5500466Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_list_of_tensors_cpu_complex64 PASSED [ 55%] 2023-03-31T07:00:29.5501032Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_list_of_tensors_cpu_float32 PASSED [ 55%] 2023-03-31T07:00:29.5501605Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_variadic_tensors_cpu_float32 SKIPPED (Skipped!) [ 55%] 2023-03-31T07:00:29.5502177Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_reduction_no_dim_cpu_float32 PASSED [ 55%] 2023-03-31T07:00:29.5502773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_minimum_cpu_float32 PASSED [ 55%] 2023-03-31T07:00:29.5503267Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mode_cpu_float32 PASSED [ 56%] 2023-03-31T07:00:29.5503749Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_movedim_cpu_float32 PASSED [ 56%] 2023-03-31T07:00:29.5504244Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_msort_cpu_float32 PASSED [ 56%] 2023-03-31T07:00:29.5504796Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mul_cpu_complex64 PASSED [ 56%] 2023-03-31T07:00:29.5505290Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mul_cpu_float32 PASSED [ 56%] 2023-03-31T07:00:29.5505763Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mv_cpu_float32 PASSED [ 57%] 2023-03-31T07:00:29.5506251Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nan_to_num_cpu_float32 PASSED [ 57%] 2023-03-31T07:00:29.5506770Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nansum_cpu_float32 SKIPPED (Skipped!) [ 57%] 2023-03-31T07:00:29.5507318Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_copy_cpu_complex64 PASSED [ 57%] 2023-03-31T07:00:29.5507871Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_copy_cpu_float32 PASSED [ 57%] 2023-03-31T07:00:29.5508373Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_cpu_float32 PASSED [ 58%] 2023-03-31T07:00:29.5508970Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_dropout_backward_cpu_float32 SKIPPED (Skipped!) [ 58%] 2023-03-31T07:00:29.5509486Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ne_cpu_float32 PASSED [ 58%] 2023-03-31T07:00:29.5509966Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_neg_cpu_float32 PASSED [ 58%] 2023-03-31T07:00:29.5510487Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_cpu_complex64 SKIPPED (Skipped!) [ 58%] 2023-03-31T07:00:29.5511117Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_strided_cpu_complex64 SKIPPED (Expected: new_empty_strided is not comparable) [ 59%] 2023-03-31T07:00:29.5511789Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_strided_cpu_float32 SKIPPED (Expected: new_empty_strided is not comparable) [ 59%] 2023-03-31T07:00:29.5512372Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_full_cpu_float32 PASSED [ 59%] 2023-03-31T07:00:29.5512880Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_zeros_cpu_complex64 PASSED [ 59%] 2023-03-31T07:00:29.5513420Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool2d_cpu_float32 XFAIL [ 59%] 2023-03-31T07:00:29.5513996Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool3d_cpu_float32 XFAIL [ 60%] 2023-03-31T07:00:29.5514561Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool2d_cpu_float32 XFAIL [ 60%] 2023-03-31T07:00:29.5515135Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool3d_cpu_float32 XFAIL [ 60%] 2023-03-31T07:00:29.5515680Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool1d_cpu_float32 PASSED [ 60%] 2023-03-31T07:00:29.5516246Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool2d_cpu_float32 PASSED [ 60%] 2023-03-31T07:00:29.5516798Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_avg_pool3d_cpu_float32 PASSED [ 61%] 2023-03-31T07:00:29.5517348Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_batch_norm_cpu_float32 PASSED [ 61%] 2023-03-31T07:00:29.5517921Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv2d_cpu_complex64 SKIPPED (Works on some configs!) [ 61%] 2023-03-31T07:00:29.5518541Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose3d_cpu_float32 SKIPPED (Skipped!) [ 61%] 2023-03-31T07:00:29.5519138Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cosine_similarity_cpu_float32 PASSED [ 61%] 2023-03-31T07:00:29.5519711Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cross_entropy_cpu_float32 PASSED [ 62%] 2023-03-31T07:00:29.5520255Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout2d_cpu_float32 XFAIL [ 62%] 2023-03-31T07:00:29.5520790Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout_cpu_float32 XFAIL [ 62%] 2023-03-31T07:00:29.5521449Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_elu_cpu_float32 PASSED [ 62%] 2023-03-31T07:00:29.5522000Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_embedding_cpu_float32 XFAIL [ 62%] 2023-03-31T07:00:29.5522567Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_with_train_cpu_float32 XFAIL [ 63%] 2023-03-31T07:00:29.5523324Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_fractional_max_pool3d_cpu_float32 XFAIL [ 63%] 2023-03-31T07:00:29.5524020Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_gaussian_nll_loss_cpu_float32 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:00:29.5524647Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_grid_sample_cpu_float32 PASSED [ 63%] 2023-03-31T07:00:29.5525229Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_group_norm_cpu_float32 XFAIL [ 63%] 2023-03-31T07:00:29.5525785Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardswish_cpu_float32 PASSED [ 64%] 2023-03-31T07:00:29.5526336Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardtanh_cpu_float32 PASSED [ 64%] 2023-03-31T07:00:29.5526891Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_huber_loss_cpu_float32 SKIPPED (Skipped!) [ 64%] 2023-03-31T07:00:29.5527461Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_instance_norm_cpu_float32 XFAIL [ 64%] 2023-03-31T07:00:29.5528029Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_area_cpu_float32 XFAIL [ 64%] 2023-03-31T07:00:29.5528611Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_bicubic_cpu_float32 XFAIL [ 65%] 2023-03-31T07:00:29.5529170Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_nearest_cpu_float32 XFAIL [ 65%] 2023-03-31T07:00:29.5529766Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_trilinear_cpu_float32 XFAIL [ 65%] 2023-03-31T07:00:29.5530329Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_l1_loss_cpu_float32 XFAIL [ 65%] 2023-03-31T07:00:29.5530876Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_layer_norm_cpu_float32 PASSED [ 65%] 2023-03-31T07:00:29.5531411Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_leaky_relu_cpu_float32 PASSED [ 66%] 2023-03-31T07:00:29.5531969Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_logsigmoid_cpu_float32 PASSED [ 66%] 2023-03-31T07:00:29.5532526Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool1d_cpu_float32 PASSED [ 66%] 2023-03-31T07:00:29.5533091Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool1d_cpu_float32 PASSED [ 66%] 2023-03-31T07:00:29.5533641Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool1d_grad_cpu_float32 PASSED [ 66%] 2023-03-31T07:00:29.5534218Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multi_margin_loss_cpu_float32 PASSED [ 67%] 2023-03-31T07:00:29.5534805Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multilabel_margin_loss_cpu_float32 PASSED [ 67%] 2023-03-31T07:00:29.5535385Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_normalize_cpu_complex64 PASSED [ 67%] 2023-03-31T07:00:29.5535950Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_circular_cpu_complex64 PASSED [ 67%] 2023-03-31T07:00:29.5536525Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_reflect_cpu_complex64 PASSED [ 67%] 2023-03-31T07:00:29.5537117Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_replicate_cpu_float32 SKIPPED (Skipped!) [ 68%] 2023-03-31T07:00:29.5537737Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pairwise_distance_cpu_float32 SKIPPED (Skipped!) [ 68%] 2023-03-31T07:00:29.5538373Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pdist_cpu_float32 PASSED [ 68%] 2023-03-31T07:00:29.5538955Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_shuffle_cpu_float32 SKIPPED (Skipped!) [ 68%] 2023-03-31T07:00:29.5539561Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_unshuffle_cpu_float32 SKIPPED (Skipped!) [ 68%] 2023-03-31T07:00:29.5540192Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_poisson_nll_loss_cpu_float32 PASSED [ 69%] 2023-03-31T07:00:29.5540773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_prelu_cpu_float32 XFAIL [ 69%] 2023-03-31T07:00:29.5541315Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_relu6_cpu_float32 PASSED [ 69%] 2023-03-31T07:00:29.5541892Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_relu_cpu_float32 PASSED [ 69%] 2023-03-31T07:00:29.5542420Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_selu_cpu_float32 PASSED [ 69%] 2023-03-31T07:00:29.5543049Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_silu_cpu_float32 PASSED [ 70%] 2023-03-31T07:00:29.5543593Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_soft_margin_loss_cpu_float32 PASSED [ 70%] 2023-03-31T07:00:29.5544163Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_cpu_float32 PASSED [ 70%] 2023-03-31T07:00:29.5544735Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cpu_float32 PASSED [ 70%] 2023-03-31T07:00:29.5545309Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softsign_cpu_complex64 PASSED [ 70%] 2023-03-31T07:00:29.5545872Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_threshold_cpu_float32 PASSED [ 71%] 2023-03-31T07:00:29.5546458Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cpu_complex64 PASSED [ 71%] 2023-03-31T07:00:29.5547041Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cpu_float32 PASSED [ 71%] 2023-03-31T07:00:29.5547658Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cpu_float32 XFAIL [ 71%] 2023-03-31T07:00:29.5548221Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nonzero_cpu_complex64 XFAIL [ 71%] 2023-03-31T07:00:29.5548723Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nonzero_cpu_float32 XFAIL [ 72%] 2023-03-31T07:00:29.5549222Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_cpu_complex64 PASSED [ 72%] 2023-03-31T07:00:29.5549722Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_cpu_float32 XFAIL [ 72%] 2023-03-31T07:00:29.5550228Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_fro_cpu_complex64 XFAIL [ 72%] 2023-03-31T07:00:29.5550742Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_nuc_cpu_complex64 XFAIL [ 72%] 2023-03-31T07:00:29.5551233Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_nuc_cpu_float32 XFAIL [ 73%] 2023-03-31T07:00:29.5551749Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_in_place_cpu_float32 XFAIL [ 73%] 2023-03-31T07:00:29.5552269Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_number_mean_cpu_float32 XFAIL [ 73%] 2023-03-31T07:00:29.5552793Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_cpu_float32 SKIPPED (Skipped!) [ 73%] 2023-03-31T07:00:29.5553320Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_like_cpu_complex64 PASSED [ 73%] 2023-03-31T07:00:29.5553846Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_like_cpu_float32 PASSED [ 74%] 2023-03-31T07:00:29.5554362Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ormqr_cpu_complex64 PASSED [ 74%] 2023-03-31T07:00:29.5554920Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_outer_cpu_complex64 PASSED [ 74%] 2023-03-31T07:00:29.5555435Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_permute_cpu_complex64 PASSED [ 74%] 2023-03-31T07:00:29.5555952Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_permute_cpu_float32 PASSED [ 74%] 2023-03-31T07:00:29.5556510Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_2_cpu_float32 SKIPPED (Skipped!) [ 75%] 2023-03-31T07:00:29.5557155Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_3_cpu_float32 SKIPPED (Skipped!) [ 75%] 2023-03-31T07:00:29.5557778Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_4_cpu_float32 SKIPPED (Skipped!) [ 75%] 2023-03-31T07:00:29.5558341Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_positive_cpu_complex64 PASSED [ 75%] 2023-03-31T07:00:29.5558902Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_positive_cpu_float32 PASSED [ 75%] 2023-03-31T07:00:29.5559424Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pow_cpu_complex64 PASSED [ 75%] 2023-03-31T07:00:29.5559936Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_prod_cpu_complex64 PASSED [ 76%] 2023-03-31T07:00:29.5560446Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_qr_cpu_complex64 PASSED [ 76%] 2023-03-31T07:00:29.5560940Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_qr_cpu_float32 PASSED [ 76%] 2023-03-31T07:00:29.5561450Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rad2deg_cpu_float32 PASSED [ 76%] 2023-03-31T07:00:29.5561970Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rand_like_cpu_complex64 XFAIL [ 76%] 2023-03-31T07:00:29.5562494Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rand_like_cpu_float32 XFAIL [ 77%] 2023-03-31T07:00:29.5562997Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_like_cpu_complex64 XFAIL [ 77%] 2023-03-31T07:00:29.5563688Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ravel_cpu_complex64 PASSED [ 77%] 2023-03-31T07:00:29.5564199Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_real_cpu_complex64 PASSED [ 77%] 2023-03-31T07:00:29.5564705Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reciprocal_cpu_complex64 PASSED [ 77%] 2023-03-31T07:00:29.5565233Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_remainder_cpu_float32 PASSED [ 78%] 2023-03-31T07:00:29.5565742Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_renorm_cpu_float32 PASSED [ 78%] 2023-03-31T07:00:29.5566256Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_cpu_complex64 PASSED [ 78%] 2023-03-31T07:00:29.5566741Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_cpu_float32 PASSED [ 78%] 2023-03-31T07:00:29.5567290Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_interleave_cpu_complex64 SKIPPED (Skipped!) [ 78%] 2023-03-31T07:00:29.5567879Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_repeat_interleave_cpu_float32 SKIPPED (Skipped!) [ 79%] 2023-03-31T07:00:29.5568431Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_cpu_float32 PASSED [ 79%] 2023-03-31T07:00:29.5568936Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize_as__cpu_complex64 PASSED [ 79%] 2023-03-31T07:00:29.5569467Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize_as__cpu_float32 PASSED [ 79%] 2023-03-31T07:00:29.5569983Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_conj_cpu_float32 PASSED [ 79%] 2023-03-31T07:00:29.5570496Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_neg_cpu_complex64 PASSED [ 80%] 2023-03-31T07:00:29.5571022Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_neg_cpu_float32 PASSED [ 80%] 2023-03-31T07:00:29.5571533Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_roll_cpu_complex64 PASSED [ 80%] 2023-03-31T07:00:29.5572041Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rot90_cpu_float32 PASSED [ 80%] 2023-03-31T07:00:29.5572681Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_0_cpu_float32 PASSED [ 80%] 2023-03-31T07:00:29.5573207Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsqrt_cpu_float32 PASSED [ 81%] 2023-03-31T07:00:29.5573717Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsub_cpu_float32 PASSED [ 81%] 2023-03-31T07:00:29.5574227Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scalar_tensor_cpu_float32 XFAIL [ 81%] 2023-03-31T07:00:29.5574802Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_cpu_complex64 PASSED [ 81%] 2023-03-31T07:00:29.5575370Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_cpu_float32 PASSED [ 81%] 2023-03-31T07:00:29.5575906Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_sum_cpu_float32 PASSED [ 82%] 2023-03-31T07:00:29.5576516Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_searchsorted_cpu_float32 SKIPPED (Expected failure!) [ 82%] 2023-03-31T07:00:29.5577073Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_cpu_complex64 PASSED [ 82%] 2023-03-31T07:00:29.5577591Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sgn_cpu_complex64 PASSED [ 82%] 2023-03-31T07:00:29.5578098Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sgn_cpu_float32 PASSED [ 82%] 2023-03-31T07:00:29.5578582Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_short_cpu_complex64 XFAIL [ 83%] 2023-03-31T07:00:29.5579097Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sign_cpu_float32 PASSED [ 83%] 2023-03-31T07:00:29.5579632Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_bartlett_cpu_float32 XFAIL [ 83%] 2023-03-31T07:00:29.5580183Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_blackman_cpu_float32 XFAIL [ 83%] 2023-03-31T07:00:29.5580734Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_general_hamming_cpu_float32 XFAIL [ 83%] 2023-03-31T07:00:29.5581305Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_hamming_cpu_float32 XFAIL [ 84%] 2023-03-31T07:00:29.5581841Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_hann_cpu_float32 XFAIL [ 84%] 2023-03-31T07:00:29.5582375Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_kaiser_cpu_float32 XFAIL [ 84%] 2023-03-31T07:00:29.5582976Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinh_cpu_float32 PASSED [ 84%] 2023-03-31T07:00:29.5583848Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_cpu_complex64 SKIPPED (variant consistency doesn't work on torch.ops) [ 84%] 2023-03-31T07:00:29.5584473Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_sampled_addmm_cpu_complex64 SKIPPED (Skipped!) [ 85%] 2023-03-31T07:00:29.5585040Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_sampled_addmm_cpu_float32 SKIPPED (Skipped!) [ 85%] 2023-03-31T07:00:29.5585609Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_j0_cpu_float32 PASSED [ 85%] 2023-03-31T07:00:29.5586151Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_j1_cpu_float32 PASSED [ 85%] 2023-03-31T07:00:29.5586685Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_y1_cpu_float32 PASSED [ 85%] 2023-03-31T07:00:29.5587226Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_hermite_polynomial_h_cpu_float32 PASSED [ 86%] 2023-03-31T07:00:29.5587773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i0e_cpu_float32 PASSED [ 86%] 2023-03-31T07:00:29.5588289Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i1e_cpu_float32 PASSED [ 86%] 2023-03-31T07:00:29.5589131Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_legendre_polynomial_p_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 86%] 2023-03-31T07:00:29.5589770Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_log_ndtr_cpu_float32 PASSED [ 86%] 2023-03-31T07:00:29.5590423Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_k0_cpu_float32 PASSED [ 87%] 2023-03-31T07:00:29.5590995Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_k1_cpu_float32 PASSED [ 87%] 2023-03-31T07:00:29.5591537Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_ndtri_cpu_float32 PASSED [ 87%] 2023-03-31T07:00:29.5592143Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_polygamma_special_polygamma_n_0_cpu_float32 XFAIL [ 87%] 2023-03-31T07:00:29.5592773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_scaled_modified_bessel_k0_cpu_float32 PASSED [ 87%] 2023-03-31T07:00:29.5593365Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_scaled_modified_bessel_k1_cpu_float32 PASSED [ 88%] 2023-03-31T07:00:29.5594312Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_u_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 88%] 2023-03-31T07:00:29.5595288Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 88%] 2023-03-31T07:00:29.5595946Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_zeta_cpu_float32 PASSED [ 88%] 2023-03-31T07:00:29.5596471Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_cpu_complex64 PASSED [ 88%] 2023-03-31T07:00:29.5597006Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_with_sizes_cpu_float32 PASSED [ 89%] 2023-03-31T07:00:29.5597506Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sqrt_cpu_float32 PASSED [ 89%] 2023-03-31T07:00:29.5598027Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_square_cpu_complex64 PASSED [ 89%] 2023-03-31T07:00:29.5598561Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_multiple_cpu_complex64 PASSED [ 89%] 2023-03-31T07:00:29.5599084Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stack_cpu_complex64 PASSED [ 89%] 2023-03-31T07:00:29.5599577Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_cpu_float32 PASSED [ 90%] 2023-03-31T07:00:29.5600105Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_unbiased_cpu_float32 PASSED [ 90%] 2023-03-31T07:00:29.5600642Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_unbiased_cpu_complex64 PASSED [ 90%] 2023-03-31T07:00:29.5601156Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_unbiased_cpu_float32 PASSED [ 90%] 2023-03-31T07:00:29.5601764Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stft_cpu_float32 SKIPPED (Skipped! stft does not match the native function) [ 90%] 2023-03-31T07:00:29.5602338Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sub_cpu_complex64 PASSED [ 91%] 2023-03-31T07:00:29.5602859Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_to_size_cpu_complex64 PASSED [ 91%] 2023-03-31T07:00:29.5603522Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_to_size_cpu_float32 XFAIL [ 91%] 2023-03-31T07:00:29.5604025Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_cpu_float32 PASSED [ 91%] 2023-03-31T07:00:29.5604526Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_t_cpu_float32 PASSED [ 91%] 2023-03-31T07:00:29.5605049Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_along_dim_cpu_complex64 PASSED [ 92%] 2023-03-31T07:00:29.5605571Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_along_dim_cpu_float32 PASSED [ 92%] 2023-03-31T07:00:29.5606096Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_cpu_complex64 PASSED [ 92%] 2023-03-31T07:00:29.5606588Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_take_cpu_float32 PASSED [ 92%] 2023-03-31T07:00:29.5607076Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tan_cpu_float32 PASSED [ 92%] 2023-03-31T07:00:29.5607703Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensor_split_cpu_float32 PASSED [ 93%] 2023-03-31T07:00:29.5608234Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensordot_cpu_complex64 PASSED [ 93%] 2023-03-31T07:00:29.5608751Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensordot_cpu_float32 PASSED [ 93%] 2023-03-31T07:00:29.5609252Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_sparse_cpu_complex64 PASSED [ 93%] 2023-03-31T07:00:29.5609829Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_topk_cpu_float32 PASSED [ 93%] 2023-03-31T07:00:29.5610373Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trace_cpu_float32 PASSED [ 94%] 2023-03-31T07:00:29.5610890Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapezoid_cpu_complex64 PASSED [ 94%] 2023-03-31T07:00:29.5611459Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapezoid_cpu_float32 PASSED [ 94%] 2023-03-31T07:00:29.5611982Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapz_cpu_complex64 PASSED [ 94%] 2023-03-31T07:00:29.5612513Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triangular_solve_cpu_complex64 PASSED [ 94%] 2023-03-31T07:00:29.5613026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tril_cpu_complex64 PASSED [ 95%] 2023-03-31T07:00:29.5613529Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triu_cpu_float32 PASSED [ 95%] 2023-03-31T07:00:29.5614045Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_true_divide_cpu_complex64 PASSED [ 95%] 2023-03-31T07:00:29.5614576Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_true_divide_cpu_float32 PASSED [ 95%] 2023-03-31T07:00:29.5615089Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unflatten_cpu_complex64 PASSED [ 95%] 2023-03-31T07:00:29.5615610Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unflatten_cpu_float32 PASSED [ 96%] 2023-03-31T07:00:29.5616126Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_cpu_float32 PASSED [ 96%] 2023-03-31T07:00:29.5616622Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_uniform_cpu_float32 XFAIL [ 96%] 2023-03-31T07:00:29.5617128Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unique_consecutive_cpu_float32 XFAIL [ 96%] 2023-03-31T07:00:29.5617651Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsqueeze_cpu_complex64 PASSED [ 96%] 2023-03-31T07:00:29.5618178Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_unbiased_cpu_float32 PASSED [ 97%] 2023-03-31T07:00:29.5618677Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vdot_cpu_complex64 PASSED [ 97%] 2023-03-31T07:00:29.5619175Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vdot_cpu_float32 PASSED [ 97%] 2023-03-31T07:00:29.5619683Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_complex_cpu_float32 PASSED [ 97%] 2023-03-31T07:00:29.5620198Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_cpu_complex64 PASSED [ 97%] 2023-03-31T07:00:29.5620695Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_cpu_float32 PASSED [ 98%] 2023-03-31T07:00:29.5621200Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_cpu_complex64 PASSED [ 98%] 2023-03-31T07:00:29.5621694Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_cpu_float32 PASSED [ 98%] 2023-03-31T07:00:29.5622197Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vsplit_cpu_complex64 PASSED [ 98%] 2023-03-31T07:00:29.5622771Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vstack_cpu_complex64 PASSED [ 98%] 2023-03-31T07:00:29.5623280Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vstack_cpu_float32 PASSED [ 99%] 2023-03-31T07:00:29.5623803Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_where_cpu_float32 SKIPPED (Skipped!) [ 99%] 2023-03-31T07:00:29.5624300Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_xlogy_cpu_float32 PASSED [ 99%] 2023-03-31T07:00:29.5624906Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_cpu_complex64 SKIPPED (Skipped!) [ 99%] 2023-03-31T07:00:29.5625453Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_cpu_float32 SKIPPED (Skipped!) [ 99%] 2023-03-31T07:00:29.5625992Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_like_cpu_complex64 PASSED [100%] 2023-03-31T07:00:29.5626279Z 2023-03-31T07:00:29.5626787Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_jit/test_ops_jit-ad78045603f8ca6d.xml - 2023-03-31T07:00:29.5627367Z ==== 361 passed, 53 skipped, 3 deselected, 90 xfailed in 530.78s (0:08:50) ===== 2023-03-31T07:00:29.5627885Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:00:29.5628170Z 2023-03-31T07:00:29.5628634Z ##[endgroup] 2023-03-31T07:00:29.5629256Z FINISHED PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_npw9qolt.log) 2023-03-31T07:00:29.5629578Z 2023-03-31T07:01:24.5548418Z 2023-03-31T07:01:24.5549228Z Expand the folded group to see the log file of test_ops_jit 2023-03-31T07:01:24.5550178Z ##[group]PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_lqzk42va.log) 2023-03-31T07:01:24.5551104Z Test results will be stored in test-reports/python-pytest/test_ops_jit/test_ops_jit-09b7eaa136a34551.xml 2023-03-31T07:01:24.5551654Z ============================= test session starts ============================== 2023-03-31T07:01:24.5552197Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:01:24.5552661Z cachedir: .pytest_cache 2023-03-31T07:01:24.5555114Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:01:24.5555664Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:01:24.5556461Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:01:24.5556998Z collecting ... collected 1079 items / 1 deselected / 1078 selected 2023-03-31T07:01:24.5622002Z Running 571 items in this shard: test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_acos_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_asin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atan2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_clamp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_trunc_rounding_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erfinv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_exp2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_expm1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_ge_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_i0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_igamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_igammac_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_lgamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_det_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_det_singular_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_lt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_matmul_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_matrix_exp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_min_binary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_movedim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_neg_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_group_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_layer_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_outer_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sigmoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sinc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_softmax_with_dtype_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sub_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_tanh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_xlogy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_H_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_H_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_T_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___radd___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___radd___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rdiv___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmatmul___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmod___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rpow___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rpow___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rsub___cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rsub___cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__segment_reduce_offsets_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__softmax_backward_data_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__upsample_bilinear2d_aa_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_abs_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acosh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_add_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addbmm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcdiv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcmul_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_decomposed_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_decomposed_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addr_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_all_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_allclose_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_allclose_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_amin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_angle_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_angle_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_any_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_arange_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argmin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argwhere_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_partial_views_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_partial_views_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asinh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_1d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_baddbmm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_block_diag_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bmm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bmm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bool_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bool_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_tensors_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_tensors_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_to_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_to_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bucketize_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cartesian_prod_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cat_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdouble_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ceil_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cfloat_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cfloat_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chalf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_inverse_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_inverse_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chunk_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chunk_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_max_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_column_stack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_combinations_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_complex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_physical_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_constant_pad_nd_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_contiguous_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_contiguous_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_copysign_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_corrcoef_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_corrcoef_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cos_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cosh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_count_nonzero_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cov_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cov_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cummax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumprod_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_embed_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_embed_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagflat_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_copy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diff_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_digamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dist_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_trunc_rounding_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_double_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dsplit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dstack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_einsum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_like_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eq_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erfc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erfinv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_as_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_as_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expm1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eye_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eye_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftshift_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfft2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfft_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfftn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flatten_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flip_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flip_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flipud_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_floor_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_floor_divide_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_like_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gather_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gather_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ge_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geometric_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_grid_sampler_2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_half_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_half_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histogramdd_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hstack_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hstack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hypot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_i0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_igamma_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_imag_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_add_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_add_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_copy_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_reduce_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_select_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_inner_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_int_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isclose_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isnan_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isneginf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isposinf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isreal_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isreal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_istft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_2inputs_2outputs_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_4inputs_with_extra_args_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_return_by_ref_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_return_by_ref_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_unary_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_unary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kron_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ldexp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ldexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_le_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cross_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cross_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_singular_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eig_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvals_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvals_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvalsh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_ex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_ex_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_ex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_grad_oriented_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_ex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_solve_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_power_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_power_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_multi_dot_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_subgradients_at_zero_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_singular_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_qr_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_qr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_ex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svd_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svd_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svdvals_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorinv_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorsolve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vander_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vander_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vecdot_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vector_norm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linspace_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log1p_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log2_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_with_dtype_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp2_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logcumsumexp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logcumsumexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logdet_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_and_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_and_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_not_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_or_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_or_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_xor_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_xor_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logspace_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_long_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lt_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_unpack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mH_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mH_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mT_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_amax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumprod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_fill_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_logaddexp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_mean_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_median_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_prod_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_prod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_select_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_softmin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_std_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_sum_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_var_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matmul_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matmul_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matrix_exp_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matrix_exp_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_reduction_no_dim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_maximum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mean_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_median_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_variadic_tensors_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_binary_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_reduction_with_dim_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_movedim_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_multinomial_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mv_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_5_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanmean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanmedian_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanquantile_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_batch_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_layer_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ne_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_neg_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_full_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_ones_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_ones_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_zeros_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nextafter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_alpha_dropout_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_bilinear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_binary_cross_entropy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_binary_cross_entropy_with_logits_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_celu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv1d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose1d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose1d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose2d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose3d_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cosine_embedding_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_ctc_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_embedding_bag_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_without_train_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_without_train_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_fractional_max_pool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_gelu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_glu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardshrink_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardsigmoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hinge_embedding_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_bilinear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_linear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_kl_div_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_l1_loss_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_linear_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_linear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_local_response_norm_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_margin_ranking_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool2d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool2d_grad_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool3d_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool3d_grad_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_mish_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_mse_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multilabel_soft_margin_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_nll_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_normalize_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_circular_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_constant_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_constant_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_reflect_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_replicate_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pairwise_distance_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_shuffle_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_unshuffle_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_rrelu_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_scaled_dot_product_attention_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_silu_complex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_smooth_l1_loss_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softplus_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softshrink_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softsign_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_tanhshrink_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_tanhshrink_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_unfold_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_unfold_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_upsample_bilinear_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_upsample_nearest_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_fro_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_inf_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_inf_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_in_place_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ormqr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_outer_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pca_lowrank_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pinverse_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pinverse_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polar_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pow_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_prod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_put_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_put_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_quantile_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randint_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randint_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_like_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ravel_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_real_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reciprocal_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_renorm_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_as_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_as_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize__cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize__cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_conj_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_roll_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rot90_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_neg_3_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsqrt_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsub_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scalar_tensor_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_add_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_add_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_amax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_amin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_prod_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_short_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sigmoid_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sigmoid_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_cosine_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_exponential_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_gaussian_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_general_cosine_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_nuttall_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signbit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sin_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sin_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinc_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_scatter_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_with_dtype_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_with_dtype_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sort_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_mm_reduce_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_airy_ai_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_y0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_t_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_u_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_v_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_w_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_entr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_erfcx_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_hermite_polynomial_he_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_laguerre_polynomial_l_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_i0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_i1_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_ndtr_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_t_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_v_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_spherical_bessel_j0_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_xlog1py_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_list_args_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_list_args_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_with_sizes_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sqrt_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_square_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_multiple_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stack_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_unbiased_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stft_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sub_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_lowrank_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_t_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tan_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tanh_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tanh_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensor_split_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tile_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tile_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_sparse_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trace_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_transpose_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_transpose_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapz_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triangular_solve_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tril_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triu_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trunc_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unbind_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unbind_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_copy_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_copy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_uniform_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unique_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsafe_split_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsafe_split_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsqueeze_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_unbiased_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_unbiased_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_unbiased_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_real_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_copy_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vsplit_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_where_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zero__cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zero__cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_like_cpu_float32 2023-03-31T07:01:24.5686044Z 2023-03-31T07:01:24.5686564Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_acos_cpu_float32 PASSED [ 0%] 2023-03-31T07:01:24.5686934Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_asin_cpu_float32 PASSED [ 0%] 2023-03-31T07:01:24.5687286Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_atan2_cpu_float32 XFAIL [ 0%] 2023-03-31T07:01:24.5687635Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_clamp_cpu_float32 PASSED [ 0%] 2023-03-31T07:01:24.5688057Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_div_trunc_rounding_cpu_float32 PASSED [ 0%] 2023-03-31T07:01:24.5688474Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_erfinv_cpu_float32 PASSED [ 1%] 2023-03-31T07:01:24.5688826Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_exp2_cpu_float32 PASSED [ 1%] 2023-03-31T07:01:24.5689226Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_expm1_cpu_float32 PASSED [ 1%] 2023-03-31T07:01:24.5689556Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_ge_cpu_float32 PASSED [ 1%] 2023-03-31T07:01:24.5689898Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_i0_cpu_float32 PASSED [ 1%] 2023-03-31T07:01:24.5690242Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_igamma_cpu_float32 XFAIL [ 1%] 2023-03-31T07:01:24.5690572Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_igammac_cpu_float32 XFAIL [ 2%] 2023-03-31T07:01:24.5690917Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_lgamma_cpu_float32 PASSED [ 2%] 2023-03-31T07:01:24.5691273Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_det_cpu_float32 PASSED [ 2%] 2023-03-31T07:01:24.5691651Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_linalg_det_singular_cpu_float32 PASSED [ 2%] 2023-03-31T07:01:24.5692003Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_lt_cpu_float32 PASSED [ 2%] 2023-03-31T07:01:24.5692345Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_matmul_cpu_float32 PASSED [ 2%] 2023-03-31T07:01:24.5692700Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_matrix_exp_cpu_float32 PASSED [ 3%] 2023-03-31T07:01:24.5693055Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_min_binary_cpu_float32 XFAIL [ 3%] 2023-03-31T07:01:24.5693392Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_movedim_cpu_float32 PASSED [ 3%] 2023-03-31T07:01:24.5693758Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_1_cpu_float32 PASSED [ 3%] 2023-03-31T07:01:24.5694142Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_mvlgamma_mvlgamma_p_3_cpu_float32 PASSED [ 3%] 2023-03-31T07:01:24.5694486Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_neg_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5694867Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose1d_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5695268Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_conv_transpose2d_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5695668Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_group_norm_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5696044Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_nn_functional_layer_norm_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5696410Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_outer_cpu_float32 PASSED [ 4%] 2023-03-31T07:01:24.5696755Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_cpu_float32 PASSED [ 5%] 2023-03-31T07:01:24.5697112Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_0_cpu_float32 PASSED [ 5%] 2023-03-31T07:01:24.5697491Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_round_decimals_3_cpu_float32 SKIPPED (Skipped!) [ 5%] 2023-03-31T07:01:24.5697865Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sigmoid_cpu_float32 PASSED [ 5%] 2023-03-31T07:01:24.5698214Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sinc_cpu_float32 PASSED [ 5%] 2023-03-31T07:01:24.5698608Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_softmax_with_dtype_cpu_float32 PASSED [ 5%] 2023-03-31T07:01:24.5698968Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_sub_cpu_float32 PASSED [ 6%] 2023-03-31T07:01:24.5699313Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_tanh_cpu_float32 PASSED [ 6%] 2023-03-31T07:01:24.5699660Z test_ops_jit.py::TestJitCPU::test_jit_alias_remapping_xlogy_cpu_float32 PASSED [ 6%] 2023-03-31T07:01:24.5699994Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_H_cpu_complex64 XFAIL [ 6%] 2023-03-31T07:01:24.5700369Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_H_cpu_float32 XFAIL [ 6%] 2023-03-31T07:01:24.5700756Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_T_cpu_float32 XFAIL [ 7%] 2023-03-31T07:01:24.5701093Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___radd___cpu_complex64 XFAIL [ 7%] 2023-03-31T07:01:24.5701481Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___radd___cpu_float32 XFAIL [ 7%] 2023-03-31T07:01:24.5701833Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rdiv___cpu_float32 XFAIL [ 7%] 2023-03-31T07:01:24.5702189Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmatmul___cpu_float32 XFAIL [ 7%] 2023-03-31T07:01:24.5702532Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rmod___cpu_float32 XFAIL [ 7%] 2023-03-31T07:01:24.5702974Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rpow___cpu_complex64 XFAIL [ 8%] 2023-03-31T07:01:24.5703332Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rpow___cpu_float32 XFAIL [ 8%] 2023-03-31T07:01:24.5703689Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rsub___cpu_complex64 XFAIL [ 8%] 2023-03-31T07:01:24.5704028Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit___rsub___cpu_float32 XFAIL [ 8%] 2023-03-31T07:01:24.5704407Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__segment_reduce_offsets_cpu_float32 PASSED [ 8%] 2023-03-31T07:01:24.5704804Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__softmax_backward_data_cpu_float32 XFAIL [ 8%] 2023-03-31T07:01:24.5705183Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit__upsample_bilinear2d_aa_cpu_float32 XFAIL [ 9%] 2023-03-31T07:01:24.5705554Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_abs_cpu_complex64 PASSED [ 9%] 2023-03-31T07:01:24.5705914Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_acosh_cpu_complex64 PASSED [ 9%] 2023-03-31T07:01:24.5706269Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_add_cpu_float32 PASSED [ 9%] 2023-03-31T07:01:24.5706610Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addbmm_cpu_float32 PASSED [ 9%] 2023-03-31T07:01:24.5706968Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcdiv_cpu_float32 PASSED [ 9%] 2023-03-31T07:01:24.5707325Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addcmul_cpu_complex64 PASSED [ 10%] 2023-03-31T07:01:24.5707671Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_cpu_float32 PASSED [ 10%] 2023-03-31T07:01:24.5708044Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_decomposed_cpu_complex64 PASSED [ 10%] 2023-03-31T07:01:24.5708432Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmm_decomposed_cpu_float32 PASSED [ 10%] 2023-03-31T07:01:24.5708801Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addmv_cpu_complex64 PASSED [ 10%] 2023-03-31T07:01:24.5709147Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addr_cpu_complex64 PASSED [ 11%] 2023-03-31T07:01:24.5709505Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_addr_cpu_float32 PASSED [ 11%] 2023-03-31T07:01:24.5709858Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_all_cpu_float32 PASSED [ 11%] 2023-03-31T07:01:24.5710232Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_allclose_cpu_complex64 SKIPPED (Skipped!) [ 11%] 2023-03-31T07:01:24.5710610Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_allclose_cpu_float32 SKIPPED (Skipped!) [ 11%] 2023-03-31T07:01:24.5711028Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_amin_cpu_float32 PASSED [ 11%] 2023-03-31T07:01:24.5711384Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_angle_cpu_complex64 PASSED [ 12%] 2023-03-31T07:01:24.5711726Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_angle_cpu_float32 PASSED [ 12%] 2023-03-31T07:01:24.5712075Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_any_cpu_float32 PASSED [ 12%] 2023-03-31T07:01:24.5712452Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_arange_cpu_float32 XFAIL [ 12%] 2023-03-31T07:01:24.5712836Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argmax_cpu_float32 PASSED [ 12%] 2023-03-31T07:01:24.5713175Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argmin_cpu_float32 PASSED [ 12%] 2023-03-31T07:01:24.5713562Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_argwhere_cpu_float32 PASSED [ 13%] 2023-03-31T07:01:24.5713947Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_partial_views_cpu_complex64 XFAIL [ 13%] 2023-03-31T07:01:24.5714349Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_as_strided_partial_views_cpu_float32 XFAIL [ 13%] 2023-03-31T07:01:24.5714702Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asin_cpu_float32 PASSED [ 13%] 2023-03-31T07:01:24.5715053Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_asinh_cpu_float32 PASSED [ 13%] 2023-03-31T07:01:24.5715402Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan2_cpu_float32 PASSED [ 14%] 2023-03-31T07:01:24.5715748Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan_cpu_complex64 PASSED [ 14%] 2023-03-31T07:01:24.5716100Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atan_cpu_float32 PASSED [ 14%] 2023-03-31T07:01:24.5716462Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_1d_cpu_complex64 PASSED [ 14%] 2023-03-31T07:01:24.5716832Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_1d_cpu_float32 XFAIL [ 14%] 2023-03-31T07:01:24.5717176Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_atleast_2d_cpu_float32 XFAIL [ 14%] 2023-03-31T07:01:24.5717531Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_baddbmm_cpu_float32 PASSED [ 15%] 2023-03-31T07:01:24.5717883Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_block_diag_cpu_float32 XFAIL [ 15%] 2023-03-31T07:01:24.5718222Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bmm_cpu_complex64 PASSED [ 15%] 2023-03-31T07:01:24.5718571Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bmm_cpu_float32 PASSED [ 15%] 2023-03-31T07:01:24.5718920Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bool_cpu_complex64 XFAIL [ 15%] 2023-03-31T07:01:24.5719266Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bool_cpu_float32 XFAIL [ 15%] 2023-03-31T07:01:24.5719622Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_tensors_cpu_complex64 PASSED [ 16%] 2023-03-31T07:01:24.5720007Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_tensors_cpu_float32 XFAIL [ 16%] 2023-03-31T07:01:24.5720391Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_to_cpu_complex64 PASSED [ 16%] 2023-03-31T07:01:24.5720766Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_broadcast_to_cpu_float32 PASSED [ 16%] 2023-03-31T07:01:24.5721147Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_bucketize_cpu_float32 SKIPPED (Expected failure!) [ 16%] 2023-03-31T07:01:24.5721546Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cartesian_prod_cpu_complex64 PASSED [ 16%] 2023-03-31T07:01:24.5721912Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cat_cpu_float32 PASSED [ 17%] 2023-03-31T07:01:24.5722248Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cdouble_cpu_float32 XFAIL [ 17%] 2023-03-31T07:01:24.5722604Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ceil_cpu_float32 PASSED [ 17%] 2023-03-31T07:01:24.5722992Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cfloat_cpu_complex64 XFAIL [ 17%] 2023-03-31T07:01:24.5723553Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cfloat_cpu_float32 XFAIL [ 17%] 2023-03-31T07:01:24.5723892Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chalf_cpu_float32 XFAIL [ 18%] 2023-03-31T07:01:24.5724256Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_cpu_complex64 PASSED [ 18%] 2023-03-31T07:01:24.5724704Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_cpu_float32 PASSED [ 18%] 2023-03-31T07:01:24.5725115Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_inverse_cpu_complex64 PASSED [ 18%] 2023-03-31T07:01:24.5725497Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_inverse_cpu_float32 PASSED [ 18%] 2023-03-31T07:01:24.5725925Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cholesky_solve_cpu_complex64 PASSED [ 18%] 2023-03-31T07:01:24.5726303Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chunk_cpu_complex64 PASSED [ 19%] 2023-03-31T07:01:24.5726645Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_chunk_cpu_float32 PASSED [ 19%] 2023-03-31T07:01:24.5727006Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_cpu_float32 PASSED [ 19%] 2023-03-31T07:01:24.5727363Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_clamp_max_cpu_float32 PASSED [ 19%] 2023-03-31T07:01:24.5727735Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_column_stack_cpu_complex64 PASSED [ 19%] 2023-03-31T07:01:24.5728105Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_combinations_cpu_complex64 PASSED [ 19%] 2023-03-31T07:01:24.5740706Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_complex_cpu_float32 PASSED [ 20%] 2023-03-31T07:01:24.5741076Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_cpu_complex64 PASSED [ 20%] 2023-03-31T07:01:24.5741487Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_conj_physical_cpu_float32 SKIPPED (Skipped!) [ 20%] 2023-03-31T07:01:24.5741876Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_constant_pad_nd_cpu_complex64 PASSED [ 20%] 2023-03-31T07:01:24.5742258Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_contiguous_cpu_complex64 PASSED [ 20%] 2023-03-31T07:01:24.5742754Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_contiguous_cpu_float32 PASSED [ 21%] 2023-03-31T07:01:24.5743126Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_copysign_cpu_float32 PASSED [ 21%] 2023-03-31T07:01:24.5743484Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_corrcoef_cpu_complex64 PASSED [ 21%] 2023-03-31T07:01:24.5743848Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_corrcoef_cpu_float32 PASSED [ 21%] 2023-03-31T07:01:24.5744202Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cos_cpu_float32 PASSED [ 21%] 2023-03-31T07:01:24.5744549Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cosh_cpu_float32 PASSED [ 21%] 2023-03-31T07:01:24.5744915Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_count_nonzero_cpu_float32 PASSED [ 22%] 2023-03-31T07:01:24.5745276Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cov_cpu_complex64 XFAIL [ 22%] 2023-03-31T07:01:24.5745628Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cov_cpu_float32 XFAIL [ 22%] 2023-03-31T07:01:24.5745965Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cummax_cpu_float32 PASSED [ 22%] 2023-03-31T07:01:24.5746321Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_cumprod_cpu_complex64 PASSED [ 22%] 2023-03-31T07:01:24.5746685Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_cpu_complex64 PASSED [ 22%] 2023-03-31T07:01:24.5747044Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_cpu_float32 PASSED [ 23%] 2023-03-31T07:01:24.5747395Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_embed_cpu_complex64 PASSED [ 23%] 2023-03-31T07:01:24.5747982Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diag_embed_cpu_float32 PASSED [ 23%] 2023-03-31T07:01:24.5748348Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagflat_cpu_float32 PASSED [ 23%] 2023-03-31T07:01:24.5748699Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_copy_cpu_float32 PASSED [ 23%] 2023-03-31T07:01:24.5749077Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_cpu_complex64 PASSED [ 23%] 2023-03-31T07:01:24.5749515Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diagonal_scatter_cpu_float32 PASSED [ 24%] 2023-03-31T07:01:24.5749944Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_diff_cpu_float32 PASSED [ 24%] 2023-03-31T07:01:24.5750284Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_digamma_cpu_float32 PASSED [ 24%] 2023-03-31T07:01:24.5750645Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dist_cpu_complex64 PASSED [ 24%] 2023-03-31T07:01:24.5751071Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_div_trunc_rounding_cpu_float32 PASSED [ 24%] 2023-03-31T07:01:24.5751422Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dot_cpu_float32 PASSED [ 25%] 2023-03-31T07:01:24.5751773Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_double_cpu_float32 XFAIL [ 25%] 2023-03-31T07:01:24.5752125Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dsplit_cpu_float32 PASSED [ 25%] 2023-03-31T07:01:24.5752482Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_dstack_cpu_float32 PASSED [ 25%] 2023-03-31T07:01:24.5752839Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_einsum_cpu_float32 SKIPPED (Skipped!) [ 25%] 2023-03-31T07:01:24.5753227Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_cpu_complex64 SKIPPED (Skipped!) [ 25%] 2023-03-31T07:01:24.5753623Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_like_cpu_complex64 SKIPPED (Skipped!) [ 26%] 2023-03-31T07:01:24.5754026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_empty_like_cpu_float32 SKIPPED (Skipped!) [ 26%] 2023-03-31T07:01:24.5754388Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eq_cpu_complex64 PASSED [ 26%] 2023-03-31T07:01:24.5754738Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erfc_cpu_float32 PASSED [ 26%] 2023-03-31T07:01:24.5755089Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_erfinv_cpu_float32 PASSED [ 26%] 2023-03-31T07:01:24.5755431Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp2_cpu_complex64 PASSED [ 26%] 2023-03-31T07:01:24.5755790Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp_cpu_complex64 PASSED [ 27%] 2023-03-31T07:01:24.5756145Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_exp_cpu_float32 PASSED [ 27%] 2023-03-31T07:01:24.5756507Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_as_cpu_complex64 PASSED [ 27%] 2023-03-31T07:01:24.5756861Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expand_as_cpu_float32 PASSED [ 27%] 2023-03-31T07:01:24.5757215Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_expm1_cpu_float32 PASSED [ 27%] 2023-03-31T07:01:24.5757562Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eye_cpu_complex64 XFAIL [ 28%] 2023-03-31T07:01:24.5757895Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_eye_cpu_float32 XFAIL [ 28%] 2023-03-31T07:01:24.5758249Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft2_cpu_complex64 PASSED [ 28%] 2023-03-31T07:01:24.5758615Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fft2_cpu_float32 PASSED [ 28%] 2023-03-31T07:01:24.5758985Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_fftshift_cpu_float32 PASSED [ 28%] 2023-03-31T07:01:24.5759340Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft2_cpu_complex64 PASSED [ 28%] 2023-03-31T07:01:24.5759702Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft2_cpu_float32 PASSED [ 29%] 2023-03-31T07:01:24.5760107Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft_cpu_complex64 PASSED [ 29%] 2023-03-31T07:01:24.5760468Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_hfft_cpu_float32 PASSED [ 29%] 2023-03-31T07:01:24.5760817Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft2_cpu_complex64 PASSED [ 29%] 2023-03-31T07:01:24.5761182Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifft2_cpu_float32 PASSED [ 29%] 2023-03-31T07:01:24.5761547Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftn_cpu_complex64 PASSED [ 29%] 2023-03-31T07:01:24.5761927Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ifftn_cpu_float32 PASSED [ 30%] 2023-03-31T07:01:24.5762320Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfft2_cpu_float32 PASSED [ 30%] 2023-03-31T07:01:24.5762689Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_ihfftn_cpu_float32 PASSED [ 30%] 2023-03-31T07:01:24.5763281Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft_cpu_complex64 PASSED [ 30%] 2023-03-31T07:01:24.5763665Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfft_cpu_float32 PASSED [ 30%] 2023-03-31T07:01:24.5764025Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_irfftn_cpu_float32 PASSED [ 30%] 2023-03-31T07:01:24.5764385Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfft2_cpu_float32 PASSED [ 31%] 2023-03-31T07:01:24.5764746Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfft_cpu_float32 PASSED [ 31%] 2023-03-31T07:01:24.5765089Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fft_rfftn_cpu_float32 PASSED [ 31%] 2023-03-31T07:01:24.5765452Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flatten_cpu_complex64 PASSED [ 31%] 2023-03-31T07:01:24.5765813Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flip_cpu_complex64 PASSED [ 31%] 2023-03-31T07:01:24.5766157Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flip_cpu_float32 PASSED [ 32%] 2023-03-31T07:01:24.5766516Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_flipud_cpu_complex64 PASSED [ 32%] 2023-03-31T07:01:24.5766875Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_floor_cpu_float32 PASSED [ 32%] 2023-03-31T07:01:24.5767258Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_floor_divide_cpu_float32 SKIPPED (Skipped!) [ 32%] 2023-03-31T07:01:24.5767621Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_fmax_cpu_float32 PASSED [ 32%] 2023-03-31T07:01:24.5767993Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_cpu_complex64 SKIPPED (Skipped!) [ 32%] 2023-03-31T07:01:24.5768372Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_full_like_cpu_complex64 PASSED [ 33%] 2023-03-31T07:01:24.5768722Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gather_cpu_complex64 PASSED [ 33%] 2023-03-31T07:01:24.5769080Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gather_cpu_float32 PASSED [ 33%] 2023-03-31T07:01:24.5769433Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ge_cpu_float32 PASSED [ 33%] 2023-03-31T07:01:24.5769787Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_geometric_cpu_float32 XFAIL [ 33%] 2023-03-31T07:01:24.5770140Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_grid_sampler_2d_cpu_float32 PASSED [ 33%] 2023-03-31T07:01:24.5770496Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_gt_cpu_float32 PASSED [ 34%] 2023-03-31T07:01:24.5770849Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_half_cpu_complex64 XFAIL [ 34%] 2023-03-31T07:01:24.5771202Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_half_cpu_float32 XFAIL [ 34%] 2023-03-31T07:01:24.5771540Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histc_cpu_float32 PASSED [ 34%] 2023-03-31T07:01:24.5771905Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_histogramdd_cpu_float32 XFAIL [ 34%] 2023-03-31T07:01:24.5772268Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hstack_cpu_complex64 PASSED [ 35%] 2023-03-31T07:01:24.5772699Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hstack_cpu_float32 PASSED [ 35%] 2023-03-31T07:01:24.5773055Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_hypot_cpu_float32 PASSED [ 35%] 2023-03-31T07:01:24.5773402Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_i0_cpu_float32 PASSED [ 35%] 2023-03-31T07:01:24.5773753Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_igamma_cpu_float32 PASSED [ 35%] 2023-03-31T07:01:24.5774137Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_imag_cpu_complex64 PASSED [ 35%] 2023-03-31T07:01:24.5774544Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_add_cpu_complex64 PASSED [ 36%] 2023-03-31T07:01:24.5774909Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_add_cpu_float32 PASSED [ 36%] 2023-03-31T07:01:24.5775319Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_copy_cpu_complex64 PASSED [ 36%] 2023-03-31T07:01:24.5775694Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_reduce_cpu_float32 PASSED [ 36%] 2023-03-31T07:01:24.5776068Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_index_select_cpu_complex64 PASSED [ 36%] 2023-03-31T07:01:24.5776431Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_inner_cpu_float32 PASSED [ 36%] 2023-03-31T07:01:24.5776767Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_int_cpu_float32 XFAIL [ 37%] 2023-03-31T07:01:24.5777121Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isclose_cpu_complex64 PASSED [ 37%] 2023-03-31T07:01:24.5777478Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isin_cpu_float32 PASSED [ 37%] 2023-03-31T07:01:24.5777828Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isnan_cpu_float32 PASSED [ 37%] 2023-03-31T07:01:24.5778168Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isneginf_cpu_float32 PASSED [ 37%] 2023-03-31T07:01:24.5778534Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isposinf_cpu_float32 PASSED [ 38%] 2023-03-31T07:01:24.5778892Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isreal_cpu_complex64 PASSED [ 38%] 2023-03-31T07:01:24.5779232Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_isreal_cpu_float32 PASSED [ 38%] 2023-03-31T07:01:24.5779650Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_istft_cpu_complex64 SKIPPED (Skipped! istft does not match the native function) [ 38%] 2023-03-31T07:01:24.5780110Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_2inputs_2outputs_cpu_float32 SKIPPED (Only runs on cuda) [ 38%] 2023-03-31T07:01:24.5780561Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_4inputs_with_extra_args_cpu_complex64 SKIPPED (Only runs on cuda) [ 38%] 2023-03-31T07:01:24.5780990Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_cpu_complex64 SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:01:24.5781431Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_return_by_ref_cpu_complex64 SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:01:24.5781873Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_binary_return_by_ref_cpu_float32 SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:01:24.5782300Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_unary_cpu_complex64 SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:01:24.5782791Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_jiterator_unary_cpu_float32 SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:01:24.5783178Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_kron_cpu_float32 PASSED [ 39%] 2023-03-31T07:01:24.5783538Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ldexp_cpu_complex64 PASSED [ 40%] 2023-03-31T07:01:24.5783897Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ldexp_cpu_float32 PASSED [ 40%] 2023-03-31T07:01:24.5784234Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_le_cpu_float32 PASSED [ 40%] 2023-03-31T07:01:24.5784647Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cross_cpu_complex64 PASSED [ 40%] 2023-03-31T07:01:24.5785025Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cross_cpu_float32 PASSED [ 40%] 2023-03-31T07:01:24.5785378Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_cpu_complex64 PASSED [ 40%] 2023-03-31T07:01:24.5785746Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_cpu_float32 PASSED [ 41%] 2023-03-31T07:01:24.5786192Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_det_singular_cpu_float32 PASSED [ 41%] 2023-03-31T07:01:24.5786598Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eig_cpu_float32 PASSED [ 41%] 2023-03-31T07:01:24.5786955Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigh_cpu_complex64 PASSED [ 41%] 2023-03-31T07:01:24.5787358Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigh_cpu_float32 PASSED [ 41%] 2023-03-31T07:01:24.5787738Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvals_cpu_complex64 PASSED [ 42%] 2023-03-31T07:01:24.5788116Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvals_cpu_float32 PASSED [ 42%] 2023-03-31T07:01:24.5788479Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_eigvalsh_cpu_float32 PASSED [ 42%] 2023-03-31T07:01:24.5788847Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_cpu_float32 PASSED [ 42%] 2023-03-31T07:01:24.5789218Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_ex_cpu_complex64 PASSED [ 42%] 2023-03-31T07:01:24.5789577Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_inv_ex_cpu_float32 PASSED [ 42%] 2023-03-31T07:01:24.5789956Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_cpu_complex64 PASSED [ 43%] 2023-03-31T07:01:24.5790351Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_factor_ex_cpu_complex64 PASSED [ 43%] 2023-03-31T07:01:24.5790739Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_ldl_solve_cpu_complex64 PASSED [ 43%] 2023-03-31T07:01:24.5791104Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_cpu_complex64 PASSED [ 43%] 2023-03-31T07:01:24.5791472Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_cpu_float32 PASSED [ 43%] 2023-03-31T07:01:24.5791863Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lstsq_grad_oriented_cpu_complex64 XFAIL [ 43%] 2023-03-31T07:01:24.5792252Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_cpu_complex64 PASSED [ 44%] 2023-03-31T07:01:24.5792607Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_cpu_float32 PASSED [ 44%] 2023-03-31T07:01:24.5792977Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_cpu_float32 PASSED [ 44%] 2023-03-31T07:01:24.5793362Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_factor_ex_cpu_complex64 PASSED [ 44%] 2023-03-31T07:01:24.5793745Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_lu_solve_cpu_complex64 PASSED [ 44%] 2023-03-31T07:01:24.5794117Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_power_cpu_complex64 PASSED [ 45%] 2023-03-31T07:01:24.5794506Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_power_cpu_float32 PASSED [ 45%] 2023-03-31T07:01:24.5794911Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_matrix_rank_cpu_complex64 SKIPPED (Skipped!) [ 45%] 2023-03-31T07:01:24.5795298Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_multi_dot_cpu_float32 PASSED [ 45%] 2023-03-31T07:01:24.5795698Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_norm_subgradients_at_zero_cpu_float32 PASSED [ 45%] 2023-03-31T07:01:24.5796089Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_cpu_complex64 PASSED [ 45%] 2023-03-31T07:01:24.5796597Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_pinv_singular_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 46%] 2023-03-31T07:01:24.5797026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_qr_cpu_complex64 PASSED [ 46%] 2023-03-31T07:01:24.5797389Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_qr_cpu_float32 PASSED [ 46%] 2023-03-31T07:01:24.5797751Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_cpu_float32 PASSED [ 46%] 2023-03-31T07:01:24.5798156Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_solve_ex_cpu_complex64 PASSED [ 46%] 2023-03-31T07:01:24.5798549Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svd_cpu_complex64 PASSED [ 46%] 2023-03-31T07:01:24.5798916Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svd_cpu_float32 PASSED [ 47%] 2023-03-31T07:01:24.5799313Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_svdvals_cpu_float32 PASSED [ 47%] 2023-03-31T07:01:24.5799679Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorinv_cpu_float32 PASSED [ 47%] 2023-03-31T07:01:24.5800058Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_tensorsolve_cpu_float32 PASSED [ 47%] 2023-03-31T07:01:24.5800439Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vander_cpu_complex64 PASSED [ 47%] 2023-03-31T07:01:24.5800816Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vander_cpu_float32 PASSED [ 47%] 2023-03-31T07:01:24.5801177Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vecdot_cpu_complex64 PASSED [ 48%] 2023-03-31T07:01:24.5801562Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_vector_norm_cpu_complex64 PASSED [ 48%] 2023-03-31T07:01:24.5801938Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linspace_cpu_complex64 PASSED [ 48%] 2023-03-31T07:01:24.5802306Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log1p_cpu_complex64 PASSED [ 48%] 2023-03-31T07:01:24.5802656Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log2_cpu_complex64 PASSED [ 48%] 2023-03-31T07:01:24.5803228Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_cpu_float32 PASSED [ 49%] 2023-03-31T07:01:24.5803644Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_log_softmax_with_dtype_cpu_complex64 PASSED [ 49%] 2023-03-31T07:01:24.5804013Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp2_cpu_float32 PASSED [ 49%] 2023-03-31T07:01:24.5804389Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp_cpu_complex64 PASSED [ 49%] 2023-03-31T07:01:24.5804762Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logaddexp_cpu_float32 PASSED [ 49%] 2023-03-31T07:01:24.5805136Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logcumsumexp_cpu_complex64 PASSED [ 49%] 2023-03-31T07:01:24.5805505Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logcumsumexp_cpu_float32 PASSED [ 50%] 2023-03-31T07:01:24.5805871Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logdet_cpu_float32 PASSED [ 50%] 2023-03-31T07:01:24.5806237Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_and_cpu_complex64 PASSED [ 50%] 2023-03-31T07:01:24.5806606Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_and_cpu_float32 PASSED [ 50%] 2023-03-31T07:01:24.5806957Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_not_cpu_float32 PASSED [ 50%] 2023-03-31T07:01:24.5807322Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_or_cpu_complex64 PASSED [ 50%] 2023-03-31T07:01:24.5807693Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_or_cpu_float32 PASSED [ 51%] 2023-03-31T07:01:24.5808050Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_xor_cpu_complex64 PASSED [ 51%] 2023-03-31T07:01:24.5808417Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logical_xor_cpu_float32 PASSED [ 51%] 2023-03-31T07:01:24.5808858Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logit_cpu_float32 PASSED [ 51%] 2023-03-31T07:01:24.5809222Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_logspace_cpu_complex64 PASSED [ 51%] 2023-03-31T07:01:24.5809568Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_long_cpu_complex64 XFAIL [ 52%] 2023-03-31T07:01:24.5809919Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lt_cpu_float32 PASSED [ 52%] 2023-03-31T07:01:24.5810261Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_cpu_float32 XFAIL [ 52%] 2023-03-31T07:01:24.5810661Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_solve_cpu_float32 PASSED [ 52%] 2023-03-31T07:01:24.5811047Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_lu_unpack_cpu_float32 PASSED [ 52%] 2023-03-31T07:01:24.5811400Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mH_cpu_complex64 XFAIL [ 52%] 2023-03-31T07:01:24.5811784Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mH_cpu_float32 XFAIL [ 53%] 2023-03-31T07:01:24.5812112Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mT_cpu_float32 XFAIL [ 53%] 2023-03-31T07:01:24.5812483Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_amax_cpu_float32 SKIPPED (Skipped!) [ 53%] 2023-03-31T07:01:24.5812888Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_cumprod_cpu_float32 SKIPPED (Skipped!) [ 53%] 2023-03-31T07:01:24.5813274Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_fill_cpu_complex64 PASSED [ 53%] 2023-03-31T07:01:24.5813657Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_logaddexp_cpu_float32 SKIPPED (Skipped!) [ 53%] 2023-03-31T07:01:24.5814043Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_mean_cpu_complex64 XFAIL [ 54%] 2023-03-31T07:01:24.5814434Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_median_cpu_float32 SKIPPED (Skipped!) [ 54%] 2023-03-31T07:01:24.5814816Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_prod_cpu_complex64 XFAIL [ 54%] 2023-03-31T07:01:24.5815172Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_prod_cpu_float32 XFAIL [ 54%] 2023-03-31T07:01:24.5815542Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_select_cpu_complex64 PASSED [ 54%] 2023-03-31T07:01:24.5815915Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_softmin_cpu_float32 XFAIL [ 54%] 2023-03-31T07:01:24.5816271Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_std_cpu_complex64 XFAIL [ 55%] 2023-03-31T07:01:24.5816639Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_sum_cpu_complex64 XFAIL [ 55%] 2023-03-31T07:01:24.5817003Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_masked_var_cpu_float32 XFAIL [ 55%] 2023-03-31T07:01:24.5817362Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matmul_cpu_complex64 PASSED [ 55%] 2023-03-31T07:01:24.5817710Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matmul_cpu_float32 PASSED [ 55%] 2023-03-31T07:01:24.5818074Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matrix_exp_cpu_complex64 PASSED [ 56%] 2023-03-31T07:01:24.5818441Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_matrix_exp_cpu_float32 PASSED [ 56%] 2023-03-31T07:01:24.5818803Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_max_reduction_no_dim_cpu_float32 PASSED [ 56%] 2023-03-31T07:01:24.5819171Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_maximum_cpu_float32 PASSED [ 56%] 2023-03-31T07:01:24.5819528Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mean_cpu_complex64 PASSED [ 56%] 2023-03-31T07:01:24.5819883Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mean_cpu_float32 PASSED [ 56%] 2023-03-31T07:01:24.5820223Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_median_cpu_float32 PASSED [ 57%] 2023-03-31T07:01:24.5820623Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_meshgrid_variadic_tensors_cpu_complex64 SKIPPED (Skipped!) [ 57%] 2023-03-31T07:01:24.5821054Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_binary_cpu_float32 PASSED [ 57%] 2023-03-31T07:01:24.5821435Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_min_reduction_with_dim_cpu_float32 PASSED [ 57%] 2023-03-31T07:01:24.5821793Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mm_cpu_complex64 PASSED [ 57%] 2023-03-31T07:01:24.5822141Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mm_cpu_float32 PASSED [ 57%] 2023-03-31T07:01:24.5822530Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_movedim_cpu_complex64 PASSED [ 58%] 2023-03-31T07:01:24.5823012Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_multinomial_cpu_float32 XFAIL [ 58%] 2023-03-31T07:01:24.5823370Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mv_cpu_complex64 PASSED [ 58%] 2023-03-31T07:01:24.5823794Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_1_cpu_float32 PASSED [ 58%] 2023-03-31T07:01:24.5824183Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_3_cpu_float32 PASSED [ 58%] 2023-03-31T07:01:24.5824559Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_mvlgamma_mvlgamma_p_5_cpu_float32 PASSED [ 59%] 2023-03-31T07:01:24.5824940Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanmean_cpu_float32 SKIPPED (Skipped!) [ 59%] 2023-03-31T07:01:24.5825312Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanmedian_cpu_float32 PASSED [ 59%] 2023-03-31T07:01:24.5825683Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nanquantile_cpu_float32 PASSED [ 59%] 2023-03-31T07:01:24.5826035Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_narrow_cpu_complex64 PASSED [ 59%] 2023-03-31T07:01:24.5826409Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_batch_norm_cpu_float32 XFAIL [ 59%] 2023-03-31T07:01:24.5826789Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_native_layer_norm_cpu_float32 XFAIL [ 60%] 2023-03-31T07:01:24.5827142Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ne_cpu_complex64 PASSED [ 60%] 2023-03-31T07:01:24.5827502Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_neg_cpu_complex64 PASSED [ 60%] 2023-03-31T07:01:24.5827872Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_empty_cpu_float32 SKIPPED (Skipped!) [ 60%] 2023-03-31T07:01:24.5828250Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_full_cpu_complex64 PASSED [ 60%] 2023-03-31T07:01:24.5828604Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_ones_cpu_complex64 PASSED [ 60%] 2023-03-31T07:01:24.5828967Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_ones_cpu_float32 PASSED [ 61%] 2023-03-31T07:01:24.5829327Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_new_zeros_cpu_float32 PASSED [ 61%] 2023-03-31T07:01:24.5829688Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nextafter_cpu_float32 PASSED [ 61%] 2023-03-31T07:01:24.5830070Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_avg_pool1d_cpu_float32 PASSED [ 61%] 2023-03-31T07:01:24.5830485Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_adaptive_max_pool1d_cpu_float32 PASSED [ 61%] 2023-03-31T07:01:24.5830887Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_alpha_dropout_cpu_float32 XFAIL [ 61%] 2023-03-31T07:01:24.5831267Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_bilinear_cpu_float32 PASSED [ 62%] 2023-03-31T07:01:24.5831668Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_binary_cross_entropy_cpu_float32 XFAIL [ 62%] 2023-03-31T07:01:24.5832102Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_binary_cross_entropy_with_logits_cpu_float32 SKIPPED (Skipped!) [ 62%] 2023-03-31T07:01:24.5832524Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_celu_cpu_float32 PASSED [ 62%] 2023-03-31T07:01:24.5832973Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv1d_cpu_complex64 SKIPPED (Skipped!) [ 62%] 2023-03-31T07:01:24.5833386Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv1d_cpu_float32 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:01:24.5833803Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv2d_cpu_float32 SKIPPED (Works on some configs!) [ 63%] 2023-03-31T07:01:24.5834241Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose1d_cpu_complex64 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:01:24.5834683Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose1d_cpu_float32 XFAIL [ 63%] 2023-03-31T07:01:24.5835163Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose2d_cpu_complex64 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:01:24.5835598Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose2d_cpu_float32 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:01:24.5836070Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_conv_transpose3d_cpu_complex64 SKIPPED (Skipped!) [ 64%] 2023-03-31T07:01:24.5836489Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_cosine_embedding_loss_cpu_float32 PASSED [ 64%] 2023-03-31T07:01:24.5836903Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_ctc_loss_cpu_float32 SKIPPED (Skipped!) [ 64%] 2023-03-31T07:01:24.5837306Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_dropout3d_cpu_float32 XFAIL [ 64%] 2023-03-31T07:01:24.5837702Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_embedding_bag_cpu_float32 XFAIL [ 64%] 2023-03-31T07:01:24.5838122Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_without_train_cpu_complex64 XFAIL [ 64%] 2023-03-31T07:01:24.5838569Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_without_train_cpu_float32 XFAIL [ 65%] 2023-03-31T07:01:24.5838995Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_fractional_max_pool2d_cpu_float32 XFAIL [ 65%] 2023-03-31T07:01:24.5839392Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_gelu_cpu_float32 PASSED [ 65%] 2023-03-31T07:01:24.5839761Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_glu_cpu_float32 PASSED [ 65%] 2023-03-31T07:01:24.5840150Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardshrink_cpu_float32 PASSED [ 65%] 2023-03-31T07:01:24.5840552Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hardsigmoid_cpu_float32 PASSED [ 66%] 2023-03-31T07:01:24.5840966Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_hinge_embedding_loss_cpu_float32 PASSED [ 66%] 2023-03-31T07:01:24.5841366Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_bilinear_cpu_float32 XFAIL [ 66%] 2023-03-31T07:01:24.5841784Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_interpolate_linear_cpu_float32 XFAIL [ 66%] 2023-03-31T07:01:24.5842183Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_kl_div_cpu_float32 PASSED [ 66%] 2023-03-31T07:01:24.5842558Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_l1_loss_cpu_complex64 PASSED [ 66%] 2023-03-31T07:01:24.5842950Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_linear_cpu_complex64 PASSED [ 67%] 2023-03-31T07:01:24.5843573Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_linear_cpu_float32 PASSED [ 67%] 2023-03-31T07:01:24.5843975Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_local_response_norm_cpu_float32 XFAIL [ 67%] 2023-03-31T07:01:24.5844371Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_margin_ranking_loss_cpu_float32 PASSED [ 67%] 2023-03-31T07:01:24.5844913Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool2d_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 67%] 2023-03-31T07:01:24.5845378Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_pool3d_cpu_float32 PASSED [ 67%] 2023-03-31T07:01:24.5845779Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool2d_cpu_float32 PASSED [ 68%] 2023-03-31T07:01:24.5846170Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool2d_grad_cpu_float32 PASSED [ 68%] 2023-03-31T07:01:24.5846615Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool3d_cpu_float32 PASSED [ 68%] 2023-03-31T07:01:24.5847066Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_max_unpool3d_grad_cpu_float32 PASSED [ 68%] 2023-03-31T07:01:24.5847462Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_mish_cpu_float32 PASSED [ 68%] 2023-03-31T07:01:24.5847876Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_mse_loss_cpu_float32 XFAIL [ 69%] 2023-03-31T07:01:24.5848284Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_multilabel_soft_margin_loss_cpu_float32 PASSED [ 69%] 2023-03-31T07:01:24.5848707Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_nll_loss_cpu_float32 SKIPPED (Skipped!) [ 69%] 2023-03-31T07:01:24.5849110Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_normalize_cpu_float32 PASSED [ 69%] 2023-03-31T07:01:24.5849508Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_circular_cpu_float32 SKIPPED (Skipped!) [ 69%] 2023-03-31T07:01:24.5849929Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_constant_cpu_complex64 PASSED [ 69%] 2023-03-31T07:01:24.5850337Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_constant_cpu_float32 PASSED [ 70%] 2023-03-31T07:01:24.5850754Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_reflect_cpu_float32 SKIPPED (Skipped!) [ 70%] 2023-03-31T07:01:24.5851157Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pad_replicate_cpu_complex64 PASSED [ 70%] 2023-03-31T07:01:24.5851587Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pairwise_distance_cpu_complex64 SKIPPED (Skipped!) [ 70%] 2023-03-31T07:01:24.5852028Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_shuffle_cpu_complex64 SKIPPED (Skipped!) [ 70%] 2023-03-31T07:01:24.5852463Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_pixel_unshuffle_cpu_complex64 SKIPPED (Skipped!) [ 70%] 2023-03-31T07:01:24.5852857Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_rrelu_cpu_float32 XFAIL [ 71%] 2023-03-31T07:01:24.5853281Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_scaled_dot_product_attention_cpu_float32 SKIPPED (Skipped!) [ 71%] 2023-03-31T07:01:24.5853720Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_silu_complex_cpu_complex64 PASSED [ 71%] 2023-03-31T07:01:24.5854109Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_smooth_l1_loss_cpu_float32 XFAIL [ 71%] 2023-03-31T07:01:24.5854518Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cpu_complex64 PASSED [ 71%] 2023-03-31T07:01:24.5854924Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softplus_cpu_float32 PASSED [ 71%] 2023-03-31T07:01:24.5855318Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softshrink_cpu_float32 PASSED [ 72%] 2023-03-31T07:01:24.5855700Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_softsign_cpu_float32 PASSED [ 72%] 2023-03-31T07:01:24.5856093Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_tanhshrink_cpu_complex64 PASSED [ 72%] 2023-03-31T07:01:24.5856540Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_tanhshrink_cpu_float32 PASSED [ 72%] 2023-03-31T07:01:24.5856964Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cpu_complex64 XFAIL [ 72%] 2023-03-31T07:01:24.5857395Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_unfold_cpu_complex64 SKIPPED (Internal assert failed!) [ 73%] 2023-03-31T07:01:24.5857837Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_unfold_cpu_float32 SKIPPED (Internal assert failed!) [ 73%] 2023-03-31T07:01:24.5858289Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_upsample_bilinear_cpu_float32 XFAIL [ 73%] 2023-03-31T07:01:24.5858732Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_nn_functional_upsample_nearest_cpu_float32 XFAIL [ 73%] 2023-03-31T07:01:24.5859100Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_fro_cpu_float32 XFAIL [ 73%] 2023-03-31T07:01:24.5859496Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_inf_cpu_complex64 PASSED [ 73%] 2023-03-31T07:01:24.5859859Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_norm_inf_cpu_float32 XFAIL [ 74%] 2023-03-31T07:01:24.5860209Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_cpu_float32 XFAIL [ 74%] 2023-03-31T07:01:24.5860561Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_normal_in_place_cpu_complex64 XFAIL [ 74%] 2023-03-31T07:01:24.5860943Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ones_cpu_complex64 SKIPPED (Skipped!) [ 74%] 2023-03-31T07:01:24.5861309Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ormqr_cpu_float32 PASSED [ 74%] 2023-03-31T07:01:24.5861649Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_outer_cpu_float32 PASSED [ 74%] 2023-03-31T07:01:24.5862010Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pca_lowrank_cpu_float32 XFAIL [ 75%] 2023-03-31T07:01:24.5862379Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pinverse_cpu_complex64 PASSED [ 75%] 2023-03-31T07:01:24.5862845Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pinverse_cpu_float32 PASSED [ 75%] 2023-03-31T07:01:24.5863192Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polar_cpu_float32 PASSED [ 75%] 2023-03-31T07:01:24.5863571Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_0_cpu_float32 PASSED [ 75%] 2023-03-31T07:01:24.5863981Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_polygamma_polygamma_n_1_cpu_float32 SKIPPED (Skipped!) [ 76%] 2023-03-31T07:01:24.5864366Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_pow_cpu_float32 PASSED [ 76%] 2023-03-31T07:01:24.5864712Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_prod_cpu_float32 PASSED [ 76%] 2023-03-31T07:01:24.5865068Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_put_cpu_complex64 PASSED [ 76%] 2023-03-31T07:01:24.5865419Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_put_cpu_float32 PASSED [ 76%] 2023-03-31T07:01:24.5865762Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_quantile_cpu_float32 PASSED [ 76%] 2023-03-31T07:01:24.5866119Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randint_cpu_float32 XFAIL [ 77%] 2023-03-31T07:01:24.5866480Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randint_like_cpu_float32 XFAIL [ 77%] 2023-03-31T07:01:24.5866835Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_cpu_complex64 XFAIL [ 77%] 2023-03-31T07:01:24.5867173Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_cpu_float32 XFAIL [ 77%] 2023-03-31T07:01:24.5867529Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_randn_like_cpu_float32 XFAIL [ 77%] 2023-03-31T07:01:24.5867884Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_ravel_cpu_float32 PASSED [ 77%] 2023-03-31T07:01:24.5868219Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_real_cpu_float32 PASSED [ 78%] 2023-03-31T07:01:24.5868584Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reciprocal_cpu_float32 PASSED [ 78%] 2023-03-31T07:01:24.5868989Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_renorm_cpu_complex64 PASSED [ 78%] 2023-03-31T07:01:24.5869358Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_as_cpu_complex64 PASSED [ 78%] 2023-03-31T07:01:24.5869716Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_as_cpu_float32 PASSED [ 78%] 2023-03-31T07:01:24.5870081Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_reshape_cpu_complex64 PASSED [ 78%] 2023-03-31T07:01:24.5870488Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize__cpu_complex64 PASSED [ 79%] 2023-03-31T07:01:24.5870878Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resize__cpu_float32 PASSED [ 79%] 2023-03-31T07:01:24.5871228Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_resolve_conj_cpu_complex64 PASSED [ 79%] 2023-03-31T07:01:24.5871622Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_roll_cpu_float32 PASSED [ 79%] 2023-03-31T07:01:24.5871974Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rot90_cpu_complex64 PASSED [ 79%] 2023-03-31T07:01:24.5872310Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_cpu_float32 PASSED [ 80%] 2023-03-31T07:01:24.5872690Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_3_cpu_float32 SKIPPED (Skipped!) [ 80%] 2023-03-31T07:01:24.5873098Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_round_decimals_neg_3_cpu_float32 SKIPPED (Skipped!) [ 80%] 2023-03-31T07:01:24.5873483Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsqrt_cpu_complex64 PASSED [ 80%] 2023-03-31T07:01:24.5873829Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_rsub_cpu_complex64 PASSED [ 80%] 2023-03-31T07:01:24.5874199Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scalar_tensor_cpu_complex64 XFAIL [ 80%] 2023-03-31T07:01:24.5874578Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_add_cpu_complex64 PASSED [ 81%] 2023-03-31T07:01:24.5874952Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_add_cpu_float32 PASSED [ 81%] 2023-03-31T07:01:24.5875316Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_amax_cpu_float32 PASSED [ 81%] 2023-03-31T07:01:24.5875699Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_amin_cpu_float32 PASSED [ 81%] 2023-03-31T07:01:24.5876080Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_mean_cpu_float32 PASSED [ 81%] 2023-03-31T07:01:24.5876453Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_scatter_reduce_prod_cpu_float32 PASSED [ 81%] 2023-03-31T07:01:24.5876822Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_cpu_float32 PASSED [ 82%] 2023-03-31T07:01:24.5877188Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_select_scatter_cpu_float32 PASSED [ 82%] 2023-03-31T07:01:24.5877553Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_short_cpu_float32 XFAIL [ 82%] 2023-03-31T07:01:24.5877901Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sigmoid_cpu_complex64 PASSED [ 82%] 2023-03-31T07:01:24.5878264Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sigmoid_cpu_float32 PASSED [ 82%] 2023-03-31T07:01:24.5878635Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_cosine_cpu_float32 XFAIL [ 83%] 2023-03-31T07:01:24.5879028Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_exponential_cpu_float32 XFAIL [ 83%] 2023-03-31T07:01:24.5879412Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_gaussian_cpu_float32 XFAIL [ 83%] 2023-03-31T07:01:24.5879811Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_general_cosine_cpu_float32 XFAIL [ 83%] 2023-03-31T07:01:24.5880206Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signal_windows_nuttall_cpu_float32 XFAIL [ 83%] 2023-03-31T07:01:24.5880566Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_signbit_cpu_float32 PASSED [ 83%] 2023-03-31T07:01:24.5881029Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sin_cpu_complex64 PASSED [ 84%] 2023-03-31T07:01:24.5881384Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sin_cpu_float32 PASSED [ 84%] 2023-03-31T07:01:24.5881738Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinc_cpu_complex64 PASSED [ 84%] 2023-03-31T07:01:24.5882078Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinc_cpu_float32 PASSED [ 84%] 2023-03-31T07:01:24.5882466Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sinh_cpu_complex64 PASSED [ 84%] 2023-03-31T07:01:24.5883401Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_cpu_float32 SKIPPED (variant consistency doesn't work on torch.ops) [ 84%] 2023-03-31T07:01:24.5883823Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_slice_scatter_cpu_float32 PASSED [ 85%] 2023-03-31T07:01:24.5884241Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_cpu_float32 PASSED [ 85%] 2023-03-31T07:01:24.5884620Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_with_dtype_cpu_complex64 PASSED [ 85%] 2023-03-31T07:01:24.5885006Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_softmax_with_dtype_cpu_float32 PASSED [ 85%] 2023-03-31T07:01:24.5885357Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sort_cpu_float32 PASSED [ 85%] 2023-03-31T07:01:24.5885739Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sparse_mm_reduce_cpu_float32 SKIPPED (Skipped!) [ 85%] 2023-03-31T07:01:24.5886133Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_airy_ai_cpu_float32 PASSED [ 86%] 2023-03-31T07:01:24.5886515Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_bessel_y0_cpu_float32 PASSED [ 86%] 2023-03-31T07:01:24.5886897Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_t_cpu_float32 PASSED [ 86%] 2023-03-31T07:01:24.5887313Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_u_cpu_float32 PASSED [ 86%] 2023-03-31T07:01:24.5887945Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 86%] 2023-03-31T07:01:24.5888613Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 87%] 2023-03-31T07:01:24.5889046Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_entr_cpu_float32 PASSED [ 87%] 2023-03-31T07:01:24.5889422Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_erfcx_cpu_float32 PASSED [ 87%] 2023-03-31T07:01:24.5889820Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_hermite_polynomial_he_cpu_float32 PASSED [ 87%] 2023-03-31T07:01:24.5890208Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_i1_cpu_float32 PASSED [ 87%] 2023-03-31T07:01:24.5890585Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_laguerre_polynomial_l_cpu_float32 PASSED [ 87%] 2023-03-31T07:01:24.5890988Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_i0_cpu_float32 PASSED [ 88%] 2023-03-31T07:01:24.5891389Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_modified_bessel_i1_cpu_float32 PASSED [ 88%] 2023-03-31T07:01:24.5891772Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_ndtr_cpu_float32 PASSED [ 88%] 2023-03-31T07:01:24.5892359Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_t_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 88%] 2023-03-31T07:01:24.5893045Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 88%] 2023-03-31T07:01:24.5893577Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_spherical_bessel_j0_cpu_float32 PASSED [ 88%] 2023-03-31T07:01:24.5893970Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_special_xlog1py_cpu_float32 PASSED [ 89%] 2023-03-31T07:01:24.5894319Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_cpu_float32 PASSED [ 89%] 2023-03-31T07:01:24.5894690Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_list_args_cpu_complex64 PASSED [ 89%] 2023-03-31T07:01:24.5895111Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_list_args_cpu_float32 PASSED [ 89%] 2023-03-31T07:01:24.5895519Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_split_with_sizes_cpu_complex64 PASSED [ 89%] 2023-03-31T07:01:24.5895890Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sqrt_cpu_complex64 PASSED [ 90%] 2023-03-31T07:01:24.5896246Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_square_cpu_float32 PASSED [ 90%] 2023-03-31T07:01:24.5896640Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_cpu_complex64 PASSED [ 90%] 2023-03-31T07:01:24.5896989Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_cpu_float32 PASSED [ 90%] 2023-03-31T07:01:24.5897362Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_squeeze_multiple_cpu_float32 PASSED [ 90%] 2023-03-31T07:01:24.5897730Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stack_cpu_float32 PASSED [ 90%] 2023-03-31T07:01:24.5898087Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_cpu_complex64 PASSED [ 91%] 2023-03-31T07:01:24.5898434Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_cpu_complex64 PASSED [ 91%] 2023-03-31T07:01:24.5898801Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_cpu_float32 PASSED [ 91%] 2023-03-31T07:01:24.5899174Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_std_mean_unbiased_cpu_complex64 PASSED [ 91%] 2023-03-31T07:01:24.5899600Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_stft_cpu_complex64 SKIPPED (Skipped! stft does not match the native function) [ 91%] 2023-03-31T07:01:24.5900003Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sub_cpu_float32 PASSED [ 91%] 2023-03-31T07:01:24.5900355Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_cpu_complex64 PASSED [ 92%] 2023-03-31T07:01:24.5900706Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_sum_cpu_float32 PASSED [ 92%] 2023-03-31T07:01:24.5901043Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_cpu_complex64 PASSED [ 92%] 2023-03-31T07:01:24.5901404Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_svd_lowrank_cpu_float32 XFAIL [ 92%] 2023-03-31T07:01:24.5901762Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_t_cpu_complex64 PASSED [ 92%] 2023-03-31T07:01:24.5902112Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tan_cpu_complex64 PASSED [ 92%] 2023-03-31T07:01:24.5902455Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tanh_cpu_complex64 PASSED [ 93%] 2023-03-31T07:01:24.5902922Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tanh_cpu_float32 PASSED [ 93%] 2023-03-31T07:01:24.5903286Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tensor_split_cpu_complex64 PASSED [ 93%] 2023-03-31T07:01:24.5903638Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tile_cpu_complex64 PASSED [ 93%] 2023-03-31T07:01:24.5903993Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tile_cpu_float32 PASSED [ 93%] 2023-03-31T07:01:24.5904370Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_cpu_complex64 SKIPPED (Skipped!) [ 94%] 2023-03-31T07:01:24.5904752Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_cpu_float32 SKIPPED (Skipped!) [ 94%] 2023-03-31T07:01:24.5905107Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_to_sparse_cpu_float32 PASSED [ 94%] 2023-03-31T07:01:24.5905472Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trace_cpu_complex64 PASSED [ 94%] 2023-03-31T07:01:24.5905882Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_transpose_cpu_complex64 PASSED [ 94%] 2023-03-31T07:01:24.5906242Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_transpose_cpu_float32 PASSED [ 94%] 2023-03-31T07:01:24.5906599Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trapz_cpu_float32 PASSED [ 95%] 2023-03-31T07:01:24.5906967Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triangular_solve_cpu_float32 PASSED [ 95%] 2023-03-31T07:01:24.5907365Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_tril_cpu_float32 PASSED [ 95%] 2023-03-31T07:01:24.5907748Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_triu_cpu_complex64 PASSED [ 95%] 2023-03-31T07:01:24.5908103Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_trunc_cpu_float32 PASSED [ 95%] 2023-03-31T07:01:24.5908462Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unbind_cpu_complex64 PASSED [ 95%] 2023-03-31T07:01:24.5908859Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unbind_cpu_float32 PASSED [ 96%] 2023-03-31T07:01:24.5909213Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_copy_cpu_complex64 PASSED [ 96%] 2023-03-31T07:01:24.5909586Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_copy_cpu_float32 PASSED [ 96%] 2023-03-31T07:01:24.5909952Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unfold_cpu_complex64 PASSED [ 96%] 2023-03-31T07:01:24.5910299Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_uniform_cpu_complex64 XFAIL [ 96%] 2023-03-31T07:01:24.5910657Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unique_cpu_float32 XFAIL [ 97%] 2023-03-31T07:01:24.5911026Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsafe_split_cpu_complex64 PASSED [ 97%] 2023-03-31T07:01:24.5911402Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsafe_split_cpu_float32 PASSED [ 97%] 2023-03-31T07:01:24.5911754Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_unsqueeze_cpu_float32 PASSED [ 97%] 2023-03-31T07:01:24.5912115Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_cpu_complex64 PASSED [ 97%] 2023-03-31T07:01:24.5912465Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_cpu_float32 PASSED [ 97%] 2023-03-31T07:01:24.5912810Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_cpu_complex64 PASSED [ 98%] 2023-03-31T07:01:24.5913171Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_cpu_float32 PASSED [ 98%] 2023-03-31T07:01:24.5913548Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_unbiased_cpu_complex64 PASSED [ 98%] 2023-03-31T07:01:24.5913934Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_mean_unbiased_cpu_float32 PASSED [ 98%] 2023-03-31T07:01:24.5914299Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_var_unbiased_cpu_complex64 PASSED [ 98%] 2023-03-31T07:01:24.5914675Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_as_real_cpu_complex64 PASSED [ 98%] 2023-03-31T07:01:24.5915046Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_view_copy_cpu_float32 PASSED [ 99%] 2023-03-31T07:01:24.5915405Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_vsplit_cpu_float32 PASSED [ 99%] 2023-03-31T07:01:24.5915762Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_where_cpu_complex64 SKIPPED (Skipped!) [ 99%] 2023-03-31T07:01:24.5916133Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zero__cpu_complex64 XFAIL [ 99%] 2023-03-31T07:01:24.5916486Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zero__cpu_float32 XFAIL [ 99%] 2023-03-31T07:01:24.5916831Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_zeros_like_cpu_float32 PASSED [100%] 2023-03-31T07:01:24.5917037Z 2023-03-31T07:01:24.5917415Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_jit/test_ops_jit-09b7eaa136a34551.xml - 2023-03-31T07:01:24.5917807Z ==== 414 passed, 61 skipped, 1 deselected, 96 xfailed in 586.02s (0:09:46) ===== 2023-03-31T07:01:24.5918192Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:01:24.5918399Z 2023-03-31T07:01:24.5918805Z ##[endgroup] 2023-03-31T07:01:24.5919210Z FINISHED PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_lqzk42va.log) 2023-03-31T07:01:24.5919439Z 2023-03-31T07:01:24.9824423Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_jit.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '-k=_linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:01:24.981941] 2023-03-31T07:01:32.7384250Z 2023-03-31T07:01:32.7384766Z Expand the folded group to see the log file of test_ops_jit 2023-03-31T07:01:32.7385989Z ##[group]PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_iuvt931v.log) 2023-03-31T07:01:32.7386648Z Test results will be stored in test-reports/python-pytest/test_ops_jit/test_ops_jit-a2be77a573a5370e.xml 2023-03-31T07:01:32.7386991Z ============================= test session starts ============================== 2023-03-31T07:01:32.7387397Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:01:32.7387679Z cachedir: .pytest_cache 2023-03-31T07:01:32.7388109Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:01:32.7388613Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:01:32.7389153Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:01:32.7389538Z collecting ... collected 1079 items / 1075 deselected / 4 selected 2023-03-31T07:01:32.7390283Z Running 4 items in this shard: test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_cpu_float32, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_ex_cpu_complex64, test/test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_ex_cpu_float32 2023-03-31T07:01:32.7390775Z 2023-03-31T07:01:32.7390967Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_cpu_complex64 PASSED [ 25%] 2023-03-31T07:01:32.7391354Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_cpu_float32 PASSED [ 50%] 2023-03-31T07:01:32.7391743Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_ex_cpu_complex64 PASSED [ 75%] 2023-03-31T07:01:32.7392120Z test_ops_jit.py::TestJitCPU::test_variant_consistency_jit_linalg_cholesky_ex_cpu_float32 PASSED [100%] 2023-03-31T07:01:32.7392329Z 2023-03-31T07:01:32.7392658Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_jit/test_ops_jit-a2be77a573a5370e.xml - 2023-03-31T07:01:32.7393027Z ====================== 4 passed, 1075 deselected in 4.53s ====================== 2023-03-31T07:01:32.7393354Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:01:32.7393555Z 2023-03-31T07:01:32.7393870Z ##[endgroup] 2023-03-31T07:01:32.7394319Z FINISHED PRINTING LOG FILE of test_ops_jit (/var/lib/jenkins/workspace/test/test-reports/test_ops_jit_iuvt931v.log) 2023-03-31T07:01:32.7394543Z 2023-03-31T07:01:32.7395148Z Running test_ops_fwd_gradients ... [2023-03-31 07:01:32.739320] 2023-03-31T07:01:34.4837179Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:01:34.4866886Z Ignoring disabled issues: [] 2023-03-31T07:01:34.5061854Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=0', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:01:34.505753] 2023-03-31T07:01:34.5161628Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:01:34.5190710Z Ignoring disabled issues: [] 2023-03-31T07:01:34.5386274Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--shard-id=1', '--num-shards=2', '-k=not _linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:01:34.538147] 2023-03-31T07:05:22.1149848Z 2023-03-31T07:05:22.1150464Z Expand the folded group to see the log file of test_ops_fwd_gradients 2023-03-31T07:05:22.1151582Z ##[group]PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_y2p_cyv0.log) 2023-03-31T07:05:22.1157263Z Test results will be stored in test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5d70d9a787657575.xml 2023-03-31T07:05:22.1157844Z ============================= test session starts ============================== 2023-03-31T07:05:22.1184335Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:05:22.1184974Z cachedir: .pytest_cache 2023-03-31T07:05:22.1186028Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:05:22.1186679Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:05:22.1187501Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:05:22.1188082Z collecting ... collected 3021 items / 5 deselected / 3016 selected 2023-03-31T07:05:22.1425142Z Running 1503 items in this shard: test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_H_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_T_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___getitem___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___radd___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___radd___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rdiv___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rdiv___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmatmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rsub___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__segment_reduce_offsets_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_abs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcdiv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_decomposed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_allclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_aminmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_arange_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argsort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argwhere_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argwhere_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_partial_views_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_baddbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_baddbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bernoulli_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bfloat16_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_block_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_block_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bool_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cartesian_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdouble_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cfloat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_char_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_inverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chunk_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_max_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_min_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_column_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_column_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_combinations_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_combinations_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_physical_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_physical_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_constant_pad_nd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_constant_pad_nd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_contiguous_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_copysign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_count_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_count_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cov_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cummax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumulative_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumulative_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_deg2rad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_embed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_embed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diff_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dist_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_no_rounding_mode_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_double_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_einsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_einsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_permuted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_equal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_equal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erfinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expm1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flip_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flip_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fliplr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flipud_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_floor_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ge_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geometric_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gradient_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gradient_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_half_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_half_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_heaviside_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histogram_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histogramdd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_igammac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_imag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_inner_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_int_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isfinite_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isinf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isnan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isposinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_istft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_2inputs_2outputs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_2inputs_2outputs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_4inputs_with_extra_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_return_by_ref_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_return_by_ref_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_unary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_unary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kron_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kron_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kthvalue_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ldexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ldexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_le_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lerp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lgamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eig_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvalsh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvalsh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_householder_product_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_grad_oriented_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_multi_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_multi_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_slogdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_triangular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vander_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vector_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log10_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log10_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log1p_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log1p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logcumsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_and_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_not_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_or_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_xor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_long_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_unpack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_unpack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mH_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matrix_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_pool2d_with_indices_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_variadic_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_movedim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanmean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanmedian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanquantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_dropout_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ne_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nextafter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_alpha_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_celu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_ctc_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_embedding_bag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_embedding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_without_train_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_fractional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_gaussian_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_glu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_grid_sample_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_group_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hinge_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_huber_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_instance_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_trilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_kl_div_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_l1_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_leaky_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_margin_ranking_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool2d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool3d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multi_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multilabel_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_circular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_constant_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_reflect_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_reflect_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_replicate_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pairwise_distance_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_shuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_poisson_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_relu6_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_rrelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_scaled_dot_product_attention_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_selu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_silu_complex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_with_distance_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_upsample_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_fro_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_nuc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_nuc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_in_place_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ormqr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ormqr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_outer_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pca_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_permute_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_permute_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pinverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_4_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_positive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_quantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rand_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rand_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reciprocal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_remainder_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_renorm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_interleave_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_neg_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scalar_tensor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_short_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_gaussian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_general_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_nuttall_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signbit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_y0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_entr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_erfcx_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i0e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i1e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_laguerre_polynomial_l_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_legendre_polynomial_p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_log_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_ndtri_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_scaled_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_xlog1py_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_zeta_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_list_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_to_size_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_t_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_along_dim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_sparse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_topk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapz_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triangular_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triangular_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tril_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_true_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unbind_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unflatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unflatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_uniform_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unique_consecutive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unique_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsafe_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsqueeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_where_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zero__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zero__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_H_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_T_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___getitem___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___getitem___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rdiv___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmatmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmatmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rpow___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__softmax_backward_data_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__upsample_bilinear2d_aa_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcdiv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_decomposed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_all_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_all_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_allclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_any_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_any_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argsort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argwhere_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argwhere_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_partial_views_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_partial_views_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_baddbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bernoulli_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bfloat16_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_block_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bool_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bool_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bucketize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_byte_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cartesian_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cauchy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdouble_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cfloat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cfloat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chalf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_char_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_inverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chunk_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_max_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_min_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clone_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_column_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_physical_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_physical_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_constant_pad_nd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_contiguous_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_copysign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_corrcoef_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_count_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_count_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cov_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumulative_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_embed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_embed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagflat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diff_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diff_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_digamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dist_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_no_rounding_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_trunc_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_double_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_einsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_permuted_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_permuted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_equal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erfc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erfinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expm1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eye_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eye_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flip_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flip_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fliplr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flipud_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_floor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_floor_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gather_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gather_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geometric_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geqrf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geqrf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gradient_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_grid_sampler_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_heaviside_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histogramdd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_igamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_igammac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_imag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_inner_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isfinite_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isinf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isneginf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isposinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isreal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_istft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_return_by_ref_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_unary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_unary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ldexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lerp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eig_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eig_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_householder_product_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_householder_product_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_multi_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_slogdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_slogdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_triangular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svdvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorsolve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vecdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vecdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log1p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logcumsumexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_not_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_xor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_long_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mH_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mT_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matrix_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matrix_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_variadic_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanmean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanmedian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nansum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_dropout_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ne_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ne_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_alpha_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_binary_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cosine_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_embedding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_with_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_fractional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_glu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_grid_sample_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_group_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardswish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_huber_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_bicubic_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_kl_div_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_local_response_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_margin_ranking_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool1d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool2d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_mse_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_circular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_constant_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_reflect_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_reflect_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pairwise_distance_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_shuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_prelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_relu6_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_rrelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_scaled_dot_product_attention_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_silu_complex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softsign_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softsign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_tanhshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_upsample_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_upsample_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_fro_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_inf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_inf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_nuc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ormqr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_outer_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_outer_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_permute_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pinverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polar_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_positive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_quantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rad2deg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rand_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randint_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ravel_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ravel_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_interleave_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_roll_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rot90_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rot90_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_searchsorted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sgn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_short_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sigmoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_gaussian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_general_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_general_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_nuttall_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signbit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_mm_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_sampled_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_sampled_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_y0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_y1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i0e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_legendre_polynomial_p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_scaled_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_xlog1py_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_list_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_with_sizes_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_with_sizes_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_multiple_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_multiple_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_to_size_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_along_dim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensor_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensordot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_sparse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_sparse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapz_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triangular_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tril_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_true_divide_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trunc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unflatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_uniform_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_uniform_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unique_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsafe_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_where_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_xlogy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zero__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zero__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_H_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_T_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___getitem___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmatmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmod___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rpow___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rsub___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rsub___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__native_batch_norm_legit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__segment_reduce_offsets_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_abs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcdiv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcdiv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_all_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_any_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_any_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argsort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_partial_views_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_baddbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bfloat16_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_block_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bool_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bucketize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_byte_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cartesian_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cauchy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdouble_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ceil_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chalf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chalf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_char_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_char_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_inverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_inverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clone_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_column_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_combinations_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_physical_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_constant_pad_nd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_contiguous_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_contiguous_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_corrcoef_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_corrcoef_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_count_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cov_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cummax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagflat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diff_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_no_rounding_mode_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_double_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_einsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erfc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expm1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eye_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eye_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftshift_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flipud_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_floor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_floor_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_frac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_frexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gather_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gradient_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gradient_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_grid_sampler_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_half_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_half_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_heaviside_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histogram_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hypot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_igammac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_imag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_inner_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isfinite_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isfinite_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isinf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isnan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isreal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isreal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_2inputs_2outputs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_return_by_ref_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_return_by_ref_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_unary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_unary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_le_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cond_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eig_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvalsh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvalsh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_multi_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_triangular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorsolve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vander_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vander_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vector_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log10_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log10_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log1p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_and_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_not_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_or_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_or_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_xor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_long_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_unpack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_unpack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mH_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mT_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matrix_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_pool2d_with_indices_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_list_of_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_list_of_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_variadic_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_msort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_multinomial_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_5_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanmean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nansum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_dropout_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ne_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nextafter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_binary_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_binary_cross_entropy_with_logits_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_celu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cosine_similarity_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_elu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_embedding_bag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_glu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardtanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_huber_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_instance_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_trilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_kl_div_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_l1_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_leaky_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_local_response_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool1d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_mish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_mse_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multi_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multilabel_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_circular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_replicate_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_replicate_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pairwise_distance_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pairwise_distance_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_shuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_silu_complex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_silu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_smooth_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softplus_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softsign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_upsample_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_fro_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_in_place_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ormqr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_outer_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pca_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pinverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pinverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polar_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_quantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rand_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randint_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randint_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ravel_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_real_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reciprocal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reciprocal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_remainder_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_interleave_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_interleave_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize_as__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_roll_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rot90_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rot90_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scalar_tensor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scalar_tensor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_searchsorted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sgn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sigmoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_blackman_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_general_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_hann_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_nuttall_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_sampled_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_airy_ai_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_j1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_y0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_y1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_erfcx_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_hermite_polynomial_he_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i1e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_laguerre_polynomial_l_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_legendre_polynomial_p_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_ndtri_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_scaled_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_scaled_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_xlog1py_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_list_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_multiple_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_to_size_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_t_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_along_dim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_along_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensordot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_sparse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_transpose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_transpose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapz_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triangular_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triangular_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trunc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unbind_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unbind_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unflatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_uniform_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unique_consecutive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsafe_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsafe_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vstack_cpu_complex128 2023-03-31T07:05:22.1601328Z 2023-03-31T07:05:22.1602511Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_H_cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1602945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_T_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1603553Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___getitem___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1603976Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___radd___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1604371Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___radd___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1604785Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rdiv___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1605439Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rdiv___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1605836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmatmul___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1606250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmul___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1606654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rsub___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1607140Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__segment_reduce_offsets_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1607586Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_abs_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1607986Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acos_cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1608436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acos_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:22.1608839Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acosh_cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:22.1609226Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_add_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1609624Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_add_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:22.1610029Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addbmm_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1610435Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addbmm_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:22.1610824Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcdiv_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1611235Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcmul_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1611640Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1612023Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:22.1612437Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_decomposed_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:22.1612848Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addr_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:22.1613545Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_allclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:22.1613995Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_amin_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:22.1614579Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_aminmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:22.1615221Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_arange_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:22.1615852Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argsort_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:22.1616479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argwhere_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 2%] 2023-03-31T07:05:22.1617124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argwhere_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 2%] 2023-03-31T07:05:22.1617606Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_partial_views_cpu_complex128 XFAIL [ 2%] 2023-03-31T07:05:22.1618116Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_scatter_cpu_float64 SKIPPED (Passes on complex128 and float64 only) [ 2%] 2023-03-31T07:05:22.1618562Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asin_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1618948Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asinh_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:22.1619350Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1619774Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atanh_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:22.1620218Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_1d_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1620623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_3d_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:22.1621077Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_3d_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1621483Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_baddbmm_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:22.1621869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_baddbmm_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1622266Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bernoulli_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:22.1622740Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bfloat16_cpu_float64 XFAIL [ 2%] 2023-03-31T07:05:22.1623152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_block_diag_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:22.1623550Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_block_diag_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1624131Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bool_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:22.1624612Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_tensors_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:22.1625045Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_tensors_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1625452Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_to_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:22.1625867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cartesian_prod_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1626504Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cat_cpu_float64 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 3%] 2023-03-31T07:05:22.1626991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdist_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1627379Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdouble_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1627782Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cfloat_cpu_complex128 XFAIL [ 3%] 2023-03-31T07:05:22.1628360Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_char_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:22.1628821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:22.1629216Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:22.1629630Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_inverse_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:22.1630057Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_solve_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1630459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chunk_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1630899Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1631301Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_max_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1631702Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clamp_min_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1632094Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_column_stack_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1632544Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_column_stack_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1632987Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_combinations_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1633403Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_combinations_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1633824Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_complex_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1634234Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_physical_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1634650Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_physical_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1635066Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_constant_pad_nd_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:22.1635473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_constant_pad_nd_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1635883Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_contiguous_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:22.1636289Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_copysign_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1636676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cos_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1637274Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_count_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 5%] 2023-03-31T07:05:22.1637924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_count_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 5%] 2023-03-31T07:05:22.1638398Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cov_cpu_complex128 SKIPPED (Barely fails) [ 5%] 2023-03-31T07:05:22.1638797Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cross_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1639193Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cummax_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1639593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumprod_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1639988Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumsum_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1640397Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumulative_trapezoid_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:22.1640831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumulative_trapezoid_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1641242Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_deg2rad_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1641640Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:22.1642024Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:22.1642425Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_embed_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:22.1642840Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diag_embed_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:22.1643530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_scatter_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1643933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diff_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1644332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dist_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1657736Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dist_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:22.1658541Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_no_rounding_mode_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1659007Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dot_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:22.1659414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_double_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1659887Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dstack_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1660299Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dstack_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:22.1660699Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_einsum_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:22.1661102Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_einsum_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:22.1661780Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:22.1662428Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:22.1663201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_permuted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:22.1663855Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_equal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 7%] 2023-03-31T07:05:22.1664487Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_equal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 7%] 2023-03-31T07:05:22.1664942Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erf_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1665332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erfinv_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1665740Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp2_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:22.1666145Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp2_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1666532Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expm1_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1666936Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft2_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:22.1667347Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft2_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1667749Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:22.1668185Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftn_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:22.1668593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftn_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1669004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftshift_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:22.1669417Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fftshift_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1669869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft2_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:22.1670277Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1670678Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1671085Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfftn_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1671513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfftn_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1671947Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft2_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1672353Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft2_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1672778Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1673180Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifft_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1673588Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftn_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1674004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftshift_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1674401Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfft2_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1674808Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft2_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1675212Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1675620Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfftn_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:22.1676017Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfftn_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:22.1676422Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfft_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1676822Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfftn_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1677225Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flip_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:22.1677610Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flip_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1678013Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fliplr_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:22.1678413Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flipud_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:22.1678801Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_cpu_float64 XFAIL [ 9%] 2023-03-31T07:05:22.1679201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_power_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1679791Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_floor_divide_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:22.1680248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmax_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1680627Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmin_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:22.1681211Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:22.1681842Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:22.1682525Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ge_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:22.1683401Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geometric_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:22.1683892Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gradient_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:22.1684304Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gradient_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:22.1684944Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:22.1685424Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_half_cpu_complex128 XFAIL [ 10%] 2023-03-31T07:05:22.1685862Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_half_cpu_float64 XFAIL [ 10%] 2023-03-31T07:05:22.1686447Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_heaviside_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:22.1687089Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histogram_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:22.1687716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histogramdd_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:22.1688183Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hsplit_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:22.1688582Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hstack_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:22.1689161Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_igammac_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:22.1689613Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_imag_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:22.1690025Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_add_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:22.1690438Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_copy_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:22.1690849Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_copy_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:22.1691254Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_fill_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:22.1691666Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_reduce_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:22.1692081Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_select_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:22.1692498Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_select_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:22.1692892Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_inner_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:22.1693473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_int_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1694114Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1694757Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isfinite_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1695378Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1696052Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isinf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1696687Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isnan_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1697320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isposinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:22.1697801Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_istft_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:22.1698272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_2inputs_2outputs_cpu_complex128 SKIPPED (Only runs on cuda) [ 11%] 2023-03-31T07:05:22.1698749Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_2inputs_2outputs_cpu_float64 SKIPPED (Only runs on cuda) [ 11%] 2023-03-31T07:05:22.1699293Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_4inputs_with_extra_args_cpu_float64 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1699775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_cpu_complex128 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1700214Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_cpu_float64 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1700690Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_return_by_ref_cpu_complex128 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1701175Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_binary_return_by_ref_cpu_float64 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1701654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_unary_cpu_complex128 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1702105Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_unary_cpu_float64 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:05:22.1702531Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kron_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:22.1703014Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kron_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:22.1703412Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_kthvalue_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:22.1703805Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ldexp_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:22.1704207Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ldexp_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:22.1704787Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_le_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 12%] 2023-03-31T07:05:22.1705231Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lerp_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:22.1705637Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lgamma_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:22.1706049Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cross_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:22.1706475Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_diagonal_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:22.1706880Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eig_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:22.1707296Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigh_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1707708Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvals_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1708131Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvalsh_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:22.1708573Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvalsh_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1709007Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_householder_product_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1709429Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1709864Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_ex_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1710482Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 13%] 2023-03-31T07:05:22.1711156Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_ex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 13%] 2023-03-31T07:05:22.1711858Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_solve_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 13%] 2023-03-31T07:05:22.1712342Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_grad_oriented_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:22.1712752Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:22.1713162Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1713577Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:22.1714002Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_ex_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:22.1714410Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_ex_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1714828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_solve_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:22.1715434Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:22.1715910Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_multi_dot_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:22.1716318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_multi_dot_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1716737Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:22.1717159Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_hermitian_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1717661Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_singular_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 14%] 2023-03-31T07:05:22.1718199Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_singular_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 14%] 2023-03-31T07:05:22.1718677Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_slogdet_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1719091Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_ex_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1719520Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_triangular_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:22.1719927Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svd_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1720339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorinv_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1720786Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vander_cpu_complex128 PASSED [ 15%] 2023-03-31T07:05:22.1721213Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vector_norm_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1721798Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 15%] 2023-03-31T07:05:22.1722466Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 15%] 2023-03-31T07:05:22.1722963Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log10_cpu_complex128 PASSED [ 15%] 2023-03-31T07:05:22.1723601Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log10_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1723994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log1p_cpu_complex128 PASSED [ 15%] 2023-03-31T07:05:22.1724394Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log1p_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1724787Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log2_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1725160Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1725575Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_with_dtype_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1726001Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp2_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1726404Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:22.1726802Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logcumsumexp_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:22.1727222Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logdet_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:22.1727619Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logdet_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:22.1728222Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_and_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1728860Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_not_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1729503Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_or_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1730151Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_xor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1730609Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logit_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:22.1731168Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_long_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1731788Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 16%] 2023-03-31T07:05:22.1732235Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:22.1732644Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_solve_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:22.1733045Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_unpack_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:22.1733449Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_unpack_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:22.1733914Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mH_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:22.1734314Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_amin_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1734895Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 17%] 2023-03-31T07:05:22.1735580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 17%] 2023-03-31T07:05:22.1736090Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumprod_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1736573Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumsum_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1736984Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_logaddexp_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1737410Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_logsumexp_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1737826Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_mean_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1738240Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_prod_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1738641Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_select_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1739053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_softmin_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1739459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_std_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1739873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_sum_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1740267Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_sum_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:22.1740676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_var_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:22.1741079Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matmul_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1741484Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matrix_exp_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1741904Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_pool2d_with_indices_backward_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1742359Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_reduction_no_dim_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1742883Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_reduction_with_dim_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1743280Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mean_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1743678Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_median_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1744100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_variadic_tensors_cpu_complex128 PASSED [ 18%] 2023-03-31T07:05:22.1744530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_binary_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1744935Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_reduction_no_dim_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1745365Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_min_reduction_with_dim_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1745784Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mm_cpu_complex128 PASSED [ 18%] 2023-03-31T07:05:22.1746213Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mm_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1746593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mode_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1746989Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_movedim_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:22.1747382Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mul_cpu_complex128 PASSED [ 19%] 2023-03-31T07:05:22.1747802Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mul_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1748209Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mv_cpu_complex128 PASSED [ 19%] 2023-03-31T07:05:22.1748602Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mv_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1749026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanmean_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1749414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanmedian_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1749825Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nanquantile_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1750429Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_copy_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 19%] 2023-03-31T07:05:22.1750899Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1751307Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_dropout_backward_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1751738Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_layer_norm_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:22.1752330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ne_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 19%] 2023-03-31T07:05:22.1752972Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 19%] 2023-03-31T07:05:22.1753621Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_strided_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 19%] 2023-03-31T07:05:22.1754294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_strided_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 19%] 2023-03-31T07:05:22.1754943Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 20%] 2023-03-31T07:05:22.1755580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nextafter_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 20%] 2023-03-31T07:05:22.1756072Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool2d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1756514Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool1d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1756964Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool3d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1757418Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_alpha_dropout_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1757867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool1d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1758287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool2d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1758761Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_avg_pool3d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1759198Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_bilinear_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1759624Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_celu_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1760040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv1d_cpu_complex128 PASSED [ 20%] 2023-03-31T07:05:22.1760495Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv1d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:22.1760962Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose1d_cpu_complex128 PASSED [ 20%] 2023-03-31T07:05:22.1761423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose2d_cpu_complex128 PASSED [ 20%] 2023-03-31T07:05:22.1761880Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_ctc_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1762310Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout2d_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1762746Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout3d_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1763427Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_dropout_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1763869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_embedding_bag_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1764308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_embedding_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1764778Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_without_train_cpu_complex128 PASSED [ 21%] 2023-03-31T07:05:22.1765247Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_fractional_max_pool3d_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1765701Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_gaussian_nll_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1766135Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_glu_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1766570Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_grid_sample_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1766995Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_group_norm_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1767432Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardshrink_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1767881Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hinge_embedding_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1768324Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_huber_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:22.1768741Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_instance_norm_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1769188Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_linear_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1769648Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_trilinear_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1770095Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_kl_div_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1770513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_l1_loss_cpu_complex128 PASSED [ 22%] 2023-03-31T07:05:22.1771029Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_layer_norm_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1771464Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_leaky_relu_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1771891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_linear_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1772318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_margin_ranking_loss_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1772806Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool1d_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1773281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool2d_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1773763Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool2d_grad_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1774196Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool3d_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1774637Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool3d_grad_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1775077Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multi_margin_loss_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:22.1775541Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multilabel_soft_margin_loss_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1775981Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_normalize_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1776417Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_circular_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:22.1776859Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_constant_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:22.1777309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_reflect_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:22.1777735Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_reflect_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1778169Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_replicate_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1778617Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pairwise_distance_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:22.1779057Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pdist_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1779479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_shuffle_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:22.1779932Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_poisson_nll_loss_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1780369Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_relu6_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1780791Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_relu_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1781200Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_rrelu_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1781652Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_scaled_dot_product_attention_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:22.1782101Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_selu_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1782869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_silu_complex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:22.1783401Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softshrink_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1783857Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_loss_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:22.1784332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_with_distance_loss_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:22.1784821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_unfold_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1785272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_upsample_nearest_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1785692Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1786124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_fro_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:22.1786530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_nuc_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:22.1786916Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_nuc_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:22.1787347Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_cpu_float64 SKIPPED (Gradients are incorrect!) [ 24%] 2023-03-31T07:05:22.1787979Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_in_place_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:22.1788633Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:22.1789262Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:22.1789726Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ormqr_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:22.1790125Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ormqr_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1790517Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_outer_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1790902Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pca_lowrank_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1791311Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_permute_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:22.1791711Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_permute_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1792105Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pinverse_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:22.1792527Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_0_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1792974Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_1_cpu_float64 SKIPPED (Skipped!) [ 25%] 2023-03-31T07:05:22.1793432Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_4_cpu_float64 SKIPPED (Skipped!) [ 25%] 2023-03-31T07:05:22.1793850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_positive_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1794250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pow_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:22.1794642Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pow_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1795030Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_put_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1795441Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_quantile_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:22.1796035Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rand_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:22.1796676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rand_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:22.1797330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:22.1797844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reciprocal_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:22.1798253Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_remainder_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:22.1798680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_renorm_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:22.1799087Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:22.1799493Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_interleave_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:22.1799913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_as_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:22.1800325Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_as_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:22.1800738Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:22.1801150Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_3_cpu_float64 SKIPPED (Skipped!) [ 26%] 2023-03-31T07:05:22.1801601Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_neg_3_cpu_float64 SKIPPED (Skipped!) [ 26%] 2023-03-31T07:05:22.1802027Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsqrt_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:22.1802423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsqrt_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:22.1802799Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsub_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1803633Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scalar_tensor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1804112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_add_cpu_complex128 PASSED [ 27%] 2023-03-31T07:05:22.1804514Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_add_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1804935Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_amin_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1805363Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_mean_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1805785Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_prod_cpu_float64 XFAIL [ 27%] 2023-03-31T07:05:22.1806186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1806771Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_short_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1807233Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sigmoid_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:22.1807832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1808565Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_gaussian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1809256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_general_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1809930Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_nuttall_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1810612Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signbit_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:22.1811118Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinc_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1811547Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinh_cpu_complex128 PASSED [ 28%] 2023-03-31T07:05:22.1811946Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1812339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1812741Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_with_dtype_cpu_complex128 PASSED [ 28%] 2023-03-31T07:05:22.1813171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_softmax_with_dtype_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1813577Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sort_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1814175Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:22.1814837Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_y0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:22.1815508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:22.1816215Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 28%] 2023-03-31T07:05:22.1816926Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 28%] 2023-03-31T07:05:22.1817413Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_entr_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1817813Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_erfcx_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1818232Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i0e_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:22.1818639Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i1e_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1819256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_laguerre_polynomial_l_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:22.1819943Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_legendre_polynomial_p_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 29%] 2023-03-31T07:05:22.1820438Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_log_ndtr_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1821057Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_i1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:22.1821775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:22.1822242Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_ndtri_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1822967Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_scaled_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:22.1823739Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 29%] 2023-03-31T07:05:22.1824499Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 29%] 2023-03-31T07:05:22.1825073Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_xlog1py_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1825652Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_zeta_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:22.1826108Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1826516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_list_args_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:22.1826928Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sqrt_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:22.1827315Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sqrt_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1827717Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1828128Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1828516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1828930Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_unbiased_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1829354Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_unbiased_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1829774Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_unbiased_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1830172Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stft_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1830568Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stft_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1830955Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1831360Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_to_size_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1831747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1832142Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_lowrank_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1832540Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_t_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:22.1832932Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_t_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:22.1833322Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_along_dim_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1833722Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tile_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1834121Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1834530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1835121Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_sparse_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 31%] 2023-03-31T07:05:22.1835576Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_topk_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1836004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trace_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1836424Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapezoid_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1836838Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapezoid_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1837277Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapz_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1837693Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triangular_solve_cpu_complex128 XFAIL [ 31%] 2023-03-31T07:05:22.1838104Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triangular_solve_cpu_float64 XFAIL [ 31%] 2023-03-31T07:05:22.1838512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tril_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1838914Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_true_divide_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:22.1839323Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unbind_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:22.1839716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unflatten_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1840122Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unflatten_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:22.1840527Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1840933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:22.1841514Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_uniform_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:22.1842167Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unique_consecutive_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:22.1842814Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unique_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:22.1843452Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsafe_split_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1843851Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsqueeze_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:22.1844259Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1844684Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_unbiased_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1845094Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_unbiased_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:22.1845516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_unbiased_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1845934Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_unbiased_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:22.1846339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vdot_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:22.1846725Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vdot_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1847193Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_complex_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1847607Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:22.1848012Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_copy_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1848402Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:22.1848855Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1849291Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vsplit_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:22.1849691Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vsplit_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1850111Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_where_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1850512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zero__cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:22.1850913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zero__cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1851486Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 33%] 2023-03-31T07:05:22.1851941Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_H_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1852326Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_T_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1852721Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___getitem___cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:22.1853117Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___getitem___cpu_float64 PASSED [ 33%] 2023-03-31T07:05:22.1853519Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rdiv___cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1853925Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmatmul___cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1854333Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmatmul___cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1854717Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmul___cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1855129Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rpow___cpu_float64 SKIPPED (Skipped!) [ 34%] 2023-03-31T07:05:22.1855555Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__softmax_backward_data_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1855989Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__upsample_bilinear2d_aa_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1856387Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acos_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1856781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acosh_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1857176Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_add_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1857578Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addbmm_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1857965Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcdiv_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1858369Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1858763Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:22.1859163Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_decomposed_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:22.1859810Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_all_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1860442Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_all_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1861073Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_allclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1861723Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_any_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1862372Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_any_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1863155Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1863784Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argsort_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1864419Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argwhere_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1865044Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argwhere_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:22.1865523Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_partial_views_cpu_complex128 XFAIL [ 35%] 2023-03-31T07:05:22.1865951Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_partial_views_cpu_float64 XFAIL [ 35%] 2023-03-31T07:05:22.1866389Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_scatter_cpu_complex128 XFAIL [ 35%] 2023-03-31T07:05:22.1866792Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_scatter_cpu_float64 XFAIL [ 35%] 2023-03-31T07:05:22.1867198Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asin_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:22.1867596Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asinh_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:22.1867998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:22.1868389Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_1d_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:22.1868796Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_2d_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:22.1869205Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_3d_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:22.1869595Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_baddbmm_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:22.1869996Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bernoulli_cpu_float64 XFAIL [ 36%] 2023-03-31T07:05:22.1870395Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bfloat16_cpu_complex128 XFAIL [ 36%] 2023-03-31T07:05:22.1870796Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_block_diag_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:22.1871184Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bmm_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:22.1871758Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bool_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:22.1872380Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bool_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:22.1872886Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_tensors_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:22.1873467Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bucketize_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:22.1874100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_byte_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:22.1874593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cartesian_prod_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:22.1875209Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cauchy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 37%] 2023-03-31T07:05:22.1875677Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdist_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1876073Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdouble_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1876472Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cfloat_cpu_complex128 XFAIL [ 37%] 2023-03-31T07:05:22.1876867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cfloat_cpu_float64 XFAIL [ 37%] 2023-03-31T07:05:22.1877248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chalf_cpu_complex128 XFAIL [ 37%] 2023-03-31T07:05:22.1877821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_char_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 37%] 2023-03-31T07:05:22.1878279Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:22.1878690Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_inverse_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1879097Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_solve_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1879501Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chunk_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:22.1879906Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_max_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1880309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_min_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1880693Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clone_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:22.1881099Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_column_stack_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:22.1881515Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_physical_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1881919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_physical_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:22.1882333Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_constant_pad_nd_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1882748Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_contiguous_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:22.1883351Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_copysign_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:22.1883763Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_corrcoef_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:22.1884166Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cos_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1884568Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cosh_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1885169Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_count_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 38%] 2023-03-31T07:05:22.1885884Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_count_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 38%] 2023-03-31T07:05:22.1886339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cov_cpu_complex128 XFAIL [ 38%] 2023-03-31T07:05:22.1886736Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cross_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1887176Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumprod_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1887626Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumulative_trapezoid_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1888047Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:22.1888478Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1888886Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_embed_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1889283Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diag_embed_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1889687Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagflat_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1890100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_copy_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1890498Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1890916Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_scatter_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1891325Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diff_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1891716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diff_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1892100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_digamma_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1892496Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dist_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1892888Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dist_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1893296Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_no_rounding_mode_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1893707Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_trunc_rounding_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:22.1894124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_double_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:22.1894521Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dsplit_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:22.1894919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_einsum_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:22.1895488Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 40%] 2023-03-31T07:05:22.1896135Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_permuted_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 40%] 2023-03-31T07:05:22.1896792Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_permuted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 40%] 2023-03-31T07:05:22.1897425Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eq_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 40%] 2023-03-31T07:05:22.1898043Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_equal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 40%] 2023-03-31T07:05:22.1898517Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erfc_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:22.1898917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erfinv_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:22.1899312Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp2_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:22.1899688Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp2_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:22.1900152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:22.1900584Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exp_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:22.1900973Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_as_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:22.1901422Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:22.1901828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:22.1902219Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expm1_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:22.1902874Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:22.1903516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eye_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:22.1904143Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eye_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:22.1904596Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft2_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1904991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft2_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:22.1905398Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1905807Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftn_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1906214Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftshift_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1906614Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft2_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1907019Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:22.1907423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfftn_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1907828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft2_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:22.1908219Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:22.1908634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftn_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1909036Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftn_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1909437Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftshift_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1909852Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifftshift_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1910257Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfft_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1910642Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfftn_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1911084Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft2_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1911488Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft2_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1911890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1912282Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfftn_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1912707Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfft2_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1913130Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfftn_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1913523Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fill_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1913928Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fill_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:22.1914323Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flatten_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:22.1914722Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flip_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:22.1915111Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flip_cpu_float64 PASSED [ 43%] 2023-03-31T07:05:22.1915498Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fliplr_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:22.1915898Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flipud_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:22.1916310Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_power_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:22.1916696Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_floor_cpu_float64 PASSED [ 43%] 2023-03-31T07:05:22.1917284Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_floor_divide_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1917741Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmin_cpu_float64 PASSED [ 43%] 2023-03-31T07:05:22.1918318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1918943Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1919402Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gather_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:22.1919798Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gather_cpu_float64 PASSED [ 43%] 2023-03-31T07:05:22.1920382Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geometric_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1921001Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geqrf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1921625Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_geqrf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:22.1922086Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gradient_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:22.1922503Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_grid_sampler_2d_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:22.1923195Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1923974Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_heaviside_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1924623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histogramdd_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1925085Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hsplit_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:22.1925660Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_igamma_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1926311Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_igammac_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1926798Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_imag_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:22.1927239Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_copy_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:22.1927638Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_copy_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:22.1928044Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_put_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:22.1928446Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_inner_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:22.1929028Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1929654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isfinite_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:22.1930288Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isinf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:22.1930921Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isneginf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:22.1931548Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isposinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:22.1932175Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isreal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:22.1932624Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_istft_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:22.1933078Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_complex128 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T07:05:22.1933550Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_cpu_float64 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T07:05:22.1934020Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_return_by_ref_cpu_float64 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T07:05:22.1934476Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_unary_cpu_complex128 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T07:05:22.1934928Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_unary_cpu_float64 SKIPPED (Only runs on cuda) [ 45%] 2023-03-31T07:05:22.1935355Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ldexp_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:22.1935753Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lerp_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:22.1936147Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cross_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:22.1936564Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_singular_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:22.1937016Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_diagonal_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:22.1937436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eig_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1937834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eig_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:22.1938236Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigh_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1938676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigh_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:22.1939114Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_householder_product_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1939597Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_householder_product_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:22.1940020Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:22.1940423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_ex_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1941012Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 46%] 2023-03-31T07:05:22.1941685Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_ex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 46%] 2023-03-31T07:05:22.1942349Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_solve_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 46%] 2023-03-31T07:05:22.1942922Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_cpu_complex128 SKIPPED (Skipped!) [ 46%] 2023-03-31T07:05:22.1943368Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1943793Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:22.1944213Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:22.1944625Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:22.1945032Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_ex_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1945456Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_ex_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:22.1945879Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_norm_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1946299Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_norm_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:22.1946701Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_power_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:22.1947314Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 47%] 2023-03-31T07:05:22.1947973Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 47%] 2023-03-31T07:05:22.1948648Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 47%] 2023-03-31T07:05:22.1949143Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_multi_dot_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1949604Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1950015Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:22.1950445Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_complex128 XFAIL [ 47%] 2023-03-31T07:05:22.1950865Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1951318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_hermitian_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:22.1951849Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_singular_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 48%] 2023-03-31T07:05:22.1952353Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_qr_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1952757Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_slogdet_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:22.1953173Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_slogdet_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1953580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1953998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_triangular_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1954411Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svdvals_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:22.1954833Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorinv_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1955261Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorsolve_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:22.1955685Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vecdot_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:22.1956080Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vecdot_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1956671Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:22.1957307Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:22.1957762Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log1p_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1958140Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log2_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:22.1958528Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1958941Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_with_dtype_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1959359Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:22.1959750Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1960170Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logcumsumexp_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:22.1960771Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_not_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:22.1961415Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_xor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:22.1962076Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:22.1962575Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logsumexp_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1963273Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_long_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:22.1963940Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:22.1964405Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:22.1964804Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_solve_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1965234Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mH_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1965619Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mT_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:22.1966188Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 50%] 2023-03-31T07:05:22.1966657Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumprod_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1967070Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumsum_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1967488Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_logaddexp_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1967893Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_logsumexp_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1968308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_mean_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:22.1968719Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_median_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1969112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_norm_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1969521Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_normalize_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:22.1969942Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_normalize_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1970356Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_prod_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:22.1970752Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_scatter_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1971163Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_select_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:22.1971571Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_softmin_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1971973Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_std_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:22.1972360Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_sum_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1972762Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_sum_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1973167Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_var_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1973574Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_var_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1973959Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matmul_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1974361Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matmul_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1974818Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matrix_exp_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1975212Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_matrix_exp_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1975623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_reduction_with_dim_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1976081Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_variadic_tensors_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1976518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_binary_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1976907Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mm_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1977322Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mul_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1977717Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mv_cpu_complex128 PASSED [ 51%] 2023-03-31T07:05:22.1978102Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mv_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:22.1978481Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanmean_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:22.1978879Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanmedian_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:22.1979275Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nansum_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:22.1979675Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_cpu_complex128 PASSED [ 52%] 2023-03-31T07:05:22.1980087Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_batch_norm_cpu_float64 SKIPPED (Skipped!) [ 52%] 2023-03-31T07:05:22.1980524Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_dropout_backward_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:22.1981124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ne_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1981747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ne_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1982177Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_neg_cpu_complex128 PASSED [ 52%] 2023-03-31T07:05:22.1982908Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1983569Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_strided_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1984211Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1984836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1985471Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:22.1985958Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool2d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:22.1986414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool3d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1986849Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_alpha_dropout_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1987327Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool1d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1987759Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool3d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1988190Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_batch_norm_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1988606Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_bilinear_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1989078Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_binary_cross_entropy_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1989575Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose2d_cpu_complex128 PASSED [ 53%] 2023-03-31T07:05:22.1990026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose2d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1990497Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cosine_embedding_loss_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1990940Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout2d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1991370Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout3d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1991803Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_dropout_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1992219Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_embedding_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:22.1992674Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_with_train_cpu_float64 XFAIL [ 53%] 2023-03-31T07:05:22.1993142Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_fractional_max_pool2d_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1993577Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_glu_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1993990Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_grid_sample_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1994418Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_group_norm_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1994854Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardsigmoid_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1995287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardswish_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1995708Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_huber_loss_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1996150Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_bicubic_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1996606Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_bilinear_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1997046Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_kl_div_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1997457Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_layer_norm_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1997879Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_linear_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1998315Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_local_response_norm_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1998751Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_margin_ranking_loss_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1999269Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool1d_grad_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:22.1999711Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool2d_grad_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2000164Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool3d_cpu_float64 SKIPPED (Skipped!) [ 55%] 2023-03-31T07:05:22.2000592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_mse_loss_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2001040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_nll_loss_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2001499Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_normalize_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:22.2001959Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_normalize_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2002380Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_circular_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2002810Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_constant_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2003363Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_reflect_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:22.2003806Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_reflect_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2004241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pairwise_distance_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:22.2004683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_shuffle_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2005129Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:22.2005574Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2005992Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_prelu_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:22.2006408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_relu6_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2006829Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_rrelu_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2007294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_scaled_dot_product_attention_cpu_float64 SKIPPED (Skipped!) [ 56%] 2023-03-31T07:05:22.2007954Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_silu_complex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 56%] 2023-03-31T07:05:22.2008454Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_soft_margin_loss_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2008887Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2009316Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softshrink_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2009736Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softsign_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:22.2010171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softsign_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2010605Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_tanhshrink_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2011053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:22.2011559Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2012030Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:22.2012518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2012998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_unfold_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:22.2013456Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_upsample_bilinear_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2013903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_upsample_nearest_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2014547Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:22.2014998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2015382Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_fro_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2015781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_inf_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:22.2016186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_inf_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2016590Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_nuc_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:22.2017151Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:22.2017777Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:22.2018406Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:22.2018859Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ormqr_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2019248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_outer_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:22.2019648Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_outer_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2020040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_permute_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:22.2020441Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pinverse_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:22.2020830Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polar_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2021238Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_0_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2021682Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_1_cpu_float64 SKIPPED (Skipped!) [ 58%] 2023-03-31T07:05:22.2022112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_positive_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2022497Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pow_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2022972Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_prod_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:22.2023428Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_prod_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2023808Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_qr_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:22.2024250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_quantile_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2024647Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rad2deg_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2025234Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rand_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:22.2025889Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randint_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:22.2026379Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ravel_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:22.2026776Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ravel_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:22.2027203Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_real_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2027605Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_interleave_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2028022Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_as_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2028432Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_as_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2028833Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2029406Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 59%] 2023-03-31T07:05:22.2030033Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 59%] 2023-03-31T07:05:22.2030500Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_conj_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2030913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_conj_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2031300Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_roll_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2031694Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rot90_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2032086Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rot90_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2032467Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsqrt_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2032855Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsqrt_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:22.2033258Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_add_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:22.2033666Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:22.2034063Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_amax_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2034482Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_amin_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2034902Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_mean_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2035320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_prod_cpu_float64 XFAIL [ 60%] 2023-03-31T07:05:22.2035904Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_searchsorted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:22.2036375Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_scatter_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2036831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sgn_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:22.2037406Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_short_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:22.2037850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sigmoid_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:22.2038276Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sigmoid_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2038697Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sign_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:22.2039300Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:22.2040003Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_gaussian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:22.2040680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_general_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:22.2041371Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_general_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:22.2042041Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:22.2042707Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_nuttall_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:22.2043458Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signbit_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:22.2043905Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sin_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:22.2044301Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinh_cpu_complex128 PASSED [ 61%] 2023-03-31T07:05:22.2044693Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinh_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:22.2045073Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_cpu_complex128 PASSED [ 61%] 2023-03-31T07:05:22.2045473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:22.2045869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_slice_scatter_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:22.2046259Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:22.2046672Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_with_dtype_cpu_complex128 PASSED [ 61%] 2023-03-31T07:05:22.2047107Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_mm_reduce_cpu_float64 SKIPPED (Skipped!) [ 61%] 2023-03-31T07:05:22.2047555Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_sampled_addmm_cpu_complex128 SKIPPED (Skipped!) [ 61%] 2023-03-31T07:05:22.2047997Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sparse_sampled_addmm_cpu_float64 SKIPPED (Skipped!) [ 61%] 2023-03-31T07:05:22.2048617Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_y0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:22.2049267Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_y1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:22.2050025Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:22.2050730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:22.2051244Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i0e_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:22.2051681Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i1_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:22.2052313Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_legendre_polynomial_p_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:22.2053045Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_i0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:22.2053711Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_i1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:22.2054187Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_ndtr_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:22.2054817Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_scaled_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:22.2055536Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:22.2056017Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_xlog1py_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:22.2056426Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_cpu_complex128 PASSED [ 62%] 2023-03-31T07:05:22.2056824Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:22.2057230Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_list_args_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2057634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_with_sizes_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2058053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_with_sizes_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2058455Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sqrt_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2058851Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2059243Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2059650Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_multiple_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2060068Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_squeeze_multiple_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2060468Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stack_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2060848Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2061267Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_unbiased_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2061687Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_unbiased_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2062080Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stft_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2062499Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stft_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:22.2062984Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sub_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:22.2063368Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2063752Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_to_size_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2064181Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2064596Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2064991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_svd_lowrank_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2065424Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_along_dim_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2065831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2066221Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tan_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2066607Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensor_split_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2067017Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensordot_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2067416Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tile_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2067801Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2068374Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_sparse_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 64%] 2023-03-31T07:05:22.2068837Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_sparse_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:22.2069236Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trace_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:22.2069631Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trace_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2070020Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapezoid_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2074950Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapz_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2078820Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triangular_solve_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2079233Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tril_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2079645Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triu_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2080040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triu_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2080431Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_true_divide_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2080831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trunc_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2081231Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unflatten_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2081654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:22.2082060Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:22.2082683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_uniform_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 65%] 2023-03-31T07:05:22.2083741Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_uniform_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 65%] 2023-03-31T07:05:22.2084372Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unique_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 65%] 2023-03-31T07:05:22.2084832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsafe_split_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2085230Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2085636Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2086051Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_unbiased_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2086505Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_mean_unbiased_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2086917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_unbiased_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2087312Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vdot_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2087715Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_real_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2088108Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_copy_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2100953Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2101386Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2101790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vstack_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2102198Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vstack_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2102705Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_where_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2103104Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_xlogy_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:22.2103658Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zero__cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:22.2104063Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zero__cpu_float64 PASSED [ 67%] 2023-03-31T07:05:22.2104855Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 67%] 2023-03-31T07:05:22.2105413Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_H_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2105948Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_T_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2106490Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___getitem___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2107036Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmatmul___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2107581Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmod___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2107788Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rpow___cpu_complex128 SKIPPED (Skipped!) [ 67%] 2023-03-31T07:05:22.2108100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rsub___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2108362Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rsub___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2108636Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__native_batch_norm_legit_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2108919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__segment_reduce_offsets_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2109314Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_abs_cpu_complex128 SKIPPED (In-place abs not supported for complex tensors) [ 67%] 2023-03-31T07:05:22.2109560Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acos_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:22.2109826Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acosh_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:22.2110084Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acosh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:22.2110286Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_add_cpu_complex128 PASSED [ 68%] 2023-03-31T07:05:22.2110494Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcdiv_cpu_complex128 PASSED [ 68%] 2023-03-31T07:05:22.2110696Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcdiv_cpu_float64 PASSED [ 68%] 2023-03-31T07:05:22.2110887Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addr_cpu_complex128 PASSED [ 68%] 2023-03-31T07:05:22.2111272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_all_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:22.2111529Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_amax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:22.2111909Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_any_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:22.2112322Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_any_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:22.2112730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:22.2113114Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argsort_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:22.2113341Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_cpu_complex128 SKIPPED (Numerous errors) [ 68%] 2023-03-31T07:05:22.2113563Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_partial_views_cpu_complex128 XFAIL [ 68%] 2023-03-31T07:05:22.2113839Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_scatter_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:22.2114100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asinh_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:22.2114288Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan2_cpu_float64 PASSED [ 69%] 2023-03-31T07:05:22.2114581Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan_cpu_complex128 PASSED [ 69%] 2023-03-31T07:05:22.2114778Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atan_cpu_float64 PASSED [ 69%] 2023-03-31T07:05:22.2115041Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atanh_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2115300Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atanh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2115571Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_1d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2115805Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_baddbmm_cpu_float64 PASSED [ 69%] 2023-03-31T07:05:22.2116066Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bfloat16_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2116330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_block_diag_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2116715Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bool_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 69%] 2023-03-31T07:05:22.2116996Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_tensors_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2117271Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_tensors_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2117534Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_to_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:22.2117917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bucketize_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 69%] 2023-03-31T07:05:22.2118291Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_byte_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 69%] 2023-03-31T07:05:22.2118591Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cartesian_prod_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2119049Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cat_cpu_float64 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 70%] 2023-03-31T07:05:22.2119428Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cauchy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 70%] 2023-03-31T07:05:22.2119694Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdouble_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2119890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ceil_cpu_float64 PASSED [ 70%] 2023-03-31T07:05:22.2120152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chalf_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2120412Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chalf_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2120794Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_char_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 70%] 2023-03-31T07:05:22.2121201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_char_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 70%] 2023-03-31T07:05:22.2121451Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2121714Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2121992Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_inverse_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2122266Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_inverse_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2122593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_solve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2122861Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cholesky_solve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:22.2123267Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clone_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2123540Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_column_stack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2123813Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_combinations_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2124074Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2124287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_physical_cpu_complex128 PASSED [ 71%] 2023-03-31T07:05:22.2124559Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_constant_pad_nd_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2124872Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_contiguous_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2125155Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_contiguous_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2125422Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_corrcoef_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2125687Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_corrcoef_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2125890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cos_cpu_complex128 PASSED [ 71%] 2023-03-31T07:05:22.2126085Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cos_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:22.2126283Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cosh_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:22.2126689Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_count_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 71%] 2023-03-31T07:05:22.2126994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cov_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:22.2127258Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cross_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2127512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cummax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2127716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumprod_cpu_complex128 PASSED [ 72%] 2023-03-31T07:05:22.2127909Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumsum_cpu_complex128 PASSED [ 72%] 2023-03-31T07:05:22.2128112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumsum_cpu_float64 PASSED [ 72%] 2023-03-31T07:05:22.2128413Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2128679Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagflat_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2128949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2129207Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diff_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2129459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dist_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2129680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_no_rounding_mode_cpu_complex128 PASSED [ 72%] 2023-03-31T07:05:22.2129937Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dot_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2130198Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_double_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2130458Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_einsum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:22.2130869Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 72%] 2023-03-31T07:05:22.2131273Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2131654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eq_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2132026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eq_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2132222Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erf_cpu_float64 PASSED [ 73%] 2023-03-31T07:05:22.2132420Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erfc_cpu_float64 PASSED [ 73%] 2023-03-31T07:05:22.2132622Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp2_cpu_complex128 PASSED [ 73%] 2023-03-31T07:05:22.2132888Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:22.2133182Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:22.2133382Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expm1_cpu_float64 PASSED [ 73%] 2023-03-31T07:05:22.2133769Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2134138Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eye_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2134515Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_eye_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:22.2134823Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft2_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:22.2135093Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftshift_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:22.2135358Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft2_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:22.2135623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2135890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfftn_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2136158Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft2_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2136416Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2136688Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftshift_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2136954Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftshift_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2137243Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2137535Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2137782Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2138048Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2138315Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfftn_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2138580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2138844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2139073Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fill_cpu_complex128 PASSED [ 74%] 2023-03-31T07:05:22.2139336Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flipud_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:22.2139592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2139793Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_floor_cpu_float64 PASSED [ 75%] 2023-03-31T07:05:22.2140188Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_floor_divide_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 75%] 2023-03-31T07:05:22.2140479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2140678Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_frac_cpu_float64 PASSED [ 75%] 2023-03-31T07:05:22.2140921Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_frexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2141302Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 75%] 2023-03-31T07:05:22.2141685Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 75%] 2023-03-31T07:05:22.2141952Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gather_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2142223Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gradient_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2142485Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gradient_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2142845Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_grid_sampler_2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2143262Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 75%] 2023-03-31T07:05:22.2143563Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_half_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2143828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_half_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:22.2144211Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_heaviside_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:22.2144586Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histc_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:22.2144956Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histogram_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:22.2145216Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hsplit_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:22.2145514Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hstack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:22.2145778Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hstack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:22.2145977Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hypot_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:22.2146171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_i0_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:22.2146553Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_igammac_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:22.2146816Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_imag_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:22.2147051Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_add_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:22.2147260Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_copy_cpu_complex128 PASSED [ 76%] 2023-03-31T07:05:22.2147462Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_copy_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:22.2147656Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_put_cpu_complex128 PASSED [ 76%] 2023-03-31T07:05:22.2147862Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_reduce_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:22.2148135Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_select_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 77%] 2023-03-31T07:05:22.2148402Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_inner_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 77%] 2023-03-31T07:05:22.2148791Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isfinite_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2149172Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isfinite_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2149544Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2149952Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isinf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2150357Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2150731Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isnan_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2151115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isreal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2151473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isreal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:22.2151715Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_2inputs_2outputs_cpu_complex128 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:05:22.2151963Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_complex128 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:05:22.2152242Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_return_by_ref_cpu_complex128 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:05:22.2152488Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_return_by_ref_cpu_float64 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:05:22.2152721Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_unary_cpu_complex128 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:05:22.2152950Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_unary_cpu_float64 SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:05:22.2153323Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_le_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 78%] 2023-03-31T07:05:22.2153596Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cond_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2153896Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cross_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2154166Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2154444Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_singular_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2154717Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_singular_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2154977Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_diagonal_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2155240Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eig_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2155512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigh_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2155786Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvals_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2156090Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvalsh_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2156390Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvalsh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2156660Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2156924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:22.2157324Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:22.2157722Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_solve_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:22.2157949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_cpu_complex128 SKIPPED (Skipped!) [ 79%] 2023-03-31T07:05:22.2158202Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_cpu_float64 SKIPPED (Skipped!) [ 79%] 2023-03-31T07:05:22.2158473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2158757Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lstsq_grad_oriented_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2159033Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2159308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_ex_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2159605Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_solve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2159879Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_norm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2160152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_power_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2160424Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_power_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:22.2160834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:22.2161255Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:22.2161671Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:22.2161945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_multi_dot_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2162239Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2162551Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_hermitian_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2162816Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_hermitian_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2163305Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_singular_cpu_complex128 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 80%] 2023-03-31T07:05:22.2163626Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_singular_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 80%] 2023-03-31T07:05:22.2163895Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_qr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2164173Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2164513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_triangular_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2164788Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorsolve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2165056Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vander_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2165326Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vander_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2165655Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vector_norm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:22.2166049Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 80%] 2023-03-31T07:05:22.2166434Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 80%] 2023-03-31T07:05:22.2166637Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log10_cpu_complex128 PASSED [ 81%] 2023-03-31T07:05:22.2166824Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log10_cpu_float64 PASSED [ 81%] 2023-03-31T07:05:22.2167026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log1p_cpu_float64 PASSED [ 81%] 2023-03-31T07:05:22.2167226Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_cpu_complex128 PASSED [ 81%] 2023-03-31T07:05:22.2167421Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_cpu_float64 PASSED [ 81%] 2023-03-31T07:05:22.2167699Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_with_dtype_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:22.2167966Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:22.2168267Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:22.2168563Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logdet_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:22.2168961Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_and_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 81%] 2023-03-31T07:05:22.2169347Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_not_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 81%] 2023-03-31T07:05:22.2169739Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_or_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 81%] 2023-03-31T07:05:22.2170108Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_or_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 81%] 2023-03-31T07:05:22.2170500Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_xor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 81%] 2023-03-31T07:05:22.2170698Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logit_cpu_float64 PASSED [ 81%] 2023-03-31T07:05:22.2171108Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:22.2171483Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_long_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:22.2171854Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lt_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:22.2172119Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_solve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2172388Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_unpack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2172681Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_unpack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2172936Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mH_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2173183Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mT_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2173574Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:22.2173838Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumprod_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2174109Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumprod_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2174382Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumsum_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2174647Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_cumsum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2174888Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_fill_cpu_complex128 PASSED [ 82%] 2023-03-31T07:05:22.2175191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_log_softmax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:22.2175468Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_logsumexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2175739Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_mean_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2176003Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_mean_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2176277Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_normalize_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2176545Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_prod_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2176843Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_prod_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2177115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_select_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2177365Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_select_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2177635Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_softmin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2177898Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_std_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2178197Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_var_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2178459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matmul_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2178716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matmul_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2178983Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matrix_exp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2179246Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_binary_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:22.2179538Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_pool2d_with_indices_backward_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2179810Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_reduction_no_dim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2180081Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_max_reduction_with_dim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2180372Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mean_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2180666Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mean_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2180940Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_list_of_tensors_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2181220Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_list_of_tensors_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2181509Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_variadic_tensors_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2181768Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2182026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2182315Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_msort_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2182517Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mul_cpu_complex128 PASSED [ 84%] 2023-03-31T07:05:22.2182790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mul_cpu_float64 PASSED [ 84%] 2023-03-31T07:05:22.2183186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_multinomial_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 84%] 2023-03-31T07:05:22.2183445Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mv_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:22.2183702Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mv_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2183955Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_1_cpu_float64 PASSED [ 85%] 2023-03-31T07:05:22.2184157Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_3_cpu_float64 PASSED [ 85%] 2023-03-31T07:05:22.2184366Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mvlgamma_mvlgamma_p_5_cpu_float64 PASSED [ 85%] 2023-03-31T07:05:22.2184631Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanmean_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2184890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nansum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2185285Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_copy_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:22.2185554Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2185812Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2186082Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_batch_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2186397Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_dropout_backward_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2186711Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_native_layer_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:22.2187091Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ne_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:22.2187493Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_strided_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:22.2187874Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_strided_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:22.2188262Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 86%] 2023-03-31T07:05:22.2188643Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 86%] 2023-03-31T07:05:22.2189054Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nextafter_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 86%] 2023-03-31T07:05:22.2189344Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2189632Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2189924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2190204Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2190517Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2190790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_bilinear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2191078Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_binary_cross_entropy_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2191380Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_binary_cross_entropy_with_logits_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2191597Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_celu_cpu_float64 PASSED [ 86%] 2023-03-31T07:05:22.2191880Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv1d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2192144Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv2d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2192414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:22.2192731Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2193050Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose2d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2193339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2193623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cosine_similarity_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2193906Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cross_entropy_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2194127Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout2d_cpu_float64 PASSED [ 87%] 2023-03-31T07:05:22.2194348Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout_cpu_float64 PASSED [ 87%] 2023-03-31T07:05:22.2194598Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_elu_cpu_float64 PASSED [ 87%] 2023-03-31T07:05:22.2194884Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_embedding_bag_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2195133Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_complex128 PASSED [ 87%] 2023-03-31T07:05:22.2195392Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_glu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2195675Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardshrink_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2195980Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardtanh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2196257Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_huber_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2196541Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_instance_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:22.2196833Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_bilinear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2197123Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_nearest_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2197420Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_trilinear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2197698Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_kl_div_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2197972Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_l1_loss_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2198273Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_l1_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2198578Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_layer_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2198794Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_leaky_relu_cpu_float64 PASSED [ 88%] 2023-03-31T07:05:22.2199064Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_linear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2199339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_local_response_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2199625Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2199947Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool1d_grad_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2200228Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2200444Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_mish_cpu_float64 PASSED [ 88%] 2023-03-31T07:05:22.2200719Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_mse_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:22.2201005Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multi_margin_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2201337Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multilabel_soft_margin_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2201611Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_nll_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2201891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_normalize_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2202171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_normalize_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2202454Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_circular_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2202746Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_replicate_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2203156Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_replicate_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2203439Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pairwise_distance_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2203813Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pairwise_distance_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2204132Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_shuffle_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2204459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_relu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2204945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_silu_complex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 89%] 2023-03-31T07:05:22.2205158Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_silu_cpu_float64 PASSED [ 89%] 2023-03-31T07:05:22.2205460Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_smooth_l1_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:22.2205744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_soft_margin_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2206071Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2206358Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2206634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softplus_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2206909Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softsign_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2207205Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2207565Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2207838Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_upsample_nearest_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2208227Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:22.2208494Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2208758Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_fro_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2209152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_in_place_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:22.2209539Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:22.2209919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:22.2210210Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ormqr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:22.2210494Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_outer_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2210758Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pca_lowrank_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2211024Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pinverse_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2211285Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pinverse_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2211542Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polar_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2211758Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_2_cpu_float64 SKIPPED (Skipped!) [ 91%] 2023-03-31T07:05:22.2212019Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_3_cpu_float64 SKIPPED (Skipped!) [ 91%] 2023-03-31T07:05:22.2212282Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pow_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2212536Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_pow_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2212790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_qr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2213051Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_quantile_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:22.2213470Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rand_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:22.2213851Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randint_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:22.2214240Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randint_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:22.2214616Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:22.2215000Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 92%] 2023-03-31T07:05:22.2215265Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ravel_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2215511Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_real_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2215720Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reciprocal_cpu_complex128 PASSED [ 92%] 2023-03-31T07:05:22.2215924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reciprocal_cpu_float64 PASSED [ 92%] 2023-03-31T07:05:22.2216129Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_remainder_cpu_float64 PASSED [ 92%] 2023-03-31T07:05:22.2216418Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2216729Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_interleave_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2217002Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_interleave_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2217268Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_as_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2217529Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2217916Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 92%] 2023-03-31T07:05:22.2218304Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 92%] 2023-03-31T07:05:22.2218706Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize_as__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 92%] 2023-03-31T07:05:22.2218974Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_conj_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:22.2219241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_neg_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2219501Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_roll_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2219791Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rot90_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2220048Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rot90_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2220247Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:22.2220458Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_0_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:22.2220684Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_3_cpu_float64 SKIPPED (Skipped!) [ 93%] 2023-03-31T07:05:22.2220891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsqrt_cpu_complex128 PASSED [ 93%] 2023-03-31T07:05:22.2221151Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsub_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2221551Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scalar_tensor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:22.2221927Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scalar_tensor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:22.2222131Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_add_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:22.2222342Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_amin_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:22.2222841Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_searchsorted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:22.2223137Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:22.2223407Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_scatter_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:22.2223608Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sgn_cpu_complex128 PASSED [ 94%] 2023-03-31T07:05:22.2223815Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sigmoid_cpu_complex128 PASSED [ 94%] 2023-03-31T07:05:22.2224011Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sign_cpu_float64 PASSED [ 94%] 2023-03-31T07:05:22.2224426Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_blackman_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:22.2224887Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_general_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:22.2225278Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:22.2225676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_hann_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:22.2226077Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_nuttall_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:22.2226279Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sin_cpu_complex128 PASSED [ 94%] 2023-03-31T07:05:22.2226508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sin_cpu_float64 PASSED [ 94%] 2023-03-31T07:05:22.2226713Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinc_cpu_complex128 PASSED [ 94%] 2023-03-31T07:05:22.2226910Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinc_cpu_float64 PASSED [ 94%] 2023-03-31T07:05:22.2227112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinh_cpu_complex128 PASSED [ 94%] 2023-03-31T07:05:22.2227306Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sinh_cpu_float64 PASSED [ 94%] 2023-03-31T07:05:22.2227570Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2227850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_with_dtype_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2228116Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_softmax_with_dtype_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2228375Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sort_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2228654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_sampled_addmm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2229080Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_airy_ai_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2229516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_j1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2229918Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_y0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2230316Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_y1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2230749Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 95%] 2023-03-31T07:05:22.2231019Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_erfcx_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2231441Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_hermite_polynomial_he_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2231743Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i1e_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:22.2232158Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_laguerre_polynomial_l_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:22.2232587Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_legendre_polynomial_p_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 95%] 2023-03-31T07:05:22.2232987Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_i1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 96%] 2023-03-31T07:05:22.2233405Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 96%] 2023-03-31T07:05:22.2233725Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_ndtr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2233994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_ndtri_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2234418Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_scaled_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 96%] 2023-03-31T07:05:22.2234845Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_scaled_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 96%] 2023-03-31T07:05:22.2235292Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 96%] 2023-03-31T07:05:22.2235567Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_xlog1py_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2235841Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_list_args_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2236048Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_cpu_complex128 PASSED [ 96%] 2023-03-31T07:05:22.2236283Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_cpu_float64 PASSED [ 96%] 2023-03-31T07:05:22.2236522Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_multiple_cpu_float64 PASSED [ 96%] 2023-03-31T07:05:22.2236768Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2237039Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2237307Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_unbiased_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:22.2237508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sub_cpu_complex128 PASSED [ 97%] 2023-03-31T07:05:22.2237705Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sub_cpu_float64 PASSED [ 97%] 2023-03-31T07:05:22.2237965Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2238256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2238527Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_to_size_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2238785Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2238983Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_t_cpu_complex128 PASSED [ 97%] 2023-03-31T07:05:22.2239183Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_t_cpu_float64 PASSED [ 97%] 2023-03-31T07:05:22.2239459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_along_dim_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2239742Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_along_dim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2240009Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2240204Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tan_cpu_float64 PASSED [ 97%] 2023-03-31T07:05:22.2240400Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tanh_cpu_float64 PASSED [ 97%] 2023-03-31T07:05:22.2240665Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensordot_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2240923Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tile_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:22.2241180Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2241573Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_sparse_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 98%] 2023-03-31T07:05:22.2241839Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trace_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2242078Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_transpose_cpu_complex128 PASSED [ 98%] 2023-03-31T07:05:22.2242310Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_transpose_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:22.2242567Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapezoid_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2242836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapezoid_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2243270Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapz_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2243570Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triangular_solve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2243844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triangular_solve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2244047Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trunc_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:22.2244366Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unbind_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2244626Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unbind_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2244889Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unflatten_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2245160Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_copy_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:22.2245427Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2245853Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_uniform_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 99%] 2023-03-31T07:05:22.2246239Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unique_consecutive_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 99%] 2023-03-31T07:05:22.2246512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsafe_split_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2246781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsafe_split_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2247048Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2247330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_unbiased_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2247604Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_unbiased_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2247871Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_unbiased_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2248173Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vdot_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2248472Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2248737Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2249007Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_real_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2249264Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2249528Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vsplit_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:22.2249780Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vstack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [100%] 2023-03-31T07:05:22.2249810Z 2023-03-31T07:05:22.2250196Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5d70d9a787657575.xml - 2023-03-31T07:05:22.2250344Z ==== 853 passed, 627 skipped, 5 deselected, 23 xfailed in 224.10s (0:03:44) ==== 2023-03-31T07:05:22.2250530Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:05:22.2250538Z 2023-03-31T07:05:22.2250903Z ##[endgroup] 2023-03-31T07:05:22.2251237Z FINISHED PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_y2p_cyv0.log) 2023-03-31T07:05:22.2251244Z 2023-03-31T07:05:47.5418398Z 2023-03-31T07:05:47.5418797Z Expand the folded group to see the log file of test_ops_fwd_gradients 2023-03-31T07:05:47.5421673Z ##[group]PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_32ydsk_x.log) 2023-03-31T07:05:47.5423388Z Test results will be stored in test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-419d8c8bac08ffa6.xml 2023-03-31T07:05:47.5423952Z ============================= test session starts ============================== 2023-03-31T07:05:47.5427037Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:05:47.5427569Z cachedir: .pytest_cache 2023-03-31T07:05:47.5428030Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:05:47.5428466Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:05:47.5429008Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:05:47.5429447Z collecting ... collected 3021 items / 7 deselected / 3014 selected 2023-03-31T07:05:47.5632069Z Running 1506 items in this shard: test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_H_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_T_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___getitem___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmatmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmod___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rpow___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rpow___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rsub___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__native_batch_norm_legit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__segment_reduce_lengths_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__softmax_backward_data_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__upsample_bilinear2d_aa_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_abs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcdiv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_decomposed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_all_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_all_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_allclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_angle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_angle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_any_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_any_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_partial_views_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bfloat16_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bool_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bucketize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_byte_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_byte_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cartesian_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cauchy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdouble_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ceil_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cfloat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chalf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chalf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_char_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_inverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chunk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clone_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clone_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_contiguous_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_corrcoef_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_corrcoef_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cov_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cross_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cummin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagflat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagflat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diff_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_digamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_floor_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_no_rounding_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_trunc_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_double_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_permuted_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erfc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expm1_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eye_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eye_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fliplr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flipud_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_floor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_frac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_frexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gather_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gather_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geqrf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geqrf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_grid_sampler_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hypot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_igamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_inner_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_int_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isfinite_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isnan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isneginf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isreal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isreal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_4inputs_with_extra_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lerp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cond_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cond_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eig_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_householder_product_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_grad_oriented_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_subgradients_at_zero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_subgradients_at_zero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_hermitian_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_slogdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_triangular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svdvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svdvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorsolve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorsolve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vander_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vecdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vecdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vector_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logcumsumexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_and_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_not_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_or_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_xor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_long_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mH_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mT_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mT_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_normalize_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matrix_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_maximum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_list_of_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_list_of_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_variadic_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_minimum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_movedim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_msort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_multinomial_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_5_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nan_to_num_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nansum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ne_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_binary_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_binary_cross_entropy_with_logits_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cosine_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cosine_similarity_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_elu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_with_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_without_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_fractional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_gelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardswish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardtanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_area_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_bicubic_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_linear_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_local_response_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_logsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool1d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_mish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_mse_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multilabel_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_circular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_constant_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_replicate_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pairwise_distance_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_shuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_unshuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_unshuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_prelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_silu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_smooth_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softplus_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softsign_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softsign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_tanhshrink_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_tanhshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_threshold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_with_distance_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_upsample_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_fro_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_inf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_inf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_in_place_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_number_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_outer_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pinverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polar_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_positive_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rad2deg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randint_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randint_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ravel_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ravel_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_real_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reciprocal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_renorm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_interleave_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize_as__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize_as__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_roll_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_roll_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rot90_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rot90_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scalar_tensor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_searchsorted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sgn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sgn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_short_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sigmoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_bartlett_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_blackman_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_general_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_hamming_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_hann_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_kaiser_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_mm_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_sampled_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_sampled_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_airy_ai_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_j1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_y1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_hermite_polynomial_h_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_hermite_polynomial_he_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_polygamma_special_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_scaled_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_spherical_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_list_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_with_sizes_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_with_sizes_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_square_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_square_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_multiple_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_multiple_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_to_size_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_along_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensor_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensor_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensordot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensordot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tile_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_sparse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_transpose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_transpose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapz_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tril_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_true_divide_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trunc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unbind_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_uniform_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsafe_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsqueeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_where_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_xlogy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_H_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_T_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___radd___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___radd___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rdiv___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmod___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rpow___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rsub___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rsub___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__native_batch_norm_legit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__segment_reduce_lengths_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__segment_reduce_offsets_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_abs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_abs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcdiv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_decomposed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_allclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_aminmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_angle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_angle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_arange_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_baddbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bfloat16_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_block_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_byte_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cartesian_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdouble_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ceil_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chalf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_char_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_inverse_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chunk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clone_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_column_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_combinations_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_combinations_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_constant_pad_nd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_contiguous_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_corrcoef_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cos_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cosh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cov_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cummax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cummin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumulative_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_deg2rad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagflat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_floor_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_no_rounding_mode_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_double_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_einsum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eq_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_equal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expm1_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fliplr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flipud_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_frac_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_frexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ge_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gradient_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_half_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_half_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histogram_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hypot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_inner_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_int_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_int_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isfinite_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isnan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isnan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isreal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_2inputs_2outputs_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_2inputs_2outputs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_return_by_ref_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kron_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kron_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kthvalue_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ldexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_le_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lerp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lgamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cond_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cond_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_singular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvalsh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvalsh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_multi_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_hermitian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_singular_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_triangular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svdvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorsolve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vander_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vander_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vector_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vector_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log10_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log10_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log1p_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logcumsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_and_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_and_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_not_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_or_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_or_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_xor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logspace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_long_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_unpack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_unpack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mH_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mT_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumprod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_pool2d_with_indices_backward_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_maximum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_list_of_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_list_of_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_variadic_tensors_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_minimum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_movedim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_movedim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_msort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_multinomial_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_5_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nan_to_num_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanquantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_layer_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_strided_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_full_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nextafter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_binary_cross_entropy_with_logits_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_celu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cosine_similarity_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cross_entropy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_ctc_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_elu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_embedding_bag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_fractional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_gaussian_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_gelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardtanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hinge_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_instance_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_area_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_nearest_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_trilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_l1_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_leaky_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_linear_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_logsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool3d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_mish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multi_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multilabel_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multilabel_soft_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_circular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_constant_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_replicate_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_replicate_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pairwise_distance_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_shuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_poisson_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_relu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_selu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_silu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_smooth_l1_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softplus_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_tanhshrink_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_threshold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_fro_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_nuc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_in_place_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_in_place_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_number_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ormqr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pca_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_permute_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pinverse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_4_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_positive_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pow_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_qr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rand_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randint_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_real_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reciprocal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reciprocal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_remainder_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_renorm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_renorm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_interleave_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize_as__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize_as__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_roll_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_neg_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scalar_tensor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scalar_tensor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sgn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_short_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_bartlett_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_blackman_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_hann_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_kaiser_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sort_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_airy_ai_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_j1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_entr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_erfcx_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_hermite_polynomial_h_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_hermite_polynomial_he_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i1e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_laguerre_polynomial_l_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_log_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_ndtri_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_polygamma_special_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_scaled_modified_bessel_k1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_w_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_spherical_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_zeta_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_list_args_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_square_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_square_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sub_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_to_size_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_t_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_along_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tan_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tanh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensor_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensordot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tile_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_topk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_transpose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_transpose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapz_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triangular_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tril_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_true_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unbind_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unbind_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unflatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unique_consecutive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsafe_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsqueeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsqueeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_where_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_H_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_T_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___getitem___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___radd___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___radd___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rdiv___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rdiv___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmatmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmul___cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmul___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rpow___cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__segment_reduce_lengths_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__softmax_backward_data_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__upsample_bilinear2d_aa_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_abs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acos_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_add_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addbmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcmul_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcmul_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_decomposed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_decomposed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_all_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_allclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_allclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_aminmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_angle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_angle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_arange_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argwhere_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argwhere_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_partial_views_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asin_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asinh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_2d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_baddbmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bernoulli_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bfloat16_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_block_diag_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bmm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bool_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_to_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_byte_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cartesian_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdouble_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cfloat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cfloat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chunk_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chunk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_max_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_min_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clone_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_column_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_combinations_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_physical_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_constant_pad_nd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_copysign_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cosh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_count_nonzero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cov_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cummin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumprod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumulative_trapezoid_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumulative_trapezoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_deg2rad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_embed_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_embed_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagflat_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_copy_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diff_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_digamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dist_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_floor_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_no_rounding_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_trunc_rounding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_double_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dstack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_einsum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_permuted_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_permuted_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_equal_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_equal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erfinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_as_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expm1_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftshift_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftn_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfft2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfftn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flatten_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flip_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flip_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fliplr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fliplr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flipud_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_power_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_power_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gather_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ge_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geometric_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geqrf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geqrf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histogramdd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hsplit_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_igamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_fill_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_select_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_inner_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_int_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_int_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isclose_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isclose_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isnan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isneginf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isposinf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_istft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_2inputs_2outputs_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kron_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kron_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kthvalue_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ldexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ldexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lerp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lerp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lgamma_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cond_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cross_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_diagonal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eig_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigh_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_householder_product_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_householder_product_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_multi_dot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_slogdet_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_slogdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_triangular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svd_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svdvals_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svdvals_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorinv_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorinv_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorsolve_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vecdot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vecdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vector_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log1p_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log2_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log2_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logcumsumexp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logcumsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logdet_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_and_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_not_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_xor_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logspace_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logsumexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_long_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_solve_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mH_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mT_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_amin_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_argmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_fill_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_logaddexp_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_normalize_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_softmax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_sum_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matrix_exp_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_maximum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_median_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_variadic_tensors_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_binary_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_reduction_no_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_reduction_with_dim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_minimum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mode_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_movedim_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_movedim_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nan_to_num_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanmedian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanquantile_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ne_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_neg_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_full_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_alpha_dropout_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_batch_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose1d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose3d_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cosine_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_ctc_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_embedding_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_with_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_fractional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_fractional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_gaussian_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_gelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_grid_sample_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_group_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardswish_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hinge_embedding_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_area_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_bicubic_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_linear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_linear_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_logsigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_margin_ranking_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool1d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool2d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool2d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool3d_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool3d_grad_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multilabel_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_circular_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_constant_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_constant_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_reflect_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_reflect_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pdist_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_shuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_poisson_nll_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_prelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_relu6_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_rrelu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_scaled_dot_product_attention_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_selu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softsign_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_tanhshrink_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_tanhshrink_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_threshold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_unfold_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_upsample_bilinear_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nonzero_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_fro_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_inf_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_inf_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_nuc_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_nuc_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_in_place_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_number_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ormqr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_outer_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_permute_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_permute_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_4_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_positive_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_positive_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_prod_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_put_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_put_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_qr_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rad2deg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rand_like_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ravel_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_real_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_renorm_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_renorm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_as_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize_as__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_conj_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_neg_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_roll_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_neg_3_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsub_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_add_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_amax_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_prod_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_sum_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sgn_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_short_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_short_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sigmoid_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_bartlett_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_exponential_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_gaussian_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_general_cosine_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_kaiser_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signbit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_scatter_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_mm_reduce_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_sampled_addmm_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_entr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_hermite_polynomial_h_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i0e_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i1_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_log_ndtr_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_i0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_k0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_polygamma_special_polygamma_n_0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_t_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_u_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_v_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_spherical_bessel_j0_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_zeta_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_list_args_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_with_sizes_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_with_sizes_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sqrt_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sqrt_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_square_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_square_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_multiple_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stack_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_unbiased_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stft_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stft_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_to_size_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_lowrank_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tan_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tanh_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensor_split_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensor_split_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensordot_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tile_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_sparse_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_topk_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trace_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapz_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tril_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tril_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triu_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triu_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_true_divide_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_true_divide_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unflatten_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_uniform_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unique_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsqueeze_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsqueeze_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_unbiased_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vdot_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_complex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_copy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vsplit_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vstack_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_where_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_where_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_xlogy_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zero__cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zero__cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_like_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_like_cpu_float64 2023-03-31T07:05:47.5809384Z 2023-03-31T07:05:47.5809642Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_H_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5810091Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_T_cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:47.5810514Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___getitem___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5810924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmatmul___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5811580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmod___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5811987Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rmul___cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:47.5812410Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rpow___cpu_complex128 SKIPPED (Skipped!) [ 0%] 2023-03-31T07:05:47.5812831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rpow___cpu_float64 SKIPPED (Skipped!) [ 0%] 2023-03-31T07:05:47.5813248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad___rsub___cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5813670Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__native_batch_norm_legit_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5814098Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__segment_reduce_lengths_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5814597Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__softmax_backward_data_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5815022Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad__upsample_bilinear2d_aa_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5815436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_abs_cpu_complex128 PASSED [ 0%] 2023-03-31T07:05:47.5815818Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_acosh_cpu_float64 PASSED [ 0%] 2023-03-31T07:05:47.5816220Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcdiv_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5816627Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addcmul_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5817042Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmm_decomposed_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5817445Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmv_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:47.5817844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addmv_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5818242Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_addr_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:47.5818939Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_all_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5819559Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_all_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5820259Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_allclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5820756Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_amax_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5821152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_angle_cpu_complex128 PASSED [ 1%] 2023-03-31T07:05:47.5821537Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_angle_cpu_float64 PASSED [ 1%] 2023-03-31T07:05:47.5822115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_any_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5822825Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_any_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5823466Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 1%] 2023-03-31T07:05:47.5824084Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_argmin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 2%] 2023-03-31T07:05:47.5824646Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_cpu_complex128 SKIPPED (Numerous errors) [ 2%] 2023-03-31T07:05:47.5825095Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_cpu_float64 SKIPPED (Numerous errors) [ 2%] 2023-03-31T07:05:47.5825528Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_partial_views_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:47.5825994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_as_strided_scatter_cpu_complex128 SKIPPED (Passes on complex128 and float64 only) [ 2%] 2023-03-31T07:05:47.5826449Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asin_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:47.5826853Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_asinh_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:47.5827287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan2_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:47.5827670Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atan_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:47.5828064Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atanh_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:47.5828467Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_1d_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:47.5828873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_2d_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:47.5829266Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_atleast_2d_cpu_float64 PASSED [ 2%] 2023-03-31T07:05:47.5829667Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bfloat16_cpu_complex128 XFAIL [ 2%] 2023-03-31T07:05:47.5830065Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bmm_cpu_complex128 PASSED [ 2%] 2023-03-31T07:05:47.5830447Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bmm_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:47.5831020Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bool_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5831479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_broadcast_to_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:47.5832069Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_bucketize_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5832798Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_byte_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5833458Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_byte_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5833924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cartesian_prod_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:47.5834569Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cat_cpu_complex128 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 3%] 2023-03-31T07:05:47.5835234Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cauchy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5835680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cdouble_cpu_complex128 PASSED [ 3%] 2023-03-31T07:05:47.5836083Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ceil_cpu_float64 PASSED [ 3%] 2023-03-31T07:05:47.5836481Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cfloat_cpu_float64 XFAIL [ 3%] 2023-03-31T07:05:47.5836873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chalf_cpu_complex128 XFAIL [ 3%] 2023-03-31T07:05:47.5837283Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chalf_cpu_float64 XFAIL [ 3%] 2023-03-31T07:05:47.5837853Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_char_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 3%] 2023-03-31T07:05:47.5838314Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_inverse_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5838728Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cholesky_solve_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5839119Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_chunk_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5839518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clone_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5839918Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_clone_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5840348Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5840732Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_conj_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5841133Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_contiguous_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5841543Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_corrcoef_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5841943Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_corrcoef_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5842330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cos_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5842724Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cosh_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5843308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cosh_cpu_float64 PASSED [ 4%] 2023-03-31T07:05:47.5843705Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cov_cpu_float64 SKIPPED (Barely fails) [ 4%] 2023-03-31T07:05:47.5844115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cross_cpu_complex128 PASSED [ 4%] 2023-03-31T07:05:47.5844516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cummin_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5844915Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumprod_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:47.5845309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_cumsum_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:47.5845781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagflat_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:47.5846229Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagflat_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5846645Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_copy_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:47.5847047Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_copy_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5847457Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_cpu_complex128 PASSED [ 5%] 2023-03-31T07:05:47.5847856Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5848265Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diagonal_scatter_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5848658Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_diff_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5849053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_digamma_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5849464Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_floor_rounding_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5849919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_no_rounding_mode_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5850344Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_div_trunc_rounding_cpu_float64 PASSED [ 5%] 2023-03-31T07:05:47.5850752Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dot_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:47.5851148Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_double_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:47.5851536Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dsplit_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:47.5851936Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_dsplit_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:47.5852533Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:47.5853218Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:47.5853860Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_empty_permuted_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:47.5854500Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eq_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:47.5855118Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eq_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 6%] 2023-03-31T07:05:47.5855561Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_erfc_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:47.5855941Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:47.5856340Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exp_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:47.5856742Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_as_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:47.5857156Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_as_cpu_float64 PASSED [ 6%] 2023-03-31T07:05:47.5857546Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_cpu_complex128 PASSED [ 6%] 2023-03-31T07:05:47.5857944Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expand_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5858377Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_expm1_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:47.5858996Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 7%] 2023-03-31T07:05:47.5859625Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eye_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 7%] 2023-03-31T07:05:47.5860244Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_eye_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 7%] 2023-03-31T07:05:47.5860693Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_fft_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5861099Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_hfft2_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:47.5861499Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftn_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5861904Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ifftshift_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5862307Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfft_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5862862Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_ihfftn_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5863266Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft2_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:47.5863670Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_irfft_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5864064Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fft_rfft2_cpu_float64 PASSED [ 7%] 2023-03-31T07:05:47.5864465Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fill_cpu_complex128 PASSED [ 7%] 2023-03-31T07:05:47.5864850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fill_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5865248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flatten_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:47.5865689Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flatten_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5866074Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fliplr_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5866468Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_flipud_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5866863Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_cpu_complex128 XFAIL [ 8%] 2023-03-31T07:05:47.5867266Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_float_power_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:47.5867656Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_floor_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5868050Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_fmod_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5868444Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_frac_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5868834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_frexp_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5869414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 8%] 2023-03-31T07:05:47.5870056Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_full_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 8%] 2023-03-31T07:05:47.5870517Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gather_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:47.5870977Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_gather_cpu_float64 PASSED [ 8%] 2023-03-31T07:05:47.5871583Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geqrf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:47.5872212Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_geqrf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:47.5872677Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_grid_sampler_2d_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5873261Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_histc_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:47.5873704Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hsplit_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:47.5874105Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hstack_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:47.5874508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_hypot_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5874901Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_i0_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5875490Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_igamma_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:47.5875946Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_add_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5876351Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_fill_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5876743Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_put_cpu_complex128 PASSED [ 9%] 2023-03-31T07:05:47.5877150Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_index_put_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5877551Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_inner_cpu_float64 PASSED [ 9%] 2023-03-31T07:05:47.5878124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_int_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 9%] 2023-03-31T07:05:47.5878780Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5879411Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isfinite_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5880033Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5880654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isnan_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5881280Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isneginf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5881904Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isreal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5882538Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_isreal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 10%] 2023-03-31T07:05:47.5883206Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_jiterator_4inputs_with_extra_args_cpu_complex128 SKIPPED (Only runs on cuda) [ 10%] 2023-03-31T07:05:47.5883650Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lerp_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:47.5884047Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cond_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:47.5884518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cond_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:47.5884969Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cross_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:47.5885381Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_cpu_complex128 PASSED [ 10%] 2023-03-31T07:05:47.5885777Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_cpu_float64 PASSED [ 10%] 2023-03-31T07:05:47.5886191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_singular_cpu_complex128 XFAIL [ 10%] 2023-03-31T07:05:47.5886623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_det_singular_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:47.5887045Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_diagonal_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:47.5887441Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eig_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:47.5887852Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigh_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:47.5888265Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_eigvals_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:47.5888747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_householder_product_cpu_complex128 SKIPPED (Skipped! Flaky) [ 11%] 2023-03-31T07:05:47.5889194Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:47.5889608Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_inv_ex_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:47.5890221Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:47.5890881Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_factor_ex_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:47.5891580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_ldl_solve_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 11%] 2023-03-31T07:05:47.5892067Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_cpu_complex128 SKIPPED (Skipped!) [ 11%] 2023-03-31T07:05:47.5892505Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_cpu_float64 SKIPPED (Skipped!) [ 11%] 2023-03-31T07:05:47.5892949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lstsq_grad_oriented_cpu_complex128 PASSED [ 11%] 2023-03-31T07:05:47.5893371Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_factor_cpu_float64 PASSED [ 11%] 2023-03-31T07:05:47.5893790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_lu_solve_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:47.5894214Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_norm_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:47.5894625Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_norm_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:47.5895054Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_power_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:47.5895482Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_power_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:47.5896100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 12%] 2023-03-31T07:05:47.5896816Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_hermitian_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 12%] 2023-03-31T07:05:47.5897518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_matrix_rank_hermitian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 12%] 2023-03-31T07:05:47.5898001Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:47.5898436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_subgradients_at_zero_cpu_complex128 XFAIL [ 12%] 2023-03-31T07:05:47.5898890Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_norm_subgradients_at_zero_cpu_float64 XFAIL [ 12%] 2023-03-31T07:05:47.5899310Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:47.5899719Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_cpu_float64 PASSED [ 12%] 2023-03-31T07:05:47.5900147Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_pinv_hermitian_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:47.5900572Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_qr_cpu_complex128 PASSED [ 12%] 2023-03-31T07:05:47.5900994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_qr_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5901404Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_slogdet_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5901822Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5902220Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5902726Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_ex_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5903167Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_solve_triangular_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5903592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svd_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5904034Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svdvals_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5904460Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_svdvals_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5904880Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorinv_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5905311Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorsolve_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5905729Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_tensorsolve_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5906146Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vander_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5906560Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vecdot_cpu_complex128 PASSED [ 13%] 2023-03-31T07:05:47.5906976Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vecdot_cpu_float64 PASSED [ 13%] 2023-03-31T07:05:47.5907387Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_vector_norm_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5907795Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log2_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5908191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5908780Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_normal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5909270Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:47.5909719Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_log_softmax_with_dtype_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5910147Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logaddexp_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5910555Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logcumsumexp_cpu_complex128 PASSED [ 14%] 2023-03-31T07:05:47.5911145Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_and_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5911796Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_not_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5912438Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_or_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5913081Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logical_xor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5913778Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5914418Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 14%] 2023-03-31T07:05:47.5914873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_logsumexp_cpu_float64 PASSED [ 14%] 2023-03-31T07:05:47.5915449Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_long_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 15%] 2023-03-31T07:05:47.5915885Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5916284Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_lu_solve_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5916710Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mH_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5917106Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mT_cpu_complex128 PASSED [ 15%] 2023-03-31T07:05:47.5917481Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mT_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5917877Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_amax_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5918280Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumprod_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5918673Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_cumsum_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5919087Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_fill_cpu_complex128 PASSED [ 15%] 2023-03-31T07:05:47.5919501Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_fill_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5919915Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_log_softmax_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5920313Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_mean_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5920720Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_median_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5921124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_norm_cpu_float64 PASSED [ 15%] 2023-03-31T07:05:47.5921539Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_normalize_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5921978Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_normalize_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5922415Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_prod_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5922832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_scatter_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5923409Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_scatter_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5923812Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_select_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5924222Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_softmax_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5924632Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_std_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5925043Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_masked_var_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5925436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matmul_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5925839Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_matrix_exp_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5926311Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_max_binary_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5926701Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_maximum_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5927103Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mean_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5927525Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_list_of_tensors_cpu_complex128 PASSED [ 16%] 2023-03-31T07:05:47.5927962Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_list_of_tensors_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:47.5928391Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_meshgrid_variadic_tensors_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5928807Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_minimum_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5929294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_movedim_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:47.5929691Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_msort_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5930275Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_multinomial_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 17%] 2023-03-31T07:05:47.5930751Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_1_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5931174Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_3_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5931594Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_mvlgamma_mvlgamma_p_5_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5931987Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nan_to_num_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5932388Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nansum_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5932972Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_copy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 17%] 2023-03-31T07:05:47.5933437Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_narrow_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:47.5934683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_native_batch_norm_cpu_float64 PASSED [ 17%] 2023-03-31T07:05:47.5935317Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ne_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 17%] 2023-03-31T07:05:47.5935803Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_neg_cpu_complex128 PASSED [ 17%] 2023-03-31T07:05:47.5936197Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_neg_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5936765Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5937404Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5938040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5938665Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5939293Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5939964Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_new_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 18%] 2023-03-31T07:05:47.5940446Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool1d_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5940897Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_avg_pool3d_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5941332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_adaptive_max_pool2d_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5941784Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_batch_norm_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5942232Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_binary_cross_entropy_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5942836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_binary_cross_entropy_with_logits_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5943289Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv2d_cpu_complex128 PASSED [ 18%] 2023-03-31T07:05:47.5943717Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv2d_cpu_float64 PASSED [ 18%] 2023-03-31T07:05:47.5944155Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose1d_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5944601Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose2d_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5945040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose3d_cpu_complex128 PASSED [ 19%] 2023-03-31T07:05:47.5945488Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_conv_transpose3d_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5945938Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cosine_embedding_loss_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5946387Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cosine_similarity_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5946814Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_cross_entropy_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5947245Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_elu_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5947731Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_with_train_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5948241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_feature_alpha_dropout_without_train_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5948700Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_fractional_max_pool2d_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5949136Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_gelu_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5949567Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardsigmoid_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5950004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardswish_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5950425Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_hardtanh_cpu_float64 PASSED [ 19%] 2023-03-31T07:05:47.5950867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_area_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5951318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_bicubic_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5951807Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_bilinear_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5952250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_interpolate_nearest_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5952692Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_l1_loss_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5953120Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_linear_cpu_complex128 PASSED [ 20%] 2023-03-31T07:05:47.5953552Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_local_response_norm_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5953983Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_logsigmoid_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5954420Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool1d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5954878Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool2d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5955313Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_pool3d_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5955735Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_max_unpool1d_grad_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5956163Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_mish_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5956589Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_mse_loss_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5957037Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_multilabel_margin_loss_cpu_float64 PASSED [ 20%] 2023-03-31T07:05:47.5957465Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_nll_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5957898Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_normalize_cpu_complex128 PASSED [ 21%] 2023-03-31T07:05:47.5958333Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_circular_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5958763Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_constant_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5959187Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pad_replicate_cpu_complex128 PASSED [ 21%] 2023-03-31T07:05:47.5959692Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pairwise_distance_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5960167Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_shuffle_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5960614Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_unshuffle_cpu_complex128 PASSED [ 21%] 2023-03-31T07:05:47.5961053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_pixel_unshuffle_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5961484Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_prelu_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5961905Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_silu_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5962337Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_smooth_l1_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5962768Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_soft_margin_loss_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5963408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_cpu_float64 PASSED [ 21%] 2023-03-31T07:05:47.5963903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_with_dtype_cpu_complex128 PASSED [ 21%] 2023-03-31T07:05:47.5964363Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softmin_with_dtype_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5964790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softplus_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5965223Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softsign_cpu_complex128 PASSED [ 22%] 2023-03-31T07:05:47.5965653Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_softsign_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5966088Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_tanhshrink_cpu_complex128 PASSED [ 22%] 2023-03-31T07:05:47.5966513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_tanhshrink_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5966988Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_threshold_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5967427Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_loss_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5967892Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_triplet_margin_with_distance_loss_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5968335Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_unfold_cpu_complex128 PASSED [ 22%] 2023-03-31T07:05:47.5968775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nn_functional_upsample_bilinear_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5969422Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 22%] 2023-03-31T07:05:47.5970064Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 22%] 2023-03-31T07:05:47.5970504Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_cpu_complex128 PASSED [ 22%] 2023-03-31T07:05:47.5970901Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_fro_cpu_float64 PASSED [ 22%] 2023-03-31T07:05:47.5971309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_inf_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:47.5971714Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_norm_inf_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:47.5972336Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_in_place_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 23%] 2023-03-31T07:05:47.5972883Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_normal_number_mean_cpu_float64 SKIPPED (Gradients are incorrect!) [ 23%] 2023-03-31T07:05:47.5973506Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 23%] 2023-03-31T07:05:47.5974130Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 23%] 2023-03-31T07:05:47.5974567Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_outer_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:47.5974965Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_pinverse_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:47.5975364Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polar_cpu_float64 XFAIL [ 23%] 2023-03-31T07:05:47.5975795Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_2_cpu_float64 SKIPPED (Skipped!) [ 23%] 2023-03-31T07:05:47.5976271Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_polygamma_polygamma_n_3_cpu_float64 SKIPPED (Skipped!) [ 23%] 2023-03-31T07:05:47.5976706Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_positive_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:47.5977114Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_prod_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:47.5977500Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_prod_cpu_float64 PASSED [ 23%] 2023-03-31T07:05:47.5977892Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_put_cpu_complex128 PASSED [ 23%] 2023-03-31T07:05:47.5978291Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_qr_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:47.5978683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_qr_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5979094Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rad2deg_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5979677Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randint_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:47.5980332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randint_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:47.5980970Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:47.5981598Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:47.5982241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_randn_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 24%] 2023-03-31T07:05:47.5982791Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ravel_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:47.5983195Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_ravel_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5983580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_real_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:47.5983973Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_real_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5984375Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reciprocal_cpu_complex128 PASSED [ 24%] 2023-03-31T07:05:47.5984818Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_renorm_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5985231Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_cpu_float64 PASSED [ 24%] 2023-03-31T07:05:47.5985642Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_repeat_interleave_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:47.5986064Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_reshape_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5986650Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 25%] 2023-03-31T07:05:47.5987279Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 25%] 2023-03-31T07:05:47.5987914Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize_as__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 25%] 2023-03-31T07:05:47.5988551Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resize_as__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 25%] 2023-03-31T07:05:47.5989022Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_conj_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:47.5989459Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_conj_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5989871Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_neg_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:47.5990281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_resolve_neg_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5990683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_roll_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:47.5991066Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_roll_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5991461Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rot90_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:47.5991855Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rot90_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5992273Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_cpu_float64 PASSED [ 25%] 2023-03-31T07:05:47.5992669Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_round_decimals_0_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5993079Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_rsub_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:47.5993674Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scalar_tensor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:47.5994143Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:47.5994531Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5994944Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_amax_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5995368Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_scatter_reduce_sum_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5995959Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_searchsorted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:47.5996425Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:47.5996834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_select_scatter_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5997233Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sgn_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:47.5997642Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sgn_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5998248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_short_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 26%] 2023-03-31T07:05:47.5998709Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sigmoid_cpu_complex128 PASSED [ 26%] 2023-03-31T07:05:47.5999113Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sign_cpu_float64 PASSED [ 26%] 2023-03-31T07:05:47.5999699Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_bartlett_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6000363Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_blackman_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6001038Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6001732Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_general_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6002463Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_hamming_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6003281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_hann_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6003949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_signal_windows_kaiser_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 27%] 2023-03-31T07:05:47.6004416Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sin_cpu_complex128 PASSED [ 27%] 2023-03-31T07:05:47.6004811Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sin_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:47.6005251Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinc_cpu_complex128 PASSED [ 27%] 2023-03-31T07:05:47.6005648Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sinh_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:47.6006042Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_cpu_complex128 PASSED [ 27%] 2023-03-31T07:05:47.6006448Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_slice_scatter_cpu_float64 PASSED [ 27%] 2023-03-31T07:05:47.6006864Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_mm_reduce_cpu_float64 SKIPPED (Skipped!) [ 27%] 2023-03-31T07:05:47.6007320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_sampled_addmm_cpu_complex128 SKIPPED (Skipped!) [ 27%] 2023-03-31T07:05:47.6007781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sparse_sampled_addmm_cpu_float64 SKIPPED (Skipped!) [ 28%] 2023-03-31T07:05:47.6008396Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_airy_ai_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6009043Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_j1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6009693Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_bessel_y1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6010374Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6011109Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_hermite_polynomial_h_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6011821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_hermite_polynomial_he_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6027330Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_i1_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:47.6028193Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_i0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6028895Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6029383Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_ndtr_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:47.6029836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_polygamma_special_polygamma_n_0_cpu_float64 PASSED [ 28%] 2023-03-31T07:05:47.6030732Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_scaled_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 28%] 2023-03-31T07:05:47.6031470Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 28%] 2023-03-31T07:05:47.6032184Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_shifted_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 28%] 2023-03-31T07:05:47.6032889Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_special_spherical_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 29%] 2023-03-31T07:05:47.6033377Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6033846Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_list_args_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6034260Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_with_sizes_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6034685Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_split_with_sizes_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6035100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_square_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6035503Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_square_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6035892Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6036312Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_multiple_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6036740Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_squeeze_multiple_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6037154Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stack_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6037540Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_stack_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6037928Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6038328Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_mean_cpu_complex128 PASSED [ 29%] 2023-03-31T07:05:47.6038727Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_std_unbiased_cpu_float64 PASSED [ 29%] 2023-03-31T07:05:47.6039222Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sub_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6039660Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sub_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6040054Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6040447Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_sum_to_size_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6040844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_svd_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6041249Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_along_dim_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6041654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6042035Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_take_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6042432Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tan_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6042826Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tan_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6043516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tanh_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6043949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tanh_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6044352Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensor_split_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6044765Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensor_split_cpu_float64 PASSED [ 30%] 2023-03-31T07:05:47.6045161Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensordot_cpu_complex128 PASSED [ 30%] 2023-03-31T07:05:47.6045568Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tensordot_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6045969Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tile_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6046370Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_to_sparse_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6046801Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trace_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6047202Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_transpose_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6047606Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_transpose_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6048002Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trapz_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6048390Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_tril_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6048790Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triu_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6049188Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_triu_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6049594Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_true_divide_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6049985Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_trunc_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6050372Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unbind_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6050775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_copy_cpu_complex128 PASSED [ 31%] 2023-03-31T07:05:47.6051172Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unfold_copy_cpu_float64 PASSED [ 31%] 2023-03-31T07:05:47.6051819Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_uniform_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:47.6052320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsafe_split_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6052736Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_unsqueeze_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:47.6053127Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:47.6053515Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6053906Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_var_mean_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6054301Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6054692Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_view_as_real_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:47.6055102Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vstack_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:47.6055500Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_vstack_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6055899Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_where_cpu_complex128 PASSED [ 32%] 2023-03-31T07:05:47.6056339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_xlogy_cpu_float64 PASSED [ 32%] 2023-03-31T07:05:47.6056924Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:47.6057554Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:47.6058191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_zeros_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 32%] 2023-03-31T07:05:47.6058633Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_H_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6059053Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_T_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6059457Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___radd___cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6059854Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___radd___cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6060236Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rdiv___cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6060623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmod___cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6061016Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rmul___cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6061419Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rpow___cpu_complex128 SKIPPED (Skipped!) [ 33%] 2023-03-31T07:05:47.6061834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rsub___cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6062237Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD___rsub___cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6062779Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__native_batch_norm_legit_cpu_float64 SKIPPED (Skipped!) [ 33%] 2023-03-31T07:05:47.6063219Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__segment_reduce_lengths_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6063647Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD__segment_reduce_offsets_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6064058Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_abs_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6064486Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_abs_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:47.6064891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acos_cpu_complex128 PASSED [ 33%] 2023-03-31T07:05:47.6065282Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_acosh_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6065669Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_add_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6066058Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addbmm_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6066441Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcdiv_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:47.6066836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcmul_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:47.6067237Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addcmul_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6067632Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmm_decomposed_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6068046Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmv_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:47.6068440Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addmv_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6068861Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addr_cpu_complex128 PASSED [ 34%] 2023-03-31T07:05:47.6069235Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_addr_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6069815Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_allclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 34%] 2023-03-31T07:05:47.6070263Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_amax_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6070653Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_amin_cpu_float64 PASSED [ 34%] 2023-03-31T07:05:47.6071213Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_aminmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 34%] 2023-03-31T07:05:47.6071699Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_angle_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:47.6072092Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_angle_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6072663Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_arange_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:47.6073279Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 35%] 2023-03-31T07:05:47.6073753Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_cpu_complex128 SKIPPED (Numerous errors) [ 35%] 2023-03-31T07:05:47.6074202Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_as_strided_cpu_float64 SKIPPED (Numerous errors) [ 35%] 2023-03-31T07:05:47.6074620Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asin_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:47.6075004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_asinh_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6075387Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan2_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6075769Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atan_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6076157Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atanh_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:47.6076531Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atanh_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6076960Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_1d_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:47.6077394Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_2d_cpu_float64 PASSED [ 35%] 2023-03-31T07:05:47.6077785Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_atleast_3d_cpu_complex128 PASSED [ 35%] 2023-03-31T07:05:47.6078188Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_baddbmm_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:47.6078577Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bfloat16_cpu_float64 XFAIL [ 36%] 2023-03-31T07:05:47.6078967Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_block_diag_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:47.6079347Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_bmm_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:47.6079754Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_tensors_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:47.6080176Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_to_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:47.6080587Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_broadcast_to_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:47.6081186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_byte_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:47.6081644Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cartesian_prod_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:47.6082273Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cat_cpu_complex128 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 36%] 2023-03-31T07:05:47.6082974Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cat_cpu_float64 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 36%] 2023-03-31T07:05:47.6083633Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cdouble_cpu_complex128 PASSED [ 36%] 2023-03-31T07:05:47.6084031Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ceil_cpu_float64 PASSED [ 36%] 2023-03-31T07:05:47.6084476Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chalf_cpu_float64 XFAIL [ 36%] 2023-03-31T07:05:47.6085058Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_char_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 36%] 2023-03-31T07:05:47.6085498Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6085910Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_inverse_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6086332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cholesky_solve_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6086742Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_chunk_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6087120Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clamp_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6087518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_clone_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6087916Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_column_stack_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6088318Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_combinations_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6088730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_combinations_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6089128Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_complex_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6089522Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6089945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_conj_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6090384Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_constant_pad_nd_cpu_float64 PASSED [ 37%] 2023-03-31T07:05:47.6090797Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_contiguous_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6091203Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_corrcoef_cpu_complex128 PASSED [ 37%] 2023-03-31T07:05:47.6091580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cos_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6091962Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cosh_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6092343Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cov_cpu_float64 XFAIL [ 38%] 2023-03-31T07:05:47.6092725Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cross_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6093107Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cummax_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6093494Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cummin_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6093919Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumprod_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6094306Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumsum_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:47.6094739Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumsum_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6095272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_cumulative_trapezoid_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6095682Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_deg2rad_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6096074Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagflat_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6096489Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_copy_cpu_complex128 PASSED [ 38%] 2023-03-31T07:05:47.6096941Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6097346Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_diagonal_scatter_cpu_float64 PASSED [ 38%] 2023-03-31T07:05:47.6097748Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_floor_rounding_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6098168Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_div_no_rounding_mode_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:47.6098581Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dot_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:47.6098970Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dot_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6099345Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_double_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6099742Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dsplit_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:47.6100145Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dstack_cpu_complex128 PASSED [ 39%] 2023-03-31T07:05:47.6100530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_dstack_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6100917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_einsum_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6101506Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 39%] 2023-03-31T07:05:47.6102194Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 39%] 2023-03-31T07:05:47.6102982Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_empty_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 39%] 2023-03-31T07:05:47.6103621Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_eq_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 39%] 2023-03-31T07:05:47.6104249Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_equal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 39%] 2023-03-31T07:05:47.6104701Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_erf_cpu_float64 PASSED [ 39%] 2023-03-31T07:05:47.6105082Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expand_as_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6105483Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_expm1_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:47.6105885Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fft_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6106281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftn_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6106671Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_fftshift_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6107111Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft2_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6107513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfft_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:47.6107913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_hfftn_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6108303Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft2_cpu_complex128 PASSED [ 40%] 2023-03-31T07:05:47.6108703Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ifft_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6109104Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_ihfft2_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6109532Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfft_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6109917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_irfftn_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6110309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fft_rfft_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6110700Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flatten_cpu_float64 PASSED [ 40%] 2023-03-31T07:05:47.6111085Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fliplr_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6111470Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_flipud_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6111860Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_cpu_complex128 XFAIL [ 41%] 2023-03-31T07:05:47.6112248Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_cpu_float64 XFAIL [ 41%] 2023-03-31T07:05:47.6112631Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_float_power_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6113027Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmax_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6113408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_fmod_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6113788Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_frac_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6114160Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_frexp_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6114782Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:47.6115448Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:47.6116062Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ge_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 41%] 2023-03-31T07:05:47.6116499Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_gradient_cpu_float64 PASSED [ 41%] 2023-03-31T07:05:47.6116893Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_half_cpu_complex128 XFAIL [ 41%] 2023-03-31T07:05:47.6117274Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_half_cpu_float64 XFAIL [ 41%] 2023-03-31T07:05:47.6117840Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histc_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 42%] 2023-03-31T07:05:47.6118454Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_histogram_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 42%] 2023-03-31T07:05:47.6118912Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hsplit_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6119313Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hstack_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:47.6119747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hstack_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6120131Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_hypot_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6120512Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_i0_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6120907Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_add_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:47.6121303Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_add_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6121710Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_fill_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:47.6122149Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_fill_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6122547Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_put_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6122936Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_reduce_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6123526Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_select_cpu_complex128 PASSED [ 42%] 2023-03-31T07:05:47.6123937Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_index_select_cpu_float64 PASSED [ 42%] 2023-03-31T07:05:47.6124337Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_inner_cpu_complex128 PASSED [ 43%] 2023-03-31T07:05:47.6124906Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_int_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6125531Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_int_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6126158Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6126788Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isfinite_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6127393Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isin_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6128099Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6128762Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isnan_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6129384Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isnan_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6129993Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_isreal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 43%] 2023-03-31T07:05:47.6130485Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_2inputs_2outputs_cpu_complex128 SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:05:47.6130961Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_2inputs_2outputs_cpu_float64 SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:05:47.6131440Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_float64 SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:05:47.6131913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_cpu_complex128 SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:05:47.6132411Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_jiterator_binary_return_by_ref_cpu_complex128 SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:05:47.6132849Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kron_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:47.6133241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kron_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6133623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_kthvalue_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6134014Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ldexp_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6134586Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_le_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 44%] 2023-03-31T07:05:47.6135029Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lerp_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6135503Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lgamma_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6135905Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cond_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:47.6136319Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cond_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6136722Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cross_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6137119Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:47.6137530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6137945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_det_singular_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:47.6138370Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_diagonal_cpu_float64 PASSED [ 44%] 2023-03-31T07:05:47.6138776Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvals_cpu_complex128 PASSED [ 44%] 2023-03-31T07:05:47.6139186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvals_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:47.6139597Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvalsh_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:47.6140009Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_eigvalsh_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:47.6140440Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:47.6140873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_inv_ex_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:47.6141479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:47.6142143Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_factor_ex_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:47.6142872Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_ldl_solve_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:47.6143360Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lstsq_cpu_float64 SKIPPED (Skipped!) [ 45%] 2023-03-31T07:05:47.6143777Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:47.6144191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_factor_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:47.6144592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_solve_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:47.6145067Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_lu_solve_cpu_float64 PASSED [ 45%] 2023-03-31T07:05:47.6145491Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_power_cpu_complex128 PASSED [ 45%] 2023-03-31T07:05:47.6146117Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_matrix_rank_hermitian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 45%] 2023-03-31T07:05:47.6146584Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_multi_dot_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6147014Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_float64 XFAIL [ 46%] 2023-03-31T07:05:47.6147438Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6147887Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_hermitian_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6148363Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_pinv_singular_cpu_float64 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 46%] 2023-03-31T07:05:47.6148832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_qr_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6149238Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6149639Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_ex_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6150051Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_ex_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6150475Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_solve_triangular_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6150902Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svd_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6151298Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svd_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6151700Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_svdvals_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6152118Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorinv_cpu_complex128 PASSED [ 46%] 2023-03-31T07:05:47.6152538Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_tensorsolve_cpu_float64 PASSED [ 46%] 2023-03-31T07:05:47.6152975Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vander_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6153410Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vander_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6153828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vector_norm_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6154246Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_vector_norm_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6154637Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log10_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6155027Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log10_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6155421Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log1p_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6155820Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log2_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6156206Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6156792Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_normal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 47%] 2023-03-31T07:05:47.6157283Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6157706Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_log_softmax_with_dtype_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6158113Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logaddexp2_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6158516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logcumsumexp_cpu_float64 PASSED [ 47%] 2023-03-31T07:05:47.6158921Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logdet_cpu_complex128 PASSED [ 47%] 2023-03-31T07:05:47.6159308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logdet_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:47.6159897Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_and_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6160578Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_and_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6161212Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_not_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6161849Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_or_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6162468Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_or_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6163284Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logical_xor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6163743Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logit_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:47.6164320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_logspace_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6164935Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_long_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 48%] 2023-03-31T07:05:47.6165378Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:47.6165836Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_solve_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:47.6166294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_unpack_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:47.6166681Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_lu_unpack_cpu_float64 PASSED [ 48%] 2023-03-31T07:05:47.6167084Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mH_cpu_complex128 PASSED [ 48%] 2023-03-31T07:05:47.6167473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mT_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6167858Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_amax_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6168256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_amin_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6168850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 49%] 2023-03-31T07:05:47.6169327Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumprod_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6169736Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_cumsum_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6170193Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_fill_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6170602Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_fill_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6171009Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_log_softmax_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6171401Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_mean_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6171803Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_prod_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6172217Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_scatter_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6172629Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_select_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6173065Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_softmax_cpu_float64 PASSED [ 49%] 2023-03-31T07:05:47.6173474Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_masked_std_cpu_complex128 PASSED [ 49%] 2023-03-31T07:05:47.6173876Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_binary_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6174293Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_pool2d_with_indices_backward_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6174716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_max_reduction_no_dim_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6175126Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_maximum_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6175526Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mean_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:47.6175903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mean_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6176296Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_median_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6176706Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_list_of_tensors_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:47.6177141Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_list_of_tensors_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6177562Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_meshgrid_variadic_tensors_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:47.6177996Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_reduction_no_dim_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6178444Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_min_reduction_with_dim_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6178877Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_minimum_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6179252Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mm_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6179634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mode_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:47.6180031Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_movedim_cpu_complex128 PASSED [ 50%] 2023-03-31T07:05:47.6180427Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_movedim_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6180802Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_msort_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6181186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mul_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6181775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_multinomial_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 51%] 2023-03-31T07:05:47.6182243Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_1_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6182803Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_3_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6183235Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_mvlgamma_mvlgamma_p_5_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6183648Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nan_to_num_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6184039Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nanquantile_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6184640Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_copy_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 51%] 2023-03-31T07:05:47.6185284Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_copy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 51%] 2023-03-31T07:05:47.6185767Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_narrow_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6186183Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_native_layer_norm_cpu_float64 SKIPPED (Skipped!) [ 51%] 2023-03-31T07:05:47.6186580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_neg_cpu_float64 PASSED [ 51%] 2023-03-31T07:05:47.6187155Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 51%] 2023-03-31T07:05:47.6187808Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_empty_strided_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:47.6188453Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_full_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:47.6189080Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:47.6189703Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_new_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:47.6190326Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nextafter_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 52%] 2023-03-31T07:05:47.6190810Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_avg_pool1d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6191299Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool1d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6191773Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool2d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6192227Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_adaptive_max_pool3d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6192663Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_avg_pool2d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6193109Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_binary_cross_entropy_with_logits_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6193558Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_celu_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6193985Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv1d_cpu_complex128 PASSED [ 52%] 2023-03-31T07:05:47.6194408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv1d_cpu_float64 PASSED [ 52%] 2023-03-31T07:05:47.6194819Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv2d_cpu_complex128 PASSED [ 52%] 2023-03-31T07:05:47.6195276Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv2d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6195710Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose1d_cpu_complex128 PASSED [ 53%] 2023-03-31T07:05:47.6196159Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose1d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6196592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose3d_cpu_complex128 PASSED [ 53%] 2023-03-31T07:05:47.6197040Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_conv_transpose3d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6197483Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cosine_similarity_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6197952Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_cross_entropy_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6198370Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_ctc_loss_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6198783Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_elu_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6199206Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_embedding_bag_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6199672Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_complex128 PASSED [ 53%] 2023-03-31T07:05:47.6200151Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6200623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_fractional_max_pool3d_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6201075Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_gaussian_nll_loss_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6201504Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_gelu_cpu_float64 PASSED [ 53%] 2023-03-31T07:05:47.6201914Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardshrink_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6202342Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hardtanh_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6202813Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_hinge_embedding_loss_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6203463Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_instance_norm_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6203895Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_area_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6204340Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_linear_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6204801Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_nearest_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6205258Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_interpolate_trilinear_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6205688Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_l1_loss_cpu_complex128 PASSED [ 54%] 2023-03-31T07:05:47.6206120Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_l1_loss_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6206549Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_leaky_relu_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6206965Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_linear_cpu_complex128 PASSED [ 54%] 2023-03-31T07:05:47.6207448Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_logsigmoid_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6207877Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool1d_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6208298Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool2d_cpu_float64 PASSED [ 54%] 2023-03-31T07:05:47.6208707Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_pool3d_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6209157Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool1d_cpu_float64 SKIPPED (Skipped!) [ 55%] 2023-03-31T07:05:47.6209616Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool2d_cpu_float64 SKIPPED (Skipped!) [ 55%] 2023-03-31T07:05:47.6210104Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_max_unpool3d_grad_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6210519Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_mish_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6210951Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multi_margin_loss_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6211397Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multilabel_margin_loss_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6211859Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_multilabel_soft_margin_loss_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6212299Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_circular_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:47.6212744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_constant_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:47.6213186Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_replicate_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:47.6213628Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pad_replicate_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6214055Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pairwise_distance_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6214483Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pdist_cpu_float64 PASSED [ 55%] 2023-03-31T07:05:47.6214954Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_pixel_shuffle_cpu_complex128 PASSED [ 55%] 2023-03-31T07:05:47.6215428Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_poisson_nll_loss_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6215842Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_relu_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6216261Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_selu_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6216676Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_silu_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6217098Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_smooth_l1_loss_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6217530Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:47.6217981Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6218415Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_softplus_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6218878Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_tanhshrink_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:47.6219301Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_threshold_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6219726Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nn_functional_unfold_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:47.6220348Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 56%] 2023-03-31T07:05:47.6220805Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:47.6221195Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_fro_cpu_complex128 PASSED [ 56%] 2023-03-31T07:05:47.6221593Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_norm_nuc_cpu_float64 PASSED [ 56%] 2023-03-31T07:05:47.6222060Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_cpu_float64 SKIPPED (Gradients are incorrect!) [ 57%] 2023-03-31T07:05:47.6222764Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_in_place_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:47.6223408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_in_place_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:47.6223903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_normal_number_mean_cpu_float64 SKIPPED (Gradients are incorrect!) [ 57%] 2023-03-31T07:05:47.6224518Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ones_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 57%] 2023-03-31T07:05:47.6224982Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_ormqr_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:47.6225373Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pca_lowrank_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:47.6225774Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_permute_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:47.6226168Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pinverse_cpu_float64 PASSED [ 57%] 2023-03-31T07:05:47.6226595Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_2_cpu_float64 SKIPPED (Skipped!) [ 57%] 2023-03-31T07:05:47.6227081Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_3_cpu_float64 SKIPPED (Skipped!) [ 57%] 2023-03-31T07:05:47.6227556Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_polygamma_polygamma_n_4_cpu_float64 SKIPPED (Skipped!) [ 57%] 2023-03-31T07:05:47.6227986Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_positive_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:47.6228386Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_pow_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:47.6228768Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_put_cpu_complex128 PASSED [ 57%] 2023-03-31T07:05:47.6229156Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_put_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6229541Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_qr_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6230100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rand_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6230730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randint_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6231359Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6232075Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6232716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6233335Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_randn_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 58%] 2023-03-31T07:05:47.6233786Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_real_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6234192Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reciprocal_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:47.6234633Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reciprocal_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6235026Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_remainder_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6235426Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_renorm_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:47.6235821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_renorm_cpu_float64 PASSED [ 58%] 2023-03-31T07:05:47.6236203Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:47.6236595Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6237002Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_repeat_interleave_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6237413Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_reshape_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:47.6237992Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize_as__cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 59%] 2023-03-31T07:05:47.6238629Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resize_as__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 59%] 2023-03-31T07:05:47.6239083Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_neg_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:47.6239487Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_resolve_neg_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6239904Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_roll_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:47.6240328Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6240728Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_0_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6241156Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_3_cpu_float64 SKIPPED (Skipped!) [ 59%] 2023-03-31T07:05:47.6241581Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_round_decimals_neg_3_cpu_float64 SKIPPED (Skipped!) [ 59%] 2023-03-31T07:05:47.6241998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsub_cpu_complex128 PASSED [ 59%] 2023-03-31T07:05:47.6242392Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_rsub_cpu_float64 PASSED [ 59%] 2023-03-31T07:05:47.6242977Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scalar_tensor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 59%] 2023-03-31T07:05:47.6243805Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scalar_tensor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6244264Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_add_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:47.6244715Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:47.6245126Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_scatter_reduce_sum_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:47.6245522Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:47.6245917Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_select_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:47.6246305Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sgn_cpu_float64 PASSED [ 60%] 2023-03-31T07:05:47.6246881Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_short_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6247564Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_bartlett_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6248232Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_blackman_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6248898Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6249555Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_hann_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6250197Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_signal_windows_kaiser_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 60%] 2023-03-31T07:05:47.6250664Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sin_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:47.6251063Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinc_cpu_complex128 PASSED [ 60%] 2023-03-31T07:05:47.6251456Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sinc_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6251850Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_softmax_with_dtype_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6252256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sort_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6252873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_airy_ai_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6253561Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6254210Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_bessel_j1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6254886Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6255565Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6256038Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_entr_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6256451Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_erfcx_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6257058Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_hermite_polynomial_h_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6257772Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_hermite_polynomial_he_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6258252Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_i1e_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6258867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_laguerre_polynomial_l_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 61%] 2023-03-31T07:05:47.6259337Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_log_ndtr_cpu_float64 PASSED [ 61%] 2023-03-31T07:05:47.6259949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:47.6260652Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:47.6261134Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_ndtri_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:47.6261560Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_polygamma_special_polygamma_n_0_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:47.6262205Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_scaled_modified_bessel_k1_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:47.6263028Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:47.6263758Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:47.6264477Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_shifted_chebyshev_polynomial_w_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 62%] 2023-03-31T07:05:47.6265162Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_spherical_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:47.6265821Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_special_zeta_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 62%] 2023-03-31T07:05:47.6266328Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_split_list_args_cpu_complex128 PASSED [ 62%] 2023-03-31T07:05:47.6266772Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sqrt_cpu_complex128 PASSED [ 62%] 2023-03-31T07:05:47.6267162Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_square_cpu_complex128 PASSED [ 62%] 2023-03-31T07:05:47.6267562Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_square_cpu_float64 PASSED [ 62%] 2023-03-31T07:05:47.6267962Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_stack_cpu_complex128 PASSED [ 62%] 2023-03-31T07:05:47.6268356Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6268731Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6269124Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_mean_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6269535Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_unbiased_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6269933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_std_unbiased_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6270335Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sub_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6270744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6271136Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_sum_to_size_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6271519Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_t_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6271903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_t_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6272296Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_along_dim_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6272702Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_take_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6273092Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tan_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6273508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tanh_cpu_complex128 PASSED [ 63%] 2023-03-31T07:05:47.6273903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tanh_cpu_float64 PASSED [ 63%] 2023-03-31T07:05:47.6274303Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensor_split_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6274709Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tensordot_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6275095Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tile_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6275492Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_to_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6275883Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_topk_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6276287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_transpose_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6276684Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_transpose_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6277087Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapezoid_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6277481Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_trapz_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6277877Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_triangular_solve_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6278281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_tril_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6278723Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_true_divide_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6279152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unbind_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6279539Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unbind_cpu_float64 PASSED [ 64%] 2023-03-31T07:05:47.6279943Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unflatten_cpu_complex128 PASSED [ 64%] 2023-03-31T07:05:47.6280353Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_copy_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6280762Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unfold_copy_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6281354Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unique_consecutive_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 65%] 2023-03-31T07:05:47.6281832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsafe_split_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6282250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsqueeze_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6282658Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_unsqueeze_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6283282Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6283678Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6284079Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_var_unbiased_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6284482Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vdot_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6284875Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_complex_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6285287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6285683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_view_as_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6286105Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vsplit_cpu_complex128 PASSED [ 65%] 2023-03-31T07:05:47.6286505Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_vsplit_cpu_float64 PASSED [ 65%] 2023-03-31T07:05:47.6286901Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_where_cpu_complex128 PASSED [ 66%] 2023-03-31T07:05:47.6287487Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 66%] 2023-03-31T07:05:47.6288123Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 66%] 2023-03-31T07:05:47.6288766Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_zeros_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 66%] 2023-03-31T07:05:47.6289287Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_H_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6289812Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_T_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6290344Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___getitem___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6290879Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___radd___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6291460Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___radd___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6292046Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rdiv___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6292585Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rdiv___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6293102Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmatmul___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6293634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmul___cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6294171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rmul___cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6294654Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD___rpow___cpu_float64 SKIPPED (Skipped!) [ 66%] 2023-03-31T07:05:47.6295191Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__segment_reduce_lengths_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 66%] 2023-03-31T07:05:47.6295744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__softmax_backward_data_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:47.6296308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD__upsample_bilinear2d_aa_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 67%] 2023-03-31T07:05:47.6296798Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_abs_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6297211Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_acos_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6297640Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_add_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6297844Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addbmm_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6298043Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addbmm_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6298243Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcmul_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6298439Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addcmul_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6298646Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6298845Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6299063Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_decomposed_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6299263Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmm_decomposed_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6299467Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmv_cpu_complex128 PASSED [ 67%] 2023-03-31T07:05:47.6299663Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addmv_cpu_float64 PASSED [ 67%] 2023-03-31T07:05:47.6299859Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_addr_cpu_float64 PASSED [ 68%] 2023-03-31T07:05:47.6300272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_all_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6300691Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_allclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6301082Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_allclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6301340Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_amin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:47.6301718Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_aminmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6301984Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_angle_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:47.6302243Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_angle_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:47.6302701Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_arange_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6303114Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6303504Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argwhere_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6303885Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_argwhere_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 68%] 2023-03-31T07:05:47.6304109Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_cpu_float64 SKIPPED (Numerous errors) [ 68%] 2023-03-31T07:05:47.6304328Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_partial_views_cpu_float64 XFAIL [ 68%] 2023-03-31T07:05:47.6304631Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_as_strided_scatter_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 68%] 2023-03-31T07:05:47.6304830Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asin_cpu_complex128 PASSED [ 69%] 2023-03-31T07:05:47.6305028Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asin_cpu_float64 PASSED [ 69%] 2023-03-31T07:05:47.6305282Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_asinh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6305544Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6305815Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_2d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6306069Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6306337Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_3d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6306596Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_atleast_3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6306832Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_baddbmm_cpu_complex128 PASSED [ 69%] 2023-03-31T07:05:47.6307121Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bernoulli_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6307390Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bfloat16_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6307657Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_block_diag_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6307914Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bmm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6308170Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bmm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 69%] 2023-03-31T07:05:47.6308547Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_bool_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 69%] 2023-03-31T07:05:47.6308867Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_broadcast_to_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6309249Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_byte_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 70%] 2023-03-31T07:05:47.6309509Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cartesian_prod_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6309942Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cat_cpu_complex128 SKIPPED (TODO(whc) fix pre-existing bug with cat for newly added opinfo for empty+nonempty) [ 70%] 2023-03-31T07:05:47.6310201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdist_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6310497Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cdouble_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6310760Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cfloat_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6311018Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cfloat_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6311277Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chunk_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6311541Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_chunk_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6311744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_cpu_float64 PASSED [ 70%] 2023-03-31T07:05:47.6311949Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_max_cpu_float64 PASSED [ 70%] 2023-03-31T07:05:47.6312150Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clamp_min_cpu_float64 PASSED [ 70%] 2023-03-31T07:05:47.6312405Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_clone_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6312656Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_column_stack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 70%] 2023-03-31T07:05:47.6312966Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_combinations_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6313293Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_complex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6313555Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6313762Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_conj_physical_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:47.6314029Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_constant_pad_nd_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6314233Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_copysign_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:47.6314440Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cosh_cpu_complex128 PASSED [ 71%] 2023-03-31T07:05:47.6314834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_count_nonzero_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 71%] 2023-03-31T07:05:47.6315127Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cov_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6315384Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cross_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6315627Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cummin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6315830Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumprod_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:47.6316115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumulative_trapezoid_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6316423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_cumulative_trapezoid_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 71%] 2023-03-31T07:05:47.6316624Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_deg2rad_cpu_float64 PASSED [ 71%] 2023-03-31T07:05:47.6316883Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6317152Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_embed_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6317414Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diag_embed_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6317683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagflat_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6317954Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_copy_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6318220Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_copy_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6318480Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6318773Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_scatter_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6319079Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diagonal_scatter_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6319340Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_diff_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6319539Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_digamma_cpu_float64 PASSED [ 72%] 2023-03-31T07:05:47.6319795Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dist_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 72%] 2023-03-31T07:05:47.6320016Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_floor_rounding_cpu_float64 PASSED [ 72%] 2023-03-31T07:05:47.6320231Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_no_rounding_mode_cpu_float64 PASSED [ 72%] 2023-03-31T07:05:47.6320446Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_div_trunc_rounding_cpu_float64 PASSED [ 72%] 2023-03-31T07:05:47.6320730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dot_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6320991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_double_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6321250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dsplit_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6321513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dsplit_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6321765Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dstack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6322050Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_dstack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6322312Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_einsum_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 73%] 2023-03-31T07:05:47.6322710Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6323303Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6323708Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_permuted_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6324100Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_empty_permuted_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6324480Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_equal_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6324857Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_equal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 73%] 2023-03-31T07:05:47.6325106Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_erfinv_cpu_float64 PASSED [ 73%] 2023-03-31T07:05:47.6325338Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp2_cpu_float64 PASSED [ 73%] 2023-03-31T07:05:47.6325540Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp_cpu_complex128 PASSED [ 74%] 2023-03-31T07:05:47.6325725Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_exp_cpu_float64 PASSED [ 74%] 2023-03-31T07:05:47.6325991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_as_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6326250Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expand_as_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6326449Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_expm1_cpu_complex128 PASSED [ 74%] 2023-03-31T07:05:47.6326713Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6326974Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6327271Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6327534Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftn_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6327793Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6328059Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_fftshift_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6328320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6328607Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6328865Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_hfftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6329125Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 74%] 2023-03-31T07:05:47.6329387Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6329655Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftn_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6329918Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ifftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6330180Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_ihfft2_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6330446Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft2_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6330748Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6331031Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_irfftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6331294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fft_rfftn_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6331491Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fill_cpu_float64 PASSED [ 75%] 2023-03-31T07:05:47.6331753Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flatten_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6331998Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flatten_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6332253Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flip_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6332516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flip_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6332805Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fliplr_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6333061Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fliplr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:47.6333317Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_flipud_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:47.6333579Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:47.6333792Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_power_cpu_complex128 PASSED [ 76%] 2023-03-31T07:05:47.6334023Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_float_power_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:47.6334279Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:47.6334474Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_fmod_cpu_float64 PASSED [ 76%] 2023-03-31T07:05:47.6334838Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6335225Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_full_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6335490Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_gather_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:47.6335861Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ge_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6336251Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geometric_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6336634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geqrf_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6337036Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_geqrf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6337445Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_histogramdd_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 76%] 2023-03-31T07:05:47.6337710Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_hsplit_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 76%] 2023-03-31T07:05:47.6338086Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_igamma_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6338292Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_add_cpu_complex128 PASSED [ 77%] 2023-03-31T07:05:47.6338497Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_fill_cpu_complex128 PASSED [ 77%] 2023-03-31T07:05:47.6338685Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_fill_cpu_float64 PASSED [ 77%] 2023-03-31T07:05:47.6338889Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_put_cpu_float64 PASSED [ 77%] 2023-03-31T07:05:47.6339193Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_index_select_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 77%] 2023-03-31T07:05:47.6339451Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_inner_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 77%] 2023-03-31T07:05:47.6339828Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_int_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6340198Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_int_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6340580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isclose_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6340985Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isclose_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6341366Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isnan_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6341747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isneginf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6342128Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_isposinf_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 77%] 2023-03-31T07:05:47.6342393Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_istft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 77%] 2023-03-31T07:05:47.6342709Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_2inputs_2outputs_cpu_float64 SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:05:47.6342966Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_4inputs_with_extra_args_cpu_float64 SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:05:47.6343200Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_cpu_complex128 SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:05:47.6343429Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_jiterator_binary_cpu_float64 SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:05:47.6343726Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kron_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6344014Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kron_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6344278Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_kthvalue_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6344545Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ldexp_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6344802Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ldexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6345003Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lerp_cpu_complex128 PASSED [ 78%] 2023-03-31T07:05:47.6345203Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lerp_cpu_float64 PASSED [ 78%] 2023-03-31T07:05:47.6345389Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lgamma_cpu_float64 PASSED [ 78%] 2023-03-31T07:05:47.6345683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cond_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6345948Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cross_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6346208Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_det_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 78%] 2023-03-31T07:05:47.6346480Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_diagonal_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6346747Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eig_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6347055Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigh_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6347324Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_eigvals_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6347611Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_householder_product_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6347891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_householder_product_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6348165Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_ex_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6348429Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_inv_ex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6348841Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:47.6349238Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_ex_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:47.6349664Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_factor_ex_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:47.6350084Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_ldl_solve_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 79%] 2023-03-31T07:05:47.6350354Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6350619Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6350888Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 79%] 2023-03-31T07:05:47.6351157Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_factor_ex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6351426Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_lu_solve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6351730Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6352130Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_matrix_rank_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 80%] 2023-03-31T07:05:47.6352397Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_multi_dot_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6352669Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6352936Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6353246Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6353533Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_norm_subgradients_at_zero_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6353800Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_pinv_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6354067Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_qr_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6354340Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_slogdet_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6354610Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_slogdet_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6354873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6355143Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_ex_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 80%] 2023-03-31T07:05:47.6355435Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_ex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6355742Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_solve_triangular_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6356014Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svd_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6356276Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svd_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6356543Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svdvals_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6356812Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_svdvals_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6357074Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorinv_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6357373Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorinv_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6357644Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_tensorsolve_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6357913Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vecdot_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6358179Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vecdot_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6358449Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_vector_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 81%] 2023-03-31T07:05:47.6358683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log1p_cpu_complex128 PASSED [ 81%] 2023-03-31T07:05:47.6358882Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log2_cpu_complex128 PASSED [ 81%] 2023-03-31T07:05:47.6359078Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log2_cpu_float64 PASSED [ 81%] 2023-03-31T07:05:47.6359460Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_normal_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6359728Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6359994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_log_softmax_with_dtype_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6360269Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logaddexp_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6360540Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logcumsumexp_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6360803Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logcumsumexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6361140Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logdet_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6361553Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_and_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6361944Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_not_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6362325Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logical_xor_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6362715Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logspace_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6362980Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_logsumexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6363612Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_long_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 82%] 2023-03-31T07:05:47.6363922Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6364173Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 82%] 2023-03-31T07:05:47.6364420Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_lu_solve_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6364683Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mH_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6364939Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mT_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6365246Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_amax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6365508Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_amin_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6365900Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_argmax_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 83%] 2023-03-31T07:05:47.6366104Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_fill_cpu_float64 PASSED [ 83%] 2023-03-31T07:05:47.6366378Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_logaddexp_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6366645Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_median_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6366912Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6367190Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_normalize_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6367404Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_scatter_cpu_complex128 PASSED [ 83%] 2023-03-31T07:05:47.6367634Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_scatter_cpu_float64 PASSED [ 83%] 2023-03-31T07:05:47.6367933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_softmax_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6368199Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_std_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6368461Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_sum_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:47.6368719Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_sum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6368991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_masked_var_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6369259Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_matrix_exp_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6369547Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_maximum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6369804Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_median_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6370087Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_meshgrid_variadic_tensors_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6370353Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_binary_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6370628Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_reduction_no_dim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6370933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_min_reduction_with_dim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6371180Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_minimum_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6371436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_mode_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6371698Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_movedim_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6371964Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_movedim_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6372171Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nan_to_num_cpu_float64 PASSED [ 84%] 2023-03-31T07:05:47.6372436Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanmedian_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 84%] 2023-03-31T07:05:47.6372699Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nanquantile_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:47.6373085Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_narrow_copy_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6373529Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ne_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6373763Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_neg_cpu_complex128 PASSED [ 85%] 2023-03-31T07:05:47.6373961Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_neg_cpu_float64 PASSED [ 85%] 2023-03-31T07:05:47.6374338Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6374716Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_empty_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6375093Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_full_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6375477Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6375866Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6376272Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_new_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 85%] 2023-03-31T07:05:47.6376559Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:47.6376848Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_avg_pool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:47.6377138Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_adaptive_max_pool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 85%] 2023-03-31T07:05:47.6377395Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_alpha_dropout_cpu_float64 PASSED [ 85%] 2023-03-31T07:05:47.6377672Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_avg_pool1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6377950Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_batch_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6378223Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6378513Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose1d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6378793Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6379079Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_conv_transpose3d_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6379367Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_cosine_embedding_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6379667Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_ctc_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6379910Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_dropout3d_cpu_float64 PASSED [ 86%] 2023-03-31T07:05:47.6380192Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_embedding_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6380432Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_with_train_cpu_float64 XFAIL [ 86%] 2023-03-31T07:05:47.6380680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_feature_alpha_dropout_without_train_cpu_float64 PASSED [ 86%] 2023-03-31T07:05:47.6380966Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_fractional_max_pool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6381251Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_fractional_max_pool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6381563Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_gaussian_nll_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 86%] 2023-03-31T07:05:47.6381840Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_gelu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6382118Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_grid_sample_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6382384Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_group_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6382691Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardsigmoid_cpu_float64 PASSED [ 87%] 2023-03-31T07:05:47.6383007Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hardswish_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6383296Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_hinge_embedding_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6383579Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_area_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6383870Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_bicubic_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6384160Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_interpolate_linear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6384439Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_linear_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6384718Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_logsigmoid_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6384999Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_margin_ranking_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6385304Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool1d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6385603Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool2d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6385872Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_pool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 87%] 2023-03-31T07:05:47.6386159Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool2d_grad_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6386421Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool3d_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6386707Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_max_unpool3d_grad_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6386999Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_multilabel_margin_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6387309Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_circular_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6387592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_constant_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6387875Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_constant_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6388160Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_reflect_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6388473Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pad_reflect_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6388749Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pdist_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6389030Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_shuffle_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6389320Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6389606Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_pixel_unshuffle_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6389894Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_poisson_nll_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6390169Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_prelu_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 88%] 2023-03-31T07:05:47.6390423Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_relu6_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6390665Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_rrelu_cpu_float64 XFAIL [ 89%] 2023-03-31T07:05:47.6390994Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_scaled_dot_product_attention_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6391209Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_selu_cpu_float64 PASSED [ 89%] 2023-03-31T07:05:47.6391498Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softmin_with_dtype_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6391775Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softshrink_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6392056Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_softsign_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6392339Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_tanhshrink_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6392659Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_tanhshrink_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6392880Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_threshold_cpu_float64 PASSED [ 89%] 2023-03-31T07:05:47.6393172Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_loss_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6393479Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_triplet_margin_with_distance_loss_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6393759Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_unfold_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6394047Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_unfold_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6394333Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nn_functional_upsample_bilinear_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 89%] 2023-03-31T07:05:47.6394729Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_nonzero_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:47.6394988Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6395256Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_fro_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6395526Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_inf_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6395787Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_inf_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6396048Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_nuc_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6396338Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_norm_nuc_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6396595Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_cpu_float64 SKIPPED (Gradients are incorrect!) [ 90%] 2023-03-31T07:05:47.6397002Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_in_place_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:47.6397242Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_normal_number_mean_cpu_float64 SKIPPED (Gradients are incorrect!) [ 90%] 2023-03-31T07:05:47.6397623Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:47.6397980Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ones_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 90%] 2023-03-31T07:05:47.6398241Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ormqr_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6398535Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_outer_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6398800Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_permute_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 90%] 2023-03-31T07:05:47.6399062Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_permute_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6399281Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_0_cpu_float64 PASSED [ 91%] 2023-03-31T07:05:47.6399515Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_1_cpu_float64 SKIPPED (Skipped!) [ 91%] 2023-03-31T07:05:47.6399744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_polygamma_polygamma_n_4_cpu_float64 SKIPPED (Skipped!) [ 91%] 2023-03-31T07:05:47.6400043Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_positive_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6400305Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_positive_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6400565Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_prod_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6400827Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_prod_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6401017Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_put_cpu_complex128 PASSED [ 91%] 2023-03-31T07:05:47.6401215Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_put_cpu_float64 PASSED [ 91%] 2023-03-31T07:05:47.6401471Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_qr_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:47.6401670Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rad2deg_cpu_float64 PASSED [ 91%] 2023-03-31T07:05:47.6402054Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rand_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:47.6402457Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:47.6402873Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_randn_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 91%] 2023-03-31T07:05:47.6403317Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_ravel_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6403582Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_real_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6403786Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_renorm_cpu_complex128 PASSED [ 92%] 2023-03-31T07:05:47.6403986Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_renorm_cpu_float64 PASSED [ 92%] 2023-03-31T07:05:47.6404238Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_repeat_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6404510Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_as_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6404830Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_reshape_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6405215Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resize_as__cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 92%] 2023-03-31T07:05:47.6405485Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_conj_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6405751Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_resolve_neg_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6406012Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_roll_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6406323Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_round_decimals_neg_3_cpu_float64 SKIPPED (Skipped!) [ 92%] 2023-03-31T07:05:47.6406522Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsqrt_cpu_float64 PASSED [ 92%] 2023-03-31T07:05:47.6406777Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_rsub_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 92%] 2023-03-31T07:05:47.6406984Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_add_cpu_complex128 PASSED [ 92%] 2023-03-31T07:05:47.6407176Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_cpu_complex128 PASSED [ 93%] 2023-03-31T07:05:47.6407379Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6407590Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_amax_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6407806Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_mean_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6408015Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_prod_cpu_float64 XFAIL [ 93%] 2023-03-31T07:05:47.6408224Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_scatter_reduce_sum_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6408490Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_select_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 93%] 2023-03-31T07:05:47.6408727Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sgn_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6409144Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_short_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6409525Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_short_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6409709Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sigmoid_cpu_float64 PASSED [ 93%] 2023-03-31T07:05:47.6410115Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_bartlett_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6410516Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6410929Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_exponential_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6411364Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_gaussian_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 93%] 2023-03-31T07:05:47.6411780Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_general_cosine_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6412179Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signal_windows_kaiser_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6412567Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_signbit_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6412831Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6413118Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6413386Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_slice_scatter_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6413655Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_mm_reduce_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6413931Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sparse_sampled_addmm_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6414317Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6414744Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6415158Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6415592Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 94%] 2023-03-31T07:05:47.6415891Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_entr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6416332Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_hermite_polynomial_h_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 94%] 2023-03-31T07:05:47.6416604Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i0e_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 94%] 2023-03-31T07:05:47.6416868Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_i1_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6417138Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_log_ndtr_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6417551Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_i0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:47.6417959Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_modified_bessel_k0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:47.6418284Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_polygamma_special_polygamma_n_0_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6418725Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_t_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 95%] 2023-03-31T07:05:47.6419165Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_u_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 95%] 2023-03-31T07:05:47.6419599Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_shifted_chebyshev_polynomial_v_cpu_float64 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 95%] 2023-03-31T07:05:47.6420033Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_spherical_bessel_j0_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:47.6420417Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_special_zeta_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 95%] 2023-03-31T07:05:47.6420680Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6420933Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6421201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_list_args_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6421476Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_with_sizes_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6421748Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_split_with_sizes_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 95%] 2023-03-31T07:05:47.6421952Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sqrt_cpu_complex128 PASSED [ 96%] 2023-03-31T07:05:47.6422149Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sqrt_cpu_float64 PASSED [ 96%] 2023-03-31T07:05:47.6422353Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_square_cpu_complex128 PASSED [ 96%] 2023-03-31T07:05:47.6422665Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_square_cpu_float64 PASSED [ 96%] 2023-03-31T07:05:47.6422903Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_squeeze_multiple_cpu_complex128 PASSED [ 96%] 2023-03-31T07:05:47.6423170Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stack_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6423430Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6423684Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6423945Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6424223Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_unbiased_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6424522Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_mean_unbiased_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6424794Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_std_unbiased_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6425055Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stft_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6425314Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_stft_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6425576Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_sum_to_size_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 96%] 2023-03-31T07:05:47.6425863Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6426112Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_svd_lowrank_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6426366Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_take_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6426564Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tan_cpu_complex128 PASSED [ 97%] 2023-03-31T07:05:47.6426765Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tanh_cpu_complex128 PASSED [ 97%] 2023-03-31T07:05:47.6427035Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensor_split_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6427308Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensor_split_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6427574Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tensordot_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6427834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tile_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6428117Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6428408Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_to_sparse_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6428668Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_topk_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6428921Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trace_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6429166Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_trapz_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 97%] 2023-03-31T07:05:47.6429366Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tril_cpu_complex128 PASSED [ 97%] 2023-03-31T07:05:47.6429565Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_tril_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:47.6429766Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triu_cpu_complex128 PASSED [ 98%] 2023-03-31T07:05:47.6429964Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_triu_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:47.6430201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_true_divide_cpu_complex128 PASSED [ 98%] 2023-03-31T07:05:47.6430404Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_true_divide_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:47.6430669Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unflatten_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6430935Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_copy_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6431197Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unfold_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6431620Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_uniform_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 98%] 2023-03-31T07:05:47.6431985Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unique_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 98%] 2023-03-31T07:05:47.6432192Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsqueeze_cpu_complex128 PASSED [ 98%] 2023-03-31T07:05:47.6432394Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_unsqueeze_cpu_float64 PASSED [ 98%] 2023-03-31T07:05:47.6432658Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6432916Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6433180Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 98%] 2023-03-31T07:05:47.6433452Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_var_mean_unbiased_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6433709Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vdot_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6434004Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_as_complex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6434294Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_copy_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6434556Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_view_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6434814Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vsplit_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6435056Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_vstack_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6435319Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_where_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6435580Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_where_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 99%] 2023-03-31T07:05:47.6435779Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_xlogy_cpu_float64 PASSED [ 99%] 2023-03-31T07:05:47.6436017Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zero__cpu_complex128 PASSED [ 99%] 2023-03-31T07:05:47.6436215Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zero__cpu_float64 PASSED [ 99%] 2023-03-31T07:05:47.6436594Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 99%] 2023-03-31T07:05:47.6436970Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 99%] 2023-03-31T07:05:47.6437361Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_like_cpu_complex128 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [ 99%] 2023-03-31T07:05:47.6437781Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_zeros_like_cpu_float64 SKIPPED (Skipped! Op doesn't support autograd for this dtype.) [100%] 2023-03-31T07:05:47.6437793Z 2023-03-31T07:05:47.6438163Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-419d8c8bac08ffa6.xml - 2023-03-31T07:05:47.6438292Z ==== 846 passed, 639 skipped, 7 deselected, 21 xfailed in 249.55s (0:04:09) ==== 2023-03-31T07:05:47.6438473Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:05:47.6438480Z 2023-03-31T07:05:47.6438889Z ##[endgroup] 2023-03-31T07:05:47.6439223Z FINISHED PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_32ydsk_x.log) 2023-03-31T07:05:47.6439231Z 2023-03-31T07:05:47.9630007Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_ops_fwd_gradients.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '-k=_linalg_cholesky_', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:05:47.962540] 2023-03-31T07:05:53.0467836Z 2023-03-31T07:05:53.0468306Z Expand the folded group to see the log file of test_ops_fwd_gradients 2023-03-31T07:05:53.0469353Z ##[group]PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_5gurcpcm.log) 2023-03-31T07:05:53.0470070Z Test results will be stored in test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4297aefcfc4b8bbf.xml 2023-03-31T07:05:53.0470445Z ============================= test session starts ============================== 2023-03-31T07:05:53.0471052Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:05:53.0471395Z cachedir: .pytest_cache 2023-03-31T07:05:53.0471911Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:05:53.0472345Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:05:53.0472907Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:05:53.0473330Z collecting ... collected 3021 items / 3009 deselected / 12 selected 2023-03-31T07:05:53.0475143Z Running 12 items in this shard: test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_ex_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_cpu_float64, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_ex_cpu_complex128, test/test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_ex_cpu_float64 2023-03-31T07:05:53.0476776Z 2023-03-31T07:05:53.0476991Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_cpu_complex128 PASSED [ 8%] 2023-03-31T07:05:53.0477451Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_cpu_float64 PASSED [ 16%] 2023-03-31T07:05:53.0478051Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_ex_cpu_complex128 PASSED [ 25%] 2023-03-31T07:05:53.0478520Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_fn_fwgrad_bwgrad_linalg_cholesky_ex_cpu_float64 PASSED [ 33%] 2023-03-31T07:05:53.0478951Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_cpu_complex128 PASSED [ 41%] 2023-03-31T07:05:53.0479352Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_cpu_float64 PASSED [ 50%] 2023-03-31T07:05:53.0479771Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_ex_cpu_complex128 PASSED [ 58%] 2023-03-31T07:05:53.0480201Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_forward_mode_AD_linalg_cholesky_ex_cpu_float64 PASSED [ 66%] 2023-03-31T07:05:53.0480834Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 75%] 2023-03-31T07:05:53.0481443Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 83%] 2023-03-31T07:05:53.0482126Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_ex_cpu_complex128 SKIPPED (Skipped! Operation does not support inplace autograd.) [ 91%] 2023-03-31T07:05:53.0482697Z test_ops_fwd_gradients.py::TestFwdGradientsCPU::test_inplace_forward_mode_AD_linalg_cholesky_ex_cpu_float64 SKIPPED (Skipped! Operation does not support inplace autograd.) [100%] 2023-03-31T07:05:53.0483321Z 2023-03-31T07:05:53.0483901Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4297aefcfc4b8bbf.xml - 2023-03-31T07:05:53.0484377Z ================ 8 passed, 4 skipped, 3009 deselected in 1.80s ================= 2023-03-31T07:05:53.0484760Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:05:53.0484967Z 2023-03-31T07:05:53.0485293Z ##[endgroup] 2023-03-31T07:05:53.0485728Z FINISHED PRINTING LOG FILE of test_ops_fwd_gradients (/var/lib/jenkins/workspace/test/test-reports/test_ops_fwd_gradients_5gurcpcm.log) 2023-03-31T07:05:53.0486047Z 2023-03-31T07:05:53.0486221Z Running test_fx ... [2023-03-31 07:05:53.047681] 2023-03-31T07:05:53.0486888Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_fx.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:05:53.048010] 2023-03-31T07:10:14.9297790Z 2023-03-31T07:10:14.9298274Z Expand the folded group to see the log file of test_fx 2023-03-31T07:10:14.9299082Z ##[group]PRINTING LOG FILE of test_fx (/var/lib/jenkins/workspace/test/test-reports/test_fx_sxdy1_ua.log) 2023-03-31T07:10:14.9301872Z Test results will be stored in test-reports/python-pytest/test_fx/test_fx-b0efee86074bcae7.xml 2023-03-31T07:10:14.9302716Z ============================= test session starts ============================== 2023-03-31T07:10:14.9303544Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:10:14.9303941Z cachedir: .pytest_cache 2023-03-31T07:10:14.9304676Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:10:14.9305319Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:10:14.9306025Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:10:14.9306521Z collecting ... collected 1182 items 2023-03-31T07:10:14.9452265Z Running 1182 items in this shard: test/test_fx.py::TestSubgraphRewriter::test_matching_pattern_with_list_type_arg, test/test_fx.py::TestSubgraphRewriter::test_replace_pattern_with_filters, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_annotations_int, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_call_method, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_correct_output_replacement, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_graph_argument_order, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_internal_pattern_nodes_cannot_have_users_that_are_not_matched, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_local_revert, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_multiple_pattern_match, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_nodes_with_kwargs, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_is_entire_graph, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_output_pattern_node_can_have_users_that_are_not_matched, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_placeholder_matching, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_preserves_logic, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_consecutive_submodules, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_with_duplicated_outputs, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_with_multiple_outputs, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replaces_referenced_submodules, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_single_pattern_match, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_traced_as_callable, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_oneliner_pattern, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_overlapping_matches, test/test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_unused_args, test/test_fx.py::TestDCE::test_dead_chain, test/test_fx.py::TestDCE::test_dead_getattr, test/test_fx.py::TestDCE::test_dead_placeholder, test/test_fx.py::TestDCE::test_dead_placeholder_with_user, test/test_fx.py::TestDCE::test_keep_module_with_side_effects, test/test_fx.py::TestDCE::test_keep_torch_assert, test/test_fx.py::TestDCE::test_simple, test/test_fx.py::TestConstFold::test_check_inline_non_const, test/test_fx.py::TestConstFold::test_check_inline_non_const_mult_return, test/test_fx.py::TestConstFold::test_check_skip_folding_quant_dequant_pattern, test/test_fx.py::TestConstFold::test_const_fold_basic_one_attr_name_collision, test/test_fx.py::TestConstFold::test_const_fold_basic_one_attr_no_name_collision, test/test_fx.py::TestConstFold::test_const_fold_basic_placeholder_reordered, test/test_fx.py::TestConstFold::test_const_fold_basic_two_attr, test/test_fx.py::TestConstFold::test_const_fold_basic_two_attr_three_input, test/test_fx.py::TestConstFold::test_const_fold_has_inlined_call_module_node, test/test_fx.py::TestConstFold::test_const_fold_module_attr, test/test_fx.py::TestConstFold::test_const_fold_multi_const_folded_attrs, test/test_fx.py::TestConstFold::test_const_fold_noop, test/test_fx.py::TestConstFold::test_const_fold_submod_hierarchy, test/test_fx.py::TestConstFold::test_const_fold_tensor_meta, test/test_fx.py::TestConstFold::test_const_fold_unused_placeholder, test/test_fx.py::TestConstFold::test_dict_output, test/test_fx.py::TestConstFold::test_fold_module, test/test_fx.py::TestConstFold::test_retain_node_meta, test/test_fx.py::TestConstFold::test_three_outputs, test/test_fx.py::TestConstFold::test_two_outputs, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_dim_const, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_ndim_const, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_nelement_const, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_numel_const, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_shape_const, test/test_fx.py::TestConstParamShapeInControlFlow::test_param_size_const, test/test_fx.py::TestPassManager::test_pass_manager, test/test_fx.py::TestPassManager::test_pass_manager_bad_checks, test/test_fx.py::TestPassManager::test_pass_manager_checks, test/test_fx.py::TestPassManager::test_pass_manager_error, test/test_fx.py::TestPassManager::test_this_before_that_pass_constraint, test/test_fx.py::TestPassManager::test_topological_sort, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationInput_cpu, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationMetadata_cpu, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationTorchTensorCall_cpu, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_Mutation_cpu, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_ReturnList_cpu, test/test_fx.py::TestCommonPass::test_correctness_CSEPass_TakeList_cpu, test/test_fx.py::TestCommonPass::test_correctness_factory_CSEPass_FactoryFunctionCall_cpu, test/test_fx.py::TestCommonPass::test_correctness_factory_CSEPass_MutationFactory_cpu, test/test_fx.py::TestCSEPass::test_banned_list, test/test_fx.py::TestCSEPass::test_empty, test/test_fx.py::TestCSEPass::test_immutable_list_multiple_entries, test/test_fx.py::TestCSEPass::test_immutable_list_type, test/test_fx.py::TestCSEPass::test_kwarg, test/test_fx.py::TestCSEPass::test_nested_immutable_list_type, test/test_fx.py::TestCSEPass::test_nochange, test/test_fx.py::TestCSEPass::test_rand_like, test/test_fx.py::TestCSEPass::test_rand_n, test/test_fx.py::TestCSEPass::test_random, test/test_fx.py::TestCSEPass::test_simple, test/test_fx.py::TestCSEPass::test_simple_2, test/test_fx.py::TestCSEPass::test_simple_multiple_same_ops, test/test_fx.py::TestCSEPass::test_two_args, test/test_fx.py::TestCSEPass::test_two_args_default, test/test_fx.py::TestMatcher::test_subgraph_matcher_with_attributes, test/test_fx.py::TestMatcher::test_subgraph_matcher_with_list, test/test_fx.py::TestMatcher::test_subgraph_matcher_with_list_bad, test/test_fx.py::VerifierTest::test_aten_dialect, test/test_fx.py::VerifierTest::test_aten_wrong_mem_format, test/test_fx.py::VerifierTest::test_aten_wrong_mem_format_buffer, test/test_fx.py::VerifierTest::test_verifier, test/test_fx.py::VerifierTest::test_verifier_no_functional, test/test_fx.py::VerifierTest::testr_verifier_call_module, test/test_fx.py::AnnotationsTest::test_annotate, test/test_fx.py::AnnotationsTest::test_annotations, test/test_fx.py::AnnotationsTest::test_broadcasting1, test/test_fx.py::AnnotationsTest::test_broadcasting2, test/test_fx.py::AnnotationsTest::test_broadcasting3, test/test_fx.py::AnnotationsTest::test_consistency, test/test_fx.py::AnnotationsTest::test_precision, test/test_fx.py::TypeCheckerTest::test_flatten_fully_static, test/test_fx.py::TypeCheckerTest::test_resnet50, test/test_fx.py::TypeCheckerTest::test_symbolic_add_with_broadcast, test/test_fx.py::TypeCheckerTest::test_symbolic_add_with_broadcast_2, test/test_fx.py::TypeCheckerTest::test_type_check_add_false, test/test_fx.py::TypeCheckerTest::test_type_check_add_true, test/test_fx.py::TypeCheckerTest::test_type_check_add_with_broadcast, test/test_fx.py::TypeCheckerTest::test_type_check_add_with_scalar, test/test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D, test/test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D_broadcast, test/test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D_false, test/test_fx.py::TypeCheckerTest::test_type_check_batch_norm_symbolic, test/test_fx.py::TypeCheckerTest::test_type_check_conv2D, test/test_fx.py::TypeCheckerTest::test_type_check_conv2D_2, test/test_fx.py::TypeCheckerTest::test_type_check_conv2D_2_fully_static, test/test_fx.py::TypeCheckerTest::test_type_check_conv2D_maxpool2d_flatten, test/test_fx.py::TypeCheckerTest::test_type_check_conv2D_types, test/test_fx.py::TypeCheckerTest::test_type_check_flatten, test/test_fx.py::TypeCheckerTest::test_type_check_flatten3, test/test_fx.py::TypeCheckerTest::test_type_check_flatten_2, test/test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_false, test/test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_true, test/test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_true_param_false, test/test_fx.py::TypeCheckerTest::test_type_check_reshape_false, test/test_fx.py::TypeCheckerTest::test_type_check_reshape_true, test/test_fx.py::TypeCheckerTest::test_type_check_symbolic_inferenceconv2D_maxpool2d_flatten, test/test_fx.py::TypeCheckerTest::test_type_check_transpose_False, test/test_fx.py::TypeCheckerTest::test_type_check_transpose_true, test/test_fx.py::TypeCheckerTest::test_type_maxpool2d_fully_static, test/test_fx.py::TypeCheckerTest::test_type_typechecl_maxpool2d_3dinput, test/test_fx.py::TypeCheckerTest::test_typecheck_basicblock, test/test_fx.py::TestFX::test_all_input_nodes, test/test_fx.py::TestFX::test_annotation_with_future, test/test_fx.py::TestFX::test_annotations_empty_tuple, test/test_fx.py::TestFX::test_annotations_with_forward_references, test/test_fx.py::TestFX::test_annotations_with_no_forward_references, test/test_fx.py::TestFX::test_annotations_with_non_torch_reference_and_internal_forward_references, test/test_fx.py::TestFX::test_annotations_with_non_torch_reference_and_no_internal_forward_references, test/test_fx.py::TestFX::test_args_kwargs, test/test_fx.py::TestFX::test_args_kwargs_no_self, test/test_fx.py::TestFX::test_assert, test/test_fx.py::TestFX::test_ast_rewriter_reassigns_submodules, test/test_fx.py::TestFX::test_ast_rewriter_rewrites_assert, test/test_fx.py::TestFX::test_ast_rewriter_rewrites_assert_with_message, test/test_fx.py::TestFX::test_ast_rewriter_wrap, test/test_fx.py::TestFX::test_ast_rewriter_wrap_fn_directly, test/test_fx.py::TestFX::test_ast_rewriter_wrap_with_submodule, test/test_fx.py::TestFX::test_ast_rewriter_wrapped_via_decorator, test/test_fx.py::TestFX::test_ast_rewriter_wrapped_via_decorator_and_transformed, test/test_fx.py::TestFX::test_autowrap_functions, test/test_fx.py::TestFX::test_concrete_arg_none_assert, test/test_fx.py::TestFX::test_construct_root_dict, test/test_fx.py::TestFX::test_control_flow_tracing, test/test_fx.py::TestFX::test_copy_it, test/test_fx.py::TestFX::test_copy_no_remap, test/test_fx.py::TestFX::test_ctx_mgr, test/test_fx.py::TestFX::test_custom_codegen, test/test_fx.py::TestFX::test_custom_codegen_with_transformer, test/test_fx.py::TestFX::test_custom_import, test/test_fx.py::TestFX::test_custom_proxy_dynamic_value, test/test_fx.py::TestFX::test_custom_proxy_input_dependent_control_flow, test/test_fx.py::TestFX::test_custom_proxy_type, test/test_fx.py::TestFX::test_custom_proxy_type_literal, test/test_fx.py::TestFX::test_custom_traceback_not_raised_when_exception_source_is_submodule, test/test_fx.py::TestFX::test_custom_traceback_raised_when_exception_source_is_graphmodule, test/test_fx.py::TestFX::test_deepcopy_graph_with_tracer_cls, test/test_fx.py::TestFX::test_deepcopy_graphmodule, test/test_fx.py::TestFX::test_deepcopy_graphmodule_with_transform, test/test_fx.py::TestFX::test_deepcopy_no_recursion, test/test_fx.py::TestFX::test_deepcopy_recursion_depth, test/test_fx.py::TestFX::test_deepcopy_tracer, test/test_fx.py::TestFX::test_deepcopy_with_submods_params, test/test_fx.py::TestFX::test_delete_unused_submodules_leaf, test/test_fx.py::TestFX::test_dict, test/test_fx.py::TestFX::test_direct_param_use, test/test_fx.py::TestFX::test_disallow_override, test/test_fx.py::TestFX::test_ellipsis, test/test_fx.py::TestFX::test_empty_graph_codegen, test/test_fx.py::TestFX::test_erase_node_error, test/test_fx.py::TestFX::test_example_shape_prop, test/test_fx.py::TestFX::test_find_uses, test/test_fx.py::TestFX::test_fn_type_annotation_empty, test/test_fx.py::TestFX::test_fn_type_annotations, test/test_fx.py::TestFX::test_fx_and_or, test/test_fx.py::TestFX::test_fx_create_arg, test/test_fx.py::TestFX::test_fx_shifts, test/test_fx.py::TestFX::test_fx_stateless, test/test_fx.py::TestFX::test_get_torch_func_signature, test/test_fx.py::TestFX::test_getitem, test/test_fx.py::TestFX::test_getitem_subproc, test/test_fx.py::TestFX::test_graph_edit_with_proxy, test/test_fx.py::TestFX::test_graph_fns, test/test_fx.py::TestFX::test_graph_module, test/test_fx.py::TestFX::test_graph_module_init_buffer_param_copied_dict_init, test/test_fx.py::TestFX::test_graph_module_init_buffer_param_copied_mod_init, test/test_fx.py::TestFX::test_graph_module_replicate_for_dp, test/test_fx.py::TestFX::test_graph_unique_names, test/test_fx.py::TestFX::test_graph_unique_names_manual, test/test_fx.py::TestFX::test_immutable_dict_pytree_ops, test/test_fx.py::TestFX::test_immutable_list_pytree_ops, test/test_fx.py::TestFX::test_imul_code_print, test/test_fx.py::TestFX::test_inf_nan, test/test_fx.py::TestFX::test_inf_nan_kwds, test/test_fx.py::TestFX::test_inline_graph, test/test_fx.py::TestFX::test_insertion_point, test/test_fx.py::TestFX::test_interpreter, test/test_fx.py::TestFX::test_interpreter_default_args, test/test_fx.py::TestFX::test_interpreter_gc_values, test/test_fx.py::TestFX::test_interpreter_noop_resnet18, test/test_fx.py::TestFX::test_interpreter_not_enough_args, test/test_fx.py::TestFX::test_interpreter_onthefly_swap, test/test_fx.py::TestFX::test_interpreter_partial_eval, test/test_fx.py::TestFX::test_interpreter_run_node_override, test/test_fx.py::TestFX::test_interpreter_star_args, test/test_fx.py::TestFX::test_interpreter_with_codegen, test/test_fx.py::TestFX::test_layout, test/test_fx.py::TestFX::test_leaf_module, test/test_fx.py::TestFX::test_matmul_tracing, test/test_fx.py::TestFX::test_module_deepcopy_edit_nodes, test/test_fx.py::TestFX::test_move_before, test/test_fx.py::TestFX::test_multi_insert_point, test/test_fx.py::TestFX::test_multiple_default_args, test/test_fx.py::TestFX::test_named_tuple_inlined, test/test_fx.py::TestFX::test_namedtuple_return_qualname, test/test_fx.py::TestFX::test_namedtuple_return_trace, test/test_fx.py::TestFX::test_native_callable, test/test_fx.py::TestFX::test_nn_module_stack, test/test_fx.py::TestFX::test_no_mutation, test/test_fx.py::TestFX::test_node_tagging, test/test_fx.py::TestFX::test_nonetype_annotation, test/test_fx.py::TestFX::test_partial_trace, test/test_fx.py::TestFX::test_pickle_custom_import, test/test_fx.py::TestFX::test_pickle_graphmodule, test/test_fx.py::TestFX::test_pickle_nonetype_annotation, test/test_fx.py::TestFX::test_pickle_torch_custom_ops, test/test_fx.py::TestFX::test_prepend_self, test/test_fx.py::TestFX::test_pretty_print, test/test_fx.py::TestFX::test_pretty_print_graph, test/test_fx.py::TestFX::test_pretty_print_node, test/test_fx.py::TestFX::test_pretty_print_targets, test/test_fx.py::TestFX::test_profiler_ranges_side_effect, test/test_fx.py::TestFX::test_pytree, test/test_fx.py::TestFX::test_pytree_concrete, test/test_fx.py::TestFX::test_reassign_args_kwargs_uses, test/test_fx.py::TestFX::test_regular_and_default_args, test/test_fx.py::TestFX::test_remove_uses, test/test_fx.py::TestFX::test_remove_uses_with_custom_filter, test/test_fx.py::TestFX::test_replace_input, test/test_fx.py::TestFX::test_replace_uses, test/test_fx.py::TestFX::test_reserved_getattr, test/test_fx.py::TestFX::test_return_tuple, test/test_fx.py::TestFX::test_return_type_exists, test/test_fx.py::TestFX::test_script_method_trace, test/test_fx.py::TestFX::test_script_tensor_constant, test/test_fx.py::TestFX::test_sequential, test/test_fx.py::TestFX::test_shape_prop_aggregate, test/test_fx.py::TestFX::test_shape_prop_layout, test/test_fx.py::TestFX::test_shape_prop_layout_3d, test/test_fx.py::TestFX::test_single_default_arg, test/test_fx.py::TestFX::test_snake_case, test/test_fx.py::TestFX::test_sqrt, test/test_fx.py::TestFX::test_stack_traces, test/test_fx.py::TestFX::test_stack_traces_with_transformer, test/test_fx.py::TestFX::test_string_literal_return, test/test_fx.py::TestFX::test_submodule_manipulation_API, test/test_fx.py::TestFX::test_symbolic_trace_assert, test/test_fx.py::TestFX::test_symbolic_trace_sequential, test/test_fx.py::TestFX::test_tensor_attribute, test/test_fx.py::TestFX::test_tensor_attribute_coalseced, test/test_fx.py::TestFX::test_tensor_constant, test/test_fx.py::TestFX::test_throw_out_variant, test/test_fx.py::TestFX::test_torch_custom_ops, test/test_fx.py::TestFX::test_torch_fx_getattr, test/test_fx.py::TestFX::test_torch_fx_len, test/test_fx.py::TestFX::test_torch_op_overloads, test/test_fx.py::TestFX::test_torchbind_class_attribute_in_fx, test/test_fx.py::TestFX::test_torchbind_class_attribute_in_fx_tensor_arg, test/test_fx.py::TestFX::test_trace_buffer_slice, test/test_fx.py::TestFX::test_trace_dict_int_keys, test/test_fx.py::TestFX::test_trace_dict_proxy_keys, test/test_fx.py::TestFX::test_trace_fn_constant, test/test_fx.py::TestFX::test_trace_function, test/test_fx.py::TestFX::test_trace_multiple_funcs, test/test_fx.py::TestFX::test_tracing_graphmodules_as_leaf_submodules, test/test_fx.py::TestFX::test_transformer_multi_outputs, test/test_fx.py::TestFX::test_transformer_noop, test/test_fx.py::TestFX::test_transformer_op_swap, test/test_fx.py::TestFX::test_transformer_preserves_nn_module_stack_for_get_attr, test/test_fx.py::TestFX::test_tuple_no_subscript, test/test_fx.py::TestFX::test_typename_print, test/test_fx.py::TestFX::test_unpack, test/test_fx.py::TestFX::test_unpack_dict_better_error, test/test_fx.py::TestFX::test_unpack_list_better_error, test/test_fx.py::TestFX::test_update_args_api, test/test_fx.py::TestFX::test_update_args_kwargs_yells_at_you, test/test_fx.py::TestFX::test_update_kwargs_api, test/test_fx.py::TestFX::test_user_friendly_call_provenance_with_function, test/test_fx.py::TestFX::test_user_friendly_call_provenance_with_module, test/test_fx.py::TestFX::test_wrap, test/test_fx.py::TestFX::test_wrap_decorated_function, test/test_fx.py::TestFX::test_wrap_fn_directly, test/test_fx.py::TestFX::test_wrap_with_submodule, test/test_fx.py::TestFX::test_wrapped_method, test/test_fx.py::TestFX::test_wrapped_retrace, test/test_fx.py::TestFX::test_wrapped_via_decorator, test/test_fx.py::TestFX::test_wrapped_via_decorator_and_transformed, test/test_fx.py::TestFX::test_wrong_target_type, test/test_fx.py::TestFX::test_wrong_topo, test/test_fx.py::TestFXAPIBackwardCompatibility::test_class_member_back_compat, test/test_fx.py::TestFXAPIBackwardCompatibility::test_function_back_compat, test/test_fx.py::TestFXAPIBackwardCompatibility::test_public_api_surface, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool1d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool2d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool3d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_affine_grid, test/test_fx.py::TestFunctionalTracing::test_nn_functional_alpha_dropout, test/test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_batch_norm, test/test_fx.py::TestFunctionalTracing::test_nn_functional_bilinear, test/test_fx.py::TestFunctionalTracing::test_nn_functional_binary_cross_entropy, test/test_fx.py::TestFunctionalTracing::test_nn_functional_binary_cross_entropy_with_logits, test/test_fx.py::TestFunctionalTracing::test_nn_functional_celu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_celu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_channel_shuffle, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv_tbc, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_cosine_embedding_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_cosine_similarity, test/test_fx.py::TestFunctionalTracing::test_nn_functional_cross_entropy, test/test_fx.py::TestFunctionalTracing::test_nn_functional_ctc_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_dropout, test/test_fx.py::TestFunctionalTracing::test_nn_functional_dropout1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_dropout2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_dropout3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_elu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_elu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_embedding, test/test_fx.py::TestFunctionalTracing::test_nn_functional_embedding_bag, test/test_fx.py::TestFunctionalTracing::test_nn_functional_feature_alpha_dropout, test/test_fx.py::TestFunctionalTracing::test_nn_functional_fold, test/test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool2d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool3d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_gaussian_nll_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_gelu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_glu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_grid_sample, test/test_fx.py::TestFunctionalTracing::test_nn_functional_group_norm, test/test_fx.py::TestFunctionalTracing::test_nn_functional_gumbel_softmax, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hardshrink, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hardsigmoid, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hardswish, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hardtanh, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hardtanh_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_hinge_embedding_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_huber_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_instance_norm, test/test_fx.py::TestFunctionalTracing::test_nn_functional_interpolate, test/test_fx.py::TestFunctionalTracing::test_nn_functional_kl_div, test/test_fx.py::TestFunctionalTracing::test_nn_functional_l1_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_layer_norm, test/test_fx.py::TestFunctionalTracing::test_nn_functional_leaky_relu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_leaky_relu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_linear, test/test_fx.py::TestFunctionalTracing::test_nn_functional_local_response_norm, test/test_fx.py::TestFunctionalTracing::test_nn_functional_log_softmax, test/test_fx.py::TestFunctionalTracing::test_nn_functional_logsigmoid, test/test_fx.py::TestFunctionalTracing::test_nn_functional_lp_pool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_lp_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_margin_ranking_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool1d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool2d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool3d_with_indices, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool1d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool2d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool3d, test/test_fx.py::TestFunctionalTracing::test_nn_functional_mish, test/test_fx.py::TestFunctionalTracing::test_nn_functional_mse_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_multi_head_attention_forward, test/test_fx.py::TestFunctionalTracing::test_nn_functional_multi_margin_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_multilabel_margin_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_multilabel_soft_margin_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_native_channel_shuffle, test/test_fx.py::TestFunctionalTracing::test_nn_functional_nll_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_normalize, test/test_fx.py::TestFunctionalTracing::test_nn_functional_one_hot, test/test_fx.py::TestFunctionalTracing::test_nn_functional_pad, test/test_fx.py::TestFunctionalTracing::test_nn_functional_pairwise_distance, test/test_fx.py::TestFunctionalTracing::test_nn_functional_pdist, test/test_fx.py::TestFunctionalTracing::test_nn_functional_pixel_shuffle, test/test_fx.py::TestFunctionalTracing::test_nn_functional_pixel_unshuffle, test/test_fx.py::TestFunctionalTracing::test_nn_functional_poisson_nll_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_prelu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_relu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_relu6, test/test_fx.py::TestFunctionalTracing::test_nn_functional_relu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_rrelu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_rrelu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_scaled_dot_product_attention, test/test_fx.py::TestFunctionalTracing::test_nn_functional_selu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_selu_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_silu, test/test_fx.py::TestFunctionalTracing::test_nn_functional_smooth_l1_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_soft_margin_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_softmax, test/test_fx.py::TestFunctionalTracing::test_nn_functional_softmin, test/test_fx.py::TestFunctionalTracing::test_nn_functional_softplus, test/test_fx.py::TestFunctionalTracing::test_nn_functional_softshrink, test/test_fx.py::TestFunctionalTracing::test_nn_functional_threshold, test/test_fx.py::TestFunctionalTracing::test_nn_functional_threshold_, test/test_fx.py::TestFunctionalTracing::test_nn_functional_triplet_margin_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_triplet_margin_with_distance_loss, test/test_fx.py::TestFunctionalTracing::test_nn_functional_unfold, test/test_fx.py::TestFunctionalTracing::test_nn_functional_upsample, test/test_fx.py::TestFunctionalTracing::test_nn_functional_upsample_bilinear, test/test_fx.py::TestFunctionalTracing::test_nn_functional_upsample_nearest, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_H_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_T_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___getitem___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___radd___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rdiv___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmatmul___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmod___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmul___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rpow___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rsub___cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__native_batch_norm_legit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__segment_reduce_lengths_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__segment_reduce_offsets_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__softmax_backward_data_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__upsample_bilinear2d_aa_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_abs_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_acos_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_acosh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_add_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addbmm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addcdiv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addcmul_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmm_decomposed_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_all_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_allclose_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_amax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_amin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_aminmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_angle_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_any_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_arange_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argmin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argsort_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argwhere_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_partial_views_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_asin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_asinh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atan2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atan_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atanh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_baddbmm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bernoulli_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bfloat16_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_block_diag_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bmm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bool_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_shapes_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_tensors_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_to_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bucketize_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_byte_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cartesian_prod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cat_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cauchy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cdist_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cdouble_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ceil_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cfloat_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_chalf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_char_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_inverse_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_chunk_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_max_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_min_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clone_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_column_stack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_combinations_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_complex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_conj_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_conj_physical_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_constant_pad_nd_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_contiguous_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_copysign_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_corrcoef_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cos_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cosh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_count_nonzero_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cov_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cross_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cummax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cummin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumprod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumsum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumulative_trapezoid_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_deg2rad_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diag_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diag_embed_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagflat_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_copy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diff_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_digamma_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dist_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_floor_rounding_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_no_rounding_mode_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_trunc_rounding_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_double_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dsplit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dstack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_einsum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_permuted_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_eq_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_equal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erfc_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erfinv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exp2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expand_as_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expand_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expm1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exponential_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_eye_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fftshift_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifftshift_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfft2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfftn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fill_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flatten_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flip_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fliplr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flipud_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_float_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_float_power_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_floor_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_floor_divide_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_frac_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_frexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_full_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_full_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gather_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ge_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_geometric_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_geqrf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gradient_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_grid_sampler_2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gt_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_half_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_heaviside_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histc_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histogram_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histogramdd_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hsplit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hstack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hypot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_i0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_igamma_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_igammac_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_add_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_copy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_fill_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_put_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_reduce_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_select_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_inner_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_int_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isclose_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isfinite_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isinf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isnan_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isneginf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isposinf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isreal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_2inputs_2outputs_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_4inputs_with_extra_args_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_binary_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_binary_return_by_ref_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_unary_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_kron_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_kthvalue_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ldexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_le_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lerp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lgamma_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cholesky_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cholesky_ex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cond_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cross_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_det_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_det_singular_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_diagonal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eig_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigvals_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigvalsh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_householder_product_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_inv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_inv_ex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_factor_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_factor_ex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lstsq_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lstsq_grad_oriented_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_factor_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_factor_ex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_power_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_rank_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_rank_hermitian_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_multi_dot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_norm_subgradients_at_zero_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_hermitian_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_singular_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_qr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_slogdet_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_ex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_triangular_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_svd_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_svdvals_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_tensorinv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_tensorsolve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vander_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vecdot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vector_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linspace_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log10_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log1p_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_normal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_softmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_softmax_with_dtype_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logaddexp2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logaddexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logcumsumexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logdet_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_and_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_not_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_or_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_xor_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logspace_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logsumexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_long_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lt_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_unpack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mH_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mT_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_amax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_amin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_argmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_argmin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_cumprod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_cumsum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_fill_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_log_softmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_logaddexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_logsumexp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_median_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_normalize_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_prod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_select_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_softmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_softmin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_std_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_sum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_var_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_matmul_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_matrix_exp_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_binary_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_pool2d_with_indices_backward_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_reduction_no_dim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_reduction_with_dim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_maximum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_median_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_meshgrid_list_of_tensors_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_meshgrid_variadic_tensors_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_binary_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_reduction_no_dim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_reduction_with_dim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_minimum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mode_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_movedim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_msort_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mul_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_multinomial_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mv_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_3_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_5_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nan_to_num_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanmean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanmedian_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanquantile_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nansum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_narrow_copy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_narrow_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_batch_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_dropout_backward_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_layer_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ne_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_neg_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_empty_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_empty_strided_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_full_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_ones_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_zeros_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nextafter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_alpha_dropout_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_batch_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_bilinear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_binary_cross_entropy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_binary_cross_entropy_with_logits_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_celu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cosine_embedding_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cosine_similarity_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cross_entropy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_ctc_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_elu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_embedding_bag_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_embedding_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_feature_alpha_dropout_with_train_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_feature_alpha_dropout_without_train_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_fractional_max_pool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_fractional_max_pool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_gaussian_nll_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_gelu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_glu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_grid_sample_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_group_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardshrink_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardsigmoid_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardswish_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardtanh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hinge_embedding_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_huber_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_instance_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_area_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_bicubic_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_bilinear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_linear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_nearest_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_trilinear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_kl_div_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_l1_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_layer_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_leaky_relu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_linear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_local_response_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_logsigmoid_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_margin_ranking_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool1d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool1d_grad_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool2d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool2d_grad_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool3d_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool3d_grad_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_mish_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_mse_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multi_margin_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multilabel_margin_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multilabel_soft_margin_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_nll_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_normalize_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_circular_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_constant_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_reflect_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_replicate_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pairwise_distance_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pdist_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pixel_shuffle_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pixel_unshuffle_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_poisson_nll_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_prelu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_relu6_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_relu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_rrelu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_scaled_dot_product_attention_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_selu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_silu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_smooth_l1_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_soft_margin_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softmin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softmin_with_dtype_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softplus_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softshrink_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softsign_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_tanhshrink_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_threshold_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_triplet_margin_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_triplet_margin_with_distance_loss_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_unfold_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_upsample_bilinear_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_upsample_nearest_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nonzero_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_fro_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_inf_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_nuc_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_in_place_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_number_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ones_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ones_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ormqr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_outer_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pca_lowrank_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_permute_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pinverse_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polar_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_2_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_3_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_4_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_positive_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pow_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_prod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_put_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_qr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_quantile_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rad2deg_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rand_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randint_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randint_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randn_like_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ravel_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_real_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reciprocal_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_remainder_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_renorm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_repeat_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_repeat_interleave_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reshape_as_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reshape_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resize__cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resize_as__cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resolve_conj_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resolve_neg_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_roll_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rot90_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_3_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_neg_3_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rsqrt_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rsub_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scalar_tensor_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_add_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_amax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_amin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_prod_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_sum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_searchsorted_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_select_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_select_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sgn_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_short_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sigmoid_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sign_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_bartlett_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_blackman_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_cosine_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_exponential_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_gaussian_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_general_cosine_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_general_hamming_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_hamming_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_hann_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_kaiser_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_nuttall_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signbit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sin_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sinc_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sinh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_slice_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_slice_scatter_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_softmax_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_softmax_with_dtype_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sort_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sparse_mm_reduce_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sparse_sampled_addmm_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_airy_ai_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_j0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_j1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_y0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_y1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_t_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_u_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_v_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_w_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_entr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_erfcx_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_hermite_polynomial_h_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_hermite_polynomial_he_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i0e_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i1e_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_laguerre_polynomial_l_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_legendre_polynomial_p_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_log_ndtr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_i0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_i1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_k0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_k1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_ndtr_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_ndtri_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_polygamma_special_polygamma_n_0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_scaled_modified_bessel_k0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_scaled_modified_bessel_k1_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_t_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_u_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_v_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_w_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_spherical_bessel_j0_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_xlog1py_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_zeta_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_list_args_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_with_sizes_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sqrt_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_square_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_squeeze_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_squeeze_multiple_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_stack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_mean_unbiased_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_unbiased_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_stft_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sub_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sum_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sum_to_size_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_svd_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_svd_lowrank_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_t_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_take_along_dim_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_take_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tan_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tanh_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tensor_split_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tensordot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tile_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_to_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_to_sparse_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_topk_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trace_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_transpose_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trapezoid_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trapz_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_triangular_solve_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tril_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_triu_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_true_divide_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trunc_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unbind_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unflatten_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unfold_copy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unfold_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_uniform_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unique_consecutive_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unique_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unsafe_split_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unsqueeze_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_mean_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_mean_unbiased_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_unbiased_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vdot_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_as_complex_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_as_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_copy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vsplit_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vstack_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_where_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_xlogy_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zero__cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zeros_cpu_float32, test/test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zeros_like_cpu_float32, test/test_fx.py::TestVisionTracing::test_torchvision_models_alexnet, test/test_fx.py::TestVisionTracing::test_torchvision_models_convnext_base, test/test_fx.py::TestVisionTracing::test_torchvision_models_convnext_large, test/test_fx.py::TestVisionTracing::test_torchvision_models_convnext_small, test/test_fx.py::TestVisionTracing::test_torchvision_models_convnext_tiny, test/test_fx.py::TestVisionTracing::test_torchvision_models_densenet121, test/test_fx.py::TestVisionTracing::test_torchvision_models_densenet161, test/test_fx.py::TestVisionTracing::test_torchvision_models_densenet169, test/test_fx.py::TestVisionTracing::test_torchvision_models_densenet201, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_mobilenet_v3_large_320_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_mobilenet_v3_large_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_resnet50_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_resnet50_fpn_v2, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_fcos_resnet50_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_keypointrcnn_resnet50_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_maskrcnn_resnet50_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_maskrcnn_resnet50_fpn_v2, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_retinanet_resnet50_fpn, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_retinanet_resnet50_fpn_v2, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_ssd300_vgg16, test/test_fx.py::TestVisionTracing::test_torchvision_models_detection_ssdlite320_mobilenet_v3_large, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b0, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b1, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b2, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b3, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b4, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b5, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b6, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b7, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_l, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_m, test/test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_s, test/test_fx.py::TestVisionTracing::test_torchvision_models_googlenet, test/test_fx.py::TestVisionTracing::test_torchvision_models_inception_v3, test/test_fx.py::TestVisionTracing::test_torchvision_models_maxvit_t, test/test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet0_5, test/test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet0_75, test/test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet1_0, test/test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet1_3, test/test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v2, test/test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v3_large, test/test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v3_small, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_16gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_1_6gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_32gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_3_2gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_400mf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_800mf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_8gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_128gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_16gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_1_6gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_32gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_3_2gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_400mf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_800mf, test/test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_8gf, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnet101, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnet152, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnet18, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnet34, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnet50, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnext101_32x8d, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnext101_64x4d, test/test_fx.py::TestVisionTracing::test_torchvision_models_resnext50_32x4d, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_mobilenet_v3_large, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_resnet101, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_resnet50, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_fcn_resnet101, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_fcn_resnet50, test/test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_lraspp_mobilenet_v3_large, test/test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x0_5, test/test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x1_0, test/test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x1_5, test/test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x2_0, test/test_fx.py::TestVisionTracing::test_torchvision_models_squeezenet1_0, test/test_fx.py::TestVisionTracing::test_torchvision_models_squeezenet1_1, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_b, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_s, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_t, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_b, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_s, test/test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_t, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg11, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg11_bn, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg13, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg13_bn, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg16, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg16_bn, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg19, test/test_fx.py::TestVisionTracing::test_torchvision_models_vgg19_bn, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_mc3_18, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_mvit_v1_b, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_mvit_v2_s, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_r2plus1d_18, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_r3d_18, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_s3d, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_b, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_s, test/test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_t, test/test_fx.py::TestVisionTracing::test_torchvision_models_vit_b_16, test/test_fx.py::TestVisionTracing::test_torchvision_models_vit_b_32, test/test_fx.py::TestVisionTracing::test_torchvision_models_vit_h_14, test/test_fx.py::TestVisionTracing::test_torchvision_models_vit_l_16, test/test_fx.py::TestVisionTracing::test_torchvision_models_vit_l_32, test/test_fx.py::TestVisionTracing::test_torchvision_models_wide_resnet101_2, test/test_fx.py::TestVisionTracing::test_torchvision_models_wide_resnet50_2 2023-03-31T07:10:14.9621209Z 2023-03-31T07:10:14.9623515Z test_fx.py::TestSubgraphRewriter::test_matching_pattern_with_list_type_arg <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9624571Z test_fx.py::TestSubgraphRewriter::test_replace_pattern_with_filters <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9625874Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_annotations_int <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9626966Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_call_method <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9628016Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_correct_output_replacement <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9629088Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_graph_argument_order <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9630254Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_internal_pattern_nodes_cannot_have_users_that_are_not_matched <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9631312Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_local_revert <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9632249Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_multiple_pattern_match <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9633277Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_nodes_with_kwargs <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9634249Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_is_entire_graph <- test/fx/test_subgraph_rewriter.py PASSED [ 0%] 2023-03-31T07:10:14.9635406Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_output_pattern_node_can_have_users_that_are_not_matched <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9636284Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_placeholder_matching <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9637136Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_preserves_logic <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9638013Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_consecutive_submodules <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9638872Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_with_duplicated_outputs <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9639754Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replace_with_multiple_outputs <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9640677Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_replaces_referenced_submodules <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9641452Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_single_pattern_match <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9642338Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_traced_as_callable <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9643479Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_oneliner_pattern <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9644472Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_overlapping_matches <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9645260Z test_fx.py::TestSubgraphRewriter::test_subgraph_rewriter_with_unused_args <- test/fx/test_subgraph_rewriter.py PASSED [ 1%] 2023-03-31T07:10:14.9645854Z test_fx.py::TestDCE::test_dead_chain <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9646366Z test_fx.py::TestDCE::test_dead_getattr <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9646903Z test_fx.py::TestDCE::test_dead_placeholder <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9647471Z test_fx.py::TestDCE::test_dead_placeholder_with_user <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9648052Z test_fx.py::TestDCE::test_keep_module_with_side_effects <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9648587Z test_fx.py::TestDCE::test_keep_torch_assert <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9649259Z test_fx.py::TestDCE::test_simple <- test/fx/test_dce_pass.py PASSED [ 2%] 2023-03-31T07:10:14.9649901Z test_fx.py::TestConstFold::test_check_inline_non_const <- test/fx/test_fx_const_fold.py PASSED [ 2%] 2023-03-31T07:10:14.9650519Z test_fx.py::TestConstFold::test_check_inline_non_const_mult_return <- test/fx/test_fx_const_fold.py PASSED [ 2%] 2023-03-31T07:10:14.9651199Z test_fx.py::TestConstFold::test_check_skip_folding_quant_dequant_pattern <- test/fx/test_fx_const_fold.py PASSED [ 2%] 2023-03-31T07:10:14.9651872Z test_fx.py::TestConstFold::test_const_fold_basic_one_attr_name_collision <- test/fx/test_fx_const_fold.py PASSED [ 2%] 2023-03-31T07:10:14.9652542Z test_fx.py::TestConstFold::test_const_fold_basic_one_attr_no_name_collision <- test/fx/test_fx_const_fold.py PASSED [ 2%] 2023-03-31T07:10:14.9653190Z test_fx.py::TestConstFold::test_const_fold_basic_placeholder_reordered <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9653832Z test_fx.py::TestConstFold::test_const_fold_basic_two_attr <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9654474Z test_fx.py::TestConstFold::test_const_fold_basic_two_attr_three_input <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9655128Z test_fx.py::TestConstFold::test_const_fold_has_inlined_call_module_node <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9655811Z test_fx.py::TestConstFold::test_const_fold_module_attr <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9656453Z test_fx.py::TestConstFold::test_const_fold_multi_const_folded_attrs <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9657048Z test_fx.py::TestConstFold::test_const_fold_noop <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9657650Z test_fx.py::TestConstFold::test_const_fold_submod_hierarchy <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9658244Z test_fx.py::TestConstFold::test_const_fold_tensor_meta <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9658867Z test_fx.py::TestConstFold::test_const_fold_unused_placeholder <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9659471Z test_fx.py::TestConstFold::test_dict_output <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9660090Z test_fx.py::TestConstFold::test_fold_module <- test/fx/test_fx_const_fold.py PASSED [ 3%] 2023-03-31T07:10:14.9660662Z test_fx.py::TestConstFold::test_retain_node_meta <- test/fx/test_fx_const_fold.py PASSED [ 4%] 2023-03-31T07:10:14.9661235Z test_fx.py::TestConstFold::test_three_outputs <- test/fx/test_fx_const_fold.py PASSED [ 4%] 2023-03-31T07:10:14.9661796Z test_fx.py::TestConstFold::test_two_outputs <- test/fx/test_fx_const_fold.py PASSED [ 4%] 2023-03-31T07:10:14.9662452Z test_fx.py::TestConstParamShapeInControlFlow::test_param_dim_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9663209Z test_fx.py::TestConstParamShapeInControlFlow::test_param_ndim_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9664070Z test_fx.py::TestConstParamShapeInControlFlow::test_param_nelement_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9664837Z test_fx.py::TestConstParamShapeInControlFlow::test_param_numel_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9665627Z test_fx.py::TestConstParamShapeInControlFlow::test_param_shape_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9666398Z test_fx.py::TestConstParamShapeInControlFlow::test_param_size_const <- test/fx/test_fx_param_shape_control_flow.py PASSED [ 4%] 2023-03-31T07:10:14.9667046Z test_fx.py::TestPassManager::test_pass_manager <- test/fx/test_pass_infra.py PASSED [ 4%] 2023-03-31T07:10:14.9667662Z test_fx.py::TestPassManager::test_pass_manager_bad_checks <- test/fx/test_pass_infra.py PASSED [ 4%] 2023-03-31T07:10:14.9668261Z test_fx.py::TestPassManager::test_pass_manager_checks <- test/fx/test_pass_infra.py PASSED [ 4%] 2023-03-31T07:10:14.9668961Z test_fx.py::TestPassManager::test_pass_manager_error <- test/fx/test_pass_infra.py PASSED [ 5%] 2023-03-31T07:10:14.9669649Z test_fx.py::TestPassManager::test_this_before_that_pass_constraint <- test/fx/test_pass_infra.py PASSED [ 5%] 2023-03-31T07:10:14.9670265Z test_fx.py::TestPassManager::test_topological_sort <- test/fx/test_pass_infra.py PASSED [ 5%] 2023-03-31T07:10:14.9670903Z test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationInput_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9671605Z test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationMetadata_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9672323Z test_fx.py::TestCommonPass::test_correctness_CSEPass_MutationTorchTensorCall_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9673005Z test_fx.py::TestCommonPass::test_correctness_CSEPass_Mutation_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9673671Z test_fx.py::TestCommonPass::test_correctness_CSEPass_ReturnList_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9674339Z test_fx.py::TestCommonPass::test_correctness_CSEPass_TakeList_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9675054Z test_fx.py::TestCommonPass::test_correctness_factory_CSEPass_FactoryFunctionCall_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9675822Z test_fx.py::TestCommonPass::test_correctness_factory_CSEPass_MutationFactory_cpu <- test/fx/test_common_passes.py PASSED [ 5%] 2023-03-31T07:10:14.9676443Z test_fx.py::TestCSEPass::test_banned_list <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9676966Z test_fx.py::TestCSEPass::test_empty <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9677546Z test_fx.py::TestCSEPass::test_immutable_list_multiple_entries <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9678114Z test_fx.py::TestCSEPass::test_immutable_list_type <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9678666Z test_fx.py::TestCSEPass::test_kwarg <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9679261Z test_fx.py::TestCSEPass::test_nested_immutable_list_type <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9679850Z test_fx.py::TestCSEPass::test_nochange <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9680391Z test_fx.py::TestCSEPass::test_rand_like <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9680908Z test_fx.py::TestCSEPass::test_rand_n <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9681435Z test_fx.py::TestCSEPass::test_random <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9681938Z test_fx.py::TestCSEPass::test_simple <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9682460Z test_fx.py::TestCSEPass::test_simple_2 <- test/fx/test_cse_pass.py PASSED [ 6%] 2023-03-31T07:10:14.9683268Z test_fx.py::TestCSEPass::test_simple_multiple_same_ops <- test/fx/test_cse_pass.py PASSED [ 7%] 2023-03-31T07:10:14.9683884Z test_fx.py::TestCSEPass::test_two_args <- test/fx/test_cse_pass.py PASSED [ 7%] 2023-03-31T07:10:14.9684458Z test_fx.py::TestCSEPass::test_two_args_default <- test/fx/test_cse_pass.py PASSED [ 7%] 2023-03-31T07:10:14.9685113Z test_fx.py::TestMatcher::test_subgraph_matcher_with_attributes <- test/fx/test_matcher_utils.py PASSED [ 7%] 2023-03-31T07:10:14.9685804Z test_fx.py::TestMatcher::test_subgraph_matcher_with_list <- test/fx/test_matcher_utils.py PASSED [ 7%] 2023-03-31T07:10:14.9686448Z test_fx.py::TestMatcher::test_subgraph_matcher_with_list_bad <- test/fx/test_matcher_utils.py PASSED [ 7%] 2023-03-31T07:10:14.9687204Z test_fx.py::VerifierTest::test_aten_dialect <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9687997Z test_fx.py::VerifierTest::test_aten_wrong_mem_format <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9688811Z test_fx.py::VerifierTest::test_aten_wrong_mem_format_buffer <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9689487Z test_fx.py::VerifierTest::test_verifier <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9690142Z test_fx.py::VerifierTest::test_verifier_no_functional <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9690834Z test_fx.py::VerifierTest::testr_verifier_call_module <- test/fx/test_verifier.py PASSED [ 7%] 2023-03-31T07:10:14.9691444Z test_fx.py::AnnotationsTest::test_annotate <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9692072Z test_fx.py::AnnotationsTest::test_annotations <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9692705Z test_fx.py::AnnotationsTest::test_broadcasting1 <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9693346Z test_fx.py::AnnotationsTest::test_broadcasting2 <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9693978Z test_fx.py::AnnotationsTest::test_broadcasting3 <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9694622Z test_fx.py::AnnotationsTest::test_consistency <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9695204Z test_fx.py::AnnotationsTest::test_precision <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9695805Z test_fx.py::TypeCheckerTest::test_flatten_fully_static <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9696526Z test_fx.py::TypeCheckerTest::test_resnet50 <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9697218Z test_fx.py::TypeCheckerTest::test_symbolic_add_with_broadcast <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9697927Z test_fx.py::TypeCheckerTest::test_symbolic_add_with_broadcast_2 <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9698606Z test_fx.py::TypeCheckerTest::test_type_check_add_false <- test/fx/test_gradual_type.py PASSED [ 8%] 2023-03-31T07:10:14.9699258Z test_fx.py::TypeCheckerTest::test_type_check_add_true <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9699893Z test_fx.py::TypeCheckerTest::test_type_check_add_with_broadcast <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9700532Z test_fx.py::TypeCheckerTest::test_type_check_add_with_scalar <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9701245Z test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9701918Z test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D_broadcast <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9702572Z test_fx.py::TypeCheckerTest::test_type_check_batch_norm_2D_false <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9703205Z test_fx.py::TypeCheckerTest::test_type_check_batch_norm_symbolic <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9703906Z test_fx.py::TypeCheckerTest::test_type_check_conv2D <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9704528Z test_fx.py::TypeCheckerTest::test_type_check_conv2D_2 <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9705177Z test_fx.py::TypeCheckerTest::test_type_check_conv2D_2_fully_static <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9705919Z test_fx.py::TypeCheckerTest::test_type_check_conv2D_maxpool2d_flatten <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9706656Z test_fx.py::TypeCheckerTest::test_type_check_conv2D_types <- test/fx/test_gradual_type.py PASSED [ 9%] 2023-03-31T07:10:14.9707283Z test_fx.py::TypeCheckerTest::test_type_check_flatten <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9707939Z test_fx.py::TypeCheckerTest::test_type_check_flatten3 <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9708600Z test_fx.py::TypeCheckerTest::test_type_check_flatten_2 <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9709336Z test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_false <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9710197Z test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_true <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9711121Z test_fx.py::TypeCheckerTest::test_type_check_reshape_dyn_true_param_false <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9711799Z test_fx.py::TypeCheckerTest::test_type_check_reshape_false <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9712528Z test_fx.py::TypeCheckerTest::test_type_check_reshape_true <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9713310Z test_fx.py::TypeCheckerTest::test_type_check_symbolic_inferenceconv2D_maxpool2d_flatten <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9714186Z test_fx.py::TypeCheckerTest::test_type_check_transpose_False <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9715000Z test_fx.py::TypeCheckerTest::test_type_check_transpose_true <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9715792Z test_fx.py::TypeCheckerTest::test_type_maxpool2d_fully_static <- test/fx/test_gradual_type.py PASSED [ 10%] 2023-03-31T07:10:14.9716481Z test_fx.py::TypeCheckerTest::test_type_typechecl_maxpool2d_3dinput <- test/fx/test_gradual_type.py PASSED [ 11%] 2023-03-31T07:10:14.9717174Z test_fx.py::TypeCheckerTest::test_typecheck_basicblock <- test/fx/test_gradual_type.py PASSED [ 11%] 2023-03-31T07:10:14.9717664Z test_fx.py::TestFX::test_all_input_nodes PASSED [ 11%] 2023-03-31T07:10:14.9718204Z test_fx.py::TestFX::test_annotation_with_future PASSED [ 11%] 2023-03-31T07:10:14.9718732Z test_fx.py::TestFX::test_annotations_empty_tuple PASSED [ 11%] 2023-03-31T07:10:14.9719175Z test_fx.py::TestFX::test_annotations_with_forward_references PASSED [ 11%] 2023-03-31T07:10:14.9719616Z test_fx.py::TestFX::test_annotations_with_no_forward_references PASSED [ 11%] 2023-03-31T07:10:14.9719993Z test_fx.py::TestFX::test_annotations_with_non_torch_reference_and_internal_forward_references PASSED [ 11%] 2023-03-31T07:10:14.9720401Z test_fx.py::TestFX::test_annotations_with_non_torch_reference_and_no_internal_forward_references PASSED [ 11%] 2023-03-31T07:10:14.9720792Z test_fx.py::TestFX::test_args_kwargs PASSED [ 11%] 2023-03-31T07:10:14.9721210Z test_fx.py::TestFX::test_args_kwargs_no_self PASSED [ 11%] 2023-03-31T07:10:14.9721509Z test_fx.py::TestFX::test_assert PASSED [ 12%] 2023-03-31T07:10:14.9721799Z test_fx.py::TestFX::test_ast_rewriter_reassigns_submodules PASSED [ 12%] 2023-03-31T07:10:14.9722111Z test_fx.py::TestFX::test_ast_rewriter_rewrites_assert PASSED [ 12%] 2023-03-31T07:10:14.9722437Z test_fx.py::TestFX::test_ast_rewriter_rewrites_assert_with_message PASSED [ 12%] 2023-03-31T07:10:14.9722754Z test_fx.py::TestFX::test_ast_rewriter_wrap PASSED [ 12%] 2023-03-31T07:10:14.9723215Z test_fx.py::TestFX::test_ast_rewriter_wrap_fn_directly PASSED [ 12%] 2023-03-31T07:10:14.9723565Z test_fx.py::TestFX::test_ast_rewriter_wrap_with_submodule PASSED [ 12%] 2023-03-31T07:10:14.9723884Z test_fx.py::TestFX::test_ast_rewriter_wrapped_via_decorator PASSED [ 12%] 2023-03-31T07:10:14.9724218Z test_fx.py::TestFX::test_ast_rewriter_wrapped_via_decorator_and_transformed PASSED [ 12%] 2023-03-31T07:10:14.9724560Z test_fx.py::TestFX::test_autowrap_functions PASSED [ 12%] 2023-03-31T07:10:14.9724870Z test_fx.py::TestFX::test_concrete_arg_none_assert PASSED [ 12%] 2023-03-31T07:10:14.9725174Z test_fx.py::TestFX::test_construct_root_dict PASSED [ 12%] 2023-03-31T07:10:14.9725458Z test_fx.py::TestFX::test_control_flow_tracing PASSED [ 13%] 2023-03-31T07:10:14.9725744Z test_fx.py::TestFX::test_copy_it PASSED [ 13%] 2023-03-31T07:10:14.9726027Z test_fx.py::TestFX::test_copy_no_remap PASSED [ 13%] 2023-03-31T07:10:14.9726302Z test_fx.py::TestFX::test_ctx_mgr PASSED [ 13%] 2023-03-31T07:10:14.9726701Z test_fx.py::TestFX::test_custom_codegen PASSED [ 13%] 2023-03-31T07:10:14.9727058Z test_fx.py::TestFX::test_custom_codegen_with_transformer PASSED [ 13%] 2023-03-31T07:10:14.9727361Z test_fx.py::TestFX::test_custom_import PASSED [ 13%] 2023-03-31T07:10:14.9727642Z test_fx.py::TestFX::test_custom_proxy_dynamic_value PASSED [ 13%] 2023-03-31T07:10:14.9727963Z test_fx.py::TestFX::test_custom_proxy_input_dependent_control_flow PASSED [ 13%] 2023-03-31T07:10:14.9728278Z test_fx.py::TestFX::test_custom_proxy_type PASSED [ 13%] 2023-03-31T07:10:14.9728562Z test_fx.py::TestFX::test_custom_proxy_type_literal PASSED [ 13%] 2023-03-31T07:10:14.9728913Z test_fx.py::TestFX::test_custom_traceback_not_raised_when_exception_source_is_submodule PASSED [ 13%] 2023-03-31T07:10:14.9729300Z test_fx.py::TestFX::test_custom_traceback_raised_when_exception_source_is_graphmodule PASSED [ 14%] 2023-03-31T07:10:14.9729655Z test_fx.py::TestFX::test_deepcopy_graph_with_tracer_cls PASSED [ 14%] 2023-03-31T07:10:14.9729958Z test_fx.py::TestFX::test_deepcopy_graphmodule PASSED [ 14%] 2023-03-31T07:10:14.9730276Z test_fx.py::TestFX::test_deepcopy_graphmodule_with_transform PASSED [ 14%] 2023-03-31T07:10:14.9730653Z test_fx.py::TestFX::test_deepcopy_no_recursion PASSED [ 14%] 2023-03-31T07:10:14.9730945Z test_fx.py::TestFX::test_deepcopy_recursion_depth PASSED [ 14%] 2023-03-31T07:10:14.9731242Z test_fx.py::TestFX::test_deepcopy_tracer PASSED [ 14%] 2023-03-31T07:10:14.9731618Z test_fx.py::TestFX::test_deepcopy_with_submods_params PASSED [ 14%] 2023-03-31T07:10:14.9732049Z test_fx.py::TestFX::test_delete_unused_submodules_leaf PASSED [ 14%] 2023-03-31T07:10:14.9732329Z test_fx.py::TestFX::test_dict PASSED [ 14%] 2023-03-31T07:10:14.9732613Z test_fx.py::TestFX::test_direct_param_use PASSED [ 14%] 2023-03-31T07:10:14.9732906Z test_fx.py::TestFX::test_disallow_override PASSED [ 14%] 2023-03-31T07:10:14.9733228Z test_fx.py::TestFX::test_ellipsis PASSED [ 15%] 2023-03-31T07:10:14.9733513Z test_fx.py::TestFX::test_empty_graph_codegen PASSED [ 15%] 2023-03-31T07:10:14.9733808Z test_fx.py::TestFX::test_erase_node_error PASSED [ 15%] 2023-03-31T07:10:14.9734100Z test_fx.py::TestFX::test_example_shape_prop PASSED [ 15%] 2023-03-31T07:10:14.9734375Z test_fx.py::TestFX::test_find_uses PASSED [ 15%] 2023-03-31T07:10:14.9734665Z test_fx.py::TestFX::test_fn_type_annotation_empty PASSED [ 15%] 2023-03-31T07:10:14.9734964Z test_fx.py::TestFX::test_fn_type_annotations PASSED [ 15%] 2023-03-31T07:10:14.9735238Z test_fx.py::TestFX::test_fx_and_or PASSED [ 15%] 2023-03-31T07:10:14.9735519Z test_fx.py::TestFX::test_fx_create_arg PASSED [ 15%] 2023-03-31T07:10:14.9735804Z test_fx.py::TestFX::test_fx_shifts PASSED [ 15%] 2023-03-31T07:10:14.9736093Z test_fx.py::TestFX::test_fx_stateless PASSED [ 15%] 2023-03-31T07:10:14.9736376Z test_fx.py::TestFX::test_get_torch_func_signature PASSED [ 15%] 2023-03-31T07:10:14.9736703Z test_fx.py::TestFX::test_getitem SKIPPED (Will be checked in test_getitem_subproc) [ 16%] 2023-03-31T07:10:14.9737030Z test_fx.py::TestFX::test_getitem_subproc PASSED [ 16%] 2023-03-31T07:10:14.9737315Z test_fx.py::TestFX::test_graph_edit_with_proxy PASSED [ 16%] 2023-03-31T07:10:14.9737605Z test_fx.py::TestFX::test_graph_fns PASSED [ 16%] 2023-03-31T07:10:14.9737892Z test_fx.py::TestFX::test_graph_module PASSED [ 16%] 2023-03-31T07:10:14.9738243Z test_fx.py::TestFX::test_graph_module_init_buffer_param_copied_dict_init PASSED [ 16%] 2023-03-31T07:10:14.9738606Z test_fx.py::TestFX::test_graph_module_init_buffer_param_copied_mod_init PASSED [ 16%] 2023-03-31T07:10:14.9738941Z test_fx.py::TestFX::test_graph_module_replicate_for_dp PASSED [ 16%] 2023-03-31T07:10:14.9739249Z test_fx.py::TestFX::test_graph_unique_names PASSED [ 16%] 2023-03-31T07:10:14.9739533Z test_fx.py::TestFX::test_graph_unique_names_manual PASSED [ 16%] 2023-03-31T07:10:14.9739833Z test_fx.py::TestFX::test_immutable_dict_pytree_ops PASSED [ 16%] 2023-03-31T07:10:14.9740137Z test_fx.py::TestFX::test_immutable_list_pytree_ops PASSED [ 17%] 2023-03-31T07:10:14.9740434Z test_fx.py::TestFX::test_imul_code_print PASSED [ 17%] 2023-03-31T07:10:14.9740702Z test_fx.py::TestFX::test_inf_nan PASSED [ 17%] 2023-03-31T07:10:14.9740981Z test_fx.py::TestFX::test_inf_nan_kwds PASSED [ 17%] 2023-03-31T07:10:14.9741265Z test_fx.py::TestFX::test_inline_graph PASSED [ 17%] 2023-03-31T07:10:14.9741540Z test_fx.py::TestFX::test_insertion_point PASSED [ 17%] 2023-03-31T07:10:14.9741824Z test_fx.py::TestFX::test_interpreter PASSED [ 17%] 2023-03-31T07:10:14.9742151Z test_fx.py::TestFX::test_interpreter_default_args PASSED [ 17%] 2023-03-31T07:10:14.9742455Z test_fx.py::TestFX::test_interpreter_gc_values PASSED [ 17%] 2023-03-31T07:10:14.9742742Z test_fx.py::TestFX::test_interpreter_noop_resnet18 PASSED [ 17%] 2023-03-31T07:10:14.9743048Z test_fx.py::TestFX::test_interpreter_not_enough_args PASSED [ 17%] 2023-03-31T07:10:14.9743353Z test_fx.py::TestFX::test_interpreter_onthefly_swap PASSED [ 17%] 2023-03-31T07:10:14.9743742Z test_fx.py::TestFX::test_interpreter_partial_eval PASSED [ 18%] 2023-03-31T07:10:14.9744050Z test_fx.py::TestFX::test_interpreter_run_node_override PASSED [ 18%] 2023-03-31T07:10:14.9744356Z test_fx.py::TestFX::test_interpreter_star_args PASSED [ 18%] 2023-03-31T07:10:14.9744692Z test_fx.py::TestFX::test_interpreter_with_codegen PASSED [ 18%] 2023-03-31T07:10:14.9744968Z test_fx.py::TestFX::test_layout PASSED [ 18%] 2023-03-31T07:10:14.9745251Z test_fx.py::TestFX::test_leaf_module PASSED [ 18%] 2023-03-31T07:10:14.9745539Z test_fx.py::TestFX::test_matmul_tracing PASSED [ 18%] 2023-03-31T07:10:14.9745825Z test_fx.py::TestFX::test_module_deepcopy_edit_nodes PASSED [ 18%] 2023-03-31T07:10:14.9746119Z test_fx.py::TestFX::test_move_before PASSED [ 18%] 2023-03-31T07:10:14.9746408Z test_fx.py::TestFX::test_multi_insert_point PASSED [ 18%] 2023-03-31T07:10:14.9746706Z test_fx.py::TestFX::test_multiple_default_args PASSED [ 18%] 2023-03-31T07:10:14.9746993Z test_fx.py::TestFX::test_named_tuple_inlined PASSED [ 18%] 2023-03-31T07:10:14.9747298Z test_fx.py::TestFX::test_namedtuple_return_qualname PASSED [ 19%] 2023-03-31T07:10:14.9747605Z test_fx.py::TestFX::test_namedtuple_return_trace PASSED [ 19%] 2023-03-31T07:10:14.9747891Z test_fx.py::TestFX::test_native_callable PASSED [ 19%] 2023-03-31T07:10:14.9748182Z test_fx.py::TestFX::test_nn_module_stack PASSED [ 19%] 2023-03-31T07:10:14.9748466Z test_fx.py::TestFX::test_no_mutation PASSED [ 19%] 2023-03-31T07:10:14.9748750Z test_fx.py::TestFX::test_node_tagging PASSED [ 19%] 2023-03-31T07:10:14.9749024Z test_fx.py::TestFX::test_nonetype_annotation PASSED [ 19%] 2023-03-31T07:10:14.9749348Z test_fx.py::TestFX::test_partial_trace PASSED [ 19%] 2023-03-31T07:10:14.9749667Z test_fx.py::TestFX::test_pickle_custom_import PASSED [ 19%] 2023-03-31T07:10:14.9749950Z test_fx.py::TestFX::test_pickle_graphmodule PASSED [ 19%] 2023-03-31T07:10:14.9750253Z test_fx.py::TestFX::test_pickle_nonetype_annotation PASSED [ 19%] 2023-03-31T07:10:14.9750557Z test_fx.py::TestFX::test_pickle_torch_custom_ops PASSED [ 19%] 2023-03-31T07:10:14.9750832Z test_fx.py::TestFX::test_prepend_self PASSED [ 20%] 2023-03-31T07:10:14.9751119Z test_fx.py::TestFX::test_pretty_print PASSED [ 20%] 2023-03-31T07:10:14.9751407Z test_fx.py::TestFX::test_pretty_print_graph PASSED [ 20%] 2023-03-31T07:10:14.9751697Z test_fx.py::TestFX::test_pretty_print_node PASSED [ 20%] 2023-03-31T07:10:14.9751979Z test_fx.py::TestFX::test_pretty_print_targets PASSED [ 20%] 2023-03-31T07:10:14.9752281Z test_fx.py::TestFX::test_profiler_ranges_side_effect PASSED [ 20%] 2023-03-31T07:10:14.9752569Z test_fx.py::TestFX::test_pytree PASSED [ 20%] 2023-03-31T07:10:14.9752842Z test_fx.py::TestFX::test_pytree_concrete PASSED [ 20%] 2023-03-31T07:10:14.9753174Z test_fx.py::TestFX::test_reassign_args_kwargs_uses PASSED [ 20%] 2023-03-31T07:10:14.9753476Z test_fx.py::TestFX::test_regular_and_default_args PASSED [ 20%] 2023-03-31T07:10:14.9753769Z test_fx.py::TestFX::test_remove_uses PASSED [ 20%] 2023-03-31T07:10:14.9754045Z test_fx.py::TestFX::test_remove_uses_with_custom_filter PASSED [ 20%] 2023-03-31T07:10:14.9754335Z test_fx.py::TestFX::test_replace_input PASSED [ 21%] 2023-03-31T07:10:14.9754625Z test_fx.py::TestFX::test_replace_uses PASSED [ 21%] 2023-03-31T07:10:14.9754911Z test_fx.py::TestFX::test_reserved_getattr PASSED [ 21%] 2023-03-31T07:10:14.9755208Z test_fx.py::TestFX::test_return_tuple PASSED [ 21%] 2023-03-31T07:10:14.9755530Z test_fx.py::TestFX::test_return_type_exists PASSED [ 21%] 2023-03-31T07:10:14.9755821Z test_fx.py::TestFX::test_script_method_trace PASSED [ 21%] 2023-03-31T07:10:14.9756106Z test_fx.py::TestFX::test_script_tensor_constant PASSED [ 21%] 2023-03-31T07:10:14.9756396Z test_fx.py::TestFX::test_sequential PASSED [ 21%] 2023-03-31T07:10:14.9756687Z test_fx.py::TestFX::test_shape_prop_aggregate PASSED [ 21%] 2023-03-31T07:10:14.9756969Z test_fx.py::TestFX::test_shape_prop_layout PASSED [ 21%] 2023-03-31T07:10:14.9757268Z test_fx.py::TestFX::test_shape_prop_layout_3d PASSED [ 21%] 2023-03-31T07:10:14.9757565Z test_fx.py::TestFX::test_single_default_arg PASSED [ 21%] 2023-03-31T07:10:14.9757855Z test_fx.py::TestFX::test_snake_case PASSED [ 22%] 2023-03-31T07:10:14.9758118Z test_fx.py::TestFX::test_sqrt PASSED [ 22%] 2023-03-31T07:10:14.9758397Z test_fx.py::TestFX::test_stack_traces PASSED [ 22%] 2023-03-31T07:10:14.9758697Z test_fx.py::TestFX::test_stack_traces_with_transformer PASSED [ 22%] 2023-03-31T07:10:14.9758986Z test_fx.py::TestFX::test_string_literal_return PASSED [ 22%] 2023-03-31T07:10:14.9759295Z test_fx.py::TestFX::test_submodule_manipulation_API PASSED [ 22%] 2023-03-31T07:10:14.9759595Z test_fx.py::TestFX::test_symbolic_trace_assert PASSED [ 22%] 2023-03-31T07:10:14.9759895Z test_fx.py::TestFX::test_symbolic_trace_sequential PASSED [ 22%] 2023-03-31T07:10:14.9760181Z test_fx.py::TestFX::test_tensor_attribute PASSED [ 22%] 2023-03-31T07:10:14.9760512Z test_fx.py::TestFX::test_tensor_attribute_coalseced PASSED [ 22%] 2023-03-31T07:10:14.9760835Z test_fx.py::TestFX::test_tensor_constant PASSED [ 22%] 2023-03-31T07:10:14.9761111Z test_fx.py::TestFX::test_throw_out_variant PASSED [ 23%] 2023-03-31T07:10:14.9761402Z test_fx.py::TestFX::test_torch_custom_ops PASSED [ 23%] 2023-03-31T07:10:14.9761689Z test_fx.py::TestFX::test_torch_fx_getattr PASSED [ 23%] 2023-03-31T07:10:14.9761977Z test_fx.py::TestFX::test_torch_fx_len PASSED [ 23%] 2023-03-31T07:10:14.9762248Z test_fx.py::TestFX::test_torch_op_overloads PASSED [ 23%] 2023-03-31T07:10:14.9762545Z test_fx.py::TestFX::test_torchbind_class_attribute_in_fx PASSED [ 23%] 2023-03-31T07:10:14.9762870Z test_fx.py::TestFX::test_torchbind_class_attribute_in_fx_tensor_arg PASSED [ 23%] 2023-03-31T07:10:14.9763406Z test_fx.py::TestFX::test_trace_buffer_slice SKIPPED (Hotfix for SEV remediation) [ 23%] 2023-03-31T07:10:14.9763737Z test_fx.py::TestFX::test_trace_dict_int_keys PASSED [ 23%] 2023-03-31T07:10:14.9764038Z test_fx.py::TestFX::test_trace_dict_proxy_keys PASSED [ 23%] 2023-03-31T07:10:14.9764332Z test_fx.py::TestFX::test_trace_fn_constant PASSED [ 23%] 2023-03-31T07:10:14.9764672Z test_fx.py::TestFX::test_trace_function PASSED [ 23%] 2023-03-31T07:10:14.9764963Z test_fx.py::TestFX::test_trace_multiple_funcs PASSED [ 24%] 2023-03-31T07:10:14.9765277Z test_fx.py::TestFX::test_tracing_graphmodules_as_leaf_submodules PASSED [ 24%] 2023-03-31T07:10:14.9765584Z test_fx.py::TestFX::test_transformer_multi_outputs PASSED [ 24%] 2023-03-31T07:10:14.9765888Z test_fx.py::TestFX::test_transformer_noop PASSED [ 24%] 2023-03-31T07:10:14.9766181Z test_fx.py::TestFX::test_transformer_op_swap PASSED [ 24%] 2023-03-31T07:10:14.9766514Z test_fx.py::TestFX::test_transformer_preserves_nn_module_stack_for_get_attr PASSED [ 24%] 2023-03-31T07:10:14.9766833Z test_fx.py::TestFX::test_tuple_no_subscript PASSED [ 24%] 2023-03-31T07:10:14.9767181Z test_fx.py::TestFX::test_typename_print PASSED [ 24%] 2023-03-31T07:10:14.9767470Z test_fx.py::TestFX::test_unpack PASSED [ 24%] 2023-03-31T07:10:14.9767745Z test_fx.py::TestFX::test_unpack_dict_better_error PASSED [ 24%] 2023-03-31T07:10:14.9768036Z test_fx.py::TestFX::test_unpack_list_better_error PASSED [ 24%] 2023-03-31T07:10:14.9768326Z test_fx.py::TestFX::test_update_args_api PASSED [ 24%] 2023-03-31T07:10:14.9768610Z test_fx.py::TestFX::test_update_args_kwargs_yells_at_you PASSED [ 25%] 2023-03-31T07:10:14.9768908Z test_fx.py::TestFX::test_update_kwargs_api PASSED [ 25%] 2023-03-31T07:10:14.9769230Z test_fx.py::TestFX::test_user_friendly_call_provenance_with_function PASSED [ 25%] 2023-03-31T07:10:14.9769571Z test_fx.py::TestFX::test_user_friendly_call_provenance_with_module PASSED [ 25%] 2023-03-31T07:10:14.9769867Z test_fx.py::TestFX::test_wrap PASSED [ 25%] 2023-03-31T07:10:14.9770164Z test_fx.py::TestFX::test_wrap_decorated_function PASSED [ 25%] 2023-03-31T07:10:14.9770463Z test_fx.py::TestFX::test_wrap_fn_directly PASSED [ 25%] 2023-03-31T07:10:14.9770745Z test_fx.py::TestFX::test_wrap_with_submodule PASSED [ 25%] 2023-03-31T07:10:14.9771036Z test_fx.py::TestFX::test_wrapped_method PASSED [ 25%] 2023-03-31T07:10:14.9771326Z test_fx.py::TestFX::test_wrapped_retrace PASSED [ 25%] 2023-03-31T07:10:14.9771626Z test_fx.py::TestFX::test_wrapped_via_decorator PASSED [ 25%] 2023-03-31T07:10:14.9771978Z test_fx.py::TestFX::test_wrapped_via_decorator_and_transformed PASSED [ 25%] 2023-03-31T07:10:14.9772319Z test_fx.py::TestFX::test_wrong_target_type PASSED [ 26%] 2023-03-31T07:10:14.9772606Z test_fx.py::TestFX::test_wrong_topo PASSED [ 26%] 2023-03-31T07:10:14.9772937Z test_fx.py::TestFXAPIBackwardCompatibility::test_class_member_back_compat PASSED [ 26%] 2023-03-31T07:10:14.9773332Z test_fx.py::TestFXAPIBackwardCompatibility::test_function_back_compat PASSED [ 26%] 2023-03-31T07:10:14.9773714Z test_fx.py::TestFXAPIBackwardCompatibility::test_public_api_surface PASSED [ 26%] 2023-03-31T07:10:14.9774086Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool1d PASSED [ 26%] 2023-03-31T07:10:14.9774438Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool2d PASSED [ 26%] 2023-03-31T07:10:14.9774803Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_avg_pool3d PASSED [ 26%] 2023-03-31T07:10:14.9775170Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool1d PASSED [ 26%] 2023-03-31T07:10:14.9775543Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool1d_with_indices PASSED [ 26%] 2023-03-31T07:10:14.9775929Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool2d PASSED [ 26%] 2023-03-31T07:10:14.9776345Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool2d_with_indices PASSED [ 26%] 2023-03-31T07:10:14.9776726Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool3d PASSED [ 27%] 2023-03-31T07:10:14.9777094Z test_fx.py::TestFunctionalTracing::test_nn_functional_adaptive_max_pool3d_with_indices PASSED [ 27%] 2023-03-31T07:10:14.9777460Z test_fx.py::TestFunctionalTracing::test_nn_functional_affine_grid PASSED [ 27%] 2023-03-31T07:10:14.9777807Z test_fx.py::TestFunctionalTracing::test_nn_functional_alpha_dropout PASSED [ 27%] 2023-03-31T07:10:14.9778142Z test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool1d PASSED [ 27%] 2023-03-31T07:10:14.9778483Z test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool2d PASSED [ 27%] 2023-03-31T07:10:14.9778819Z test_fx.py::TestFunctionalTracing::test_nn_functional_avg_pool3d PASSED [ 27%] 2023-03-31T07:10:14.9779196Z test_fx.py::TestFunctionalTracing::test_nn_functional_batch_norm PASSED [ 27%] 2023-03-31T07:10:14.9779525Z test_fx.py::TestFunctionalTracing::test_nn_functional_bilinear PASSED [ 27%] 2023-03-31T07:10:14.9779883Z test_fx.py::TestFunctionalTracing::test_nn_functional_binary_cross_entropy PASSED [ 27%] 2023-03-31T07:10:14.9780266Z test_fx.py::TestFunctionalTracing::test_nn_functional_binary_cross_entropy_with_logits PASSED [ 27%] 2023-03-31T07:10:14.9780613Z test_fx.py::TestFunctionalTracing::test_nn_functional_celu PASSED [ 28%] 2023-03-31T07:10:14.9780947Z test_fx.py::TestFunctionalTracing::test_nn_functional_celu_ PASSED [ 28%] 2023-03-31T07:10:14.9781291Z test_fx.py::TestFunctionalTracing::test_nn_functional_channel_shuffle PASSED [ 28%] 2023-03-31T07:10:14.9781637Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv1d PASSED [ 28%] 2023-03-31T07:10:14.9781960Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv2d PASSED [ 28%] 2023-03-31T07:10:14.9782292Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv3d PASSED [ 28%] 2023-03-31T07:10:14.9782624Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv_tbc PASSED [ 28%] 2023-03-31T07:10:14.9782962Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose1d PASSED [ 28%] 2023-03-31T07:10:14.9783323Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose2d PASSED [ 28%] 2023-03-31T07:10:14.9783767Z test_fx.py::TestFunctionalTracing::test_nn_functional_conv_transpose3d PASSED [ 28%] 2023-03-31T07:10:14.9784131Z test_fx.py::TestFunctionalTracing::test_nn_functional_cosine_embedding_loss PASSED [ 28%] 2023-03-31T07:10:14.9784481Z test_fx.py::TestFunctionalTracing::test_nn_functional_cosine_similarity PASSED [ 28%] 2023-03-31T07:10:14.9784873Z test_fx.py::TestFunctionalTracing::test_nn_functional_cross_entropy PASSED [ 29%] 2023-03-31T07:10:14.9785247Z test_fx.py::TestFunctionalTracing::test_nn_functional_ctc_loss PASSED [ 29%] 2023-03-31T07:10:14.9785576Z test_fx.py::TestFunctionalTracing::test_nn_functional_dropout PASSED [ 29%] 2023-03-31T07:10:14.9785918Z test_fx.py::TestFunctionalTracing::test_nn_functional_dropout1d PASSED [ 29%] 2023-03-31T07:10:14.9786254Z test_fx.py::TestFunctionalTracing::test_nn_functional_dropout2d PASSED [ 29%] 2023-03-31T07:10:14.9786591Z test_fx.py::TestFunctionalTracing::test_nn_functional_dropout3d PASSED [ 29%] 2023-03-31T07:10:14.9786909Z test_fx.py::TestFunctionalTracing::test_nn_functional_elu PASSED [ 29%] 2023-03-31T07:10:14.9787232Z test_fx.py::TestFunctionalTracing::test_nn_functional_elu_ PASSED [ 29%] 2023-03-31T07:10:14.9787562Z test_fx.py::TestFunctionalTracing::test_nn_functional_embedding PASSED [ 29%] 2023-03-31T07:10:14.9787897Z test_fx.py::TestFunctionalTracing::test_nn_functional_embedding_bag PASSED [ 29%] 2023-03-31T07:10:14.9788262Z test_fx.py::TestFunctionalTracing::test_nn_functional_feature_alpha_dropout PASSED [ 29%] 2023-03-31T07:10:14.9788616Z test_fx.py::TestFunctionalTracing::test_nn_functional_fold PASSED [ 29%] 2023-03-31T07:10:14.9788965Z test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool2d PASSED [ 30%] 2023-03-31T07:10:14.9789382Z test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool2d_with_indices PASSED [ 30%] 2023-03-31T07:10:14.9789764Z test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool3d PASSED [ 30%] 2023-03-31T07:10:14.9790147Z test_fx.py::TestFunctionalTracing::test_nn_functional_fractional_max_pool3d_with_indices PASSED [ 30%] 2023-03-31T07:10:14.9790520Z test_fx.py::TestFunctionalTracing::test_nn_functional_gaussian_nll_loss PASSED [ 30%] 2023-03-31T07:10:14.9790850Z test_fx.py::TestFunctionalTracing::test_nn_functional_gelu PASSED [ 30%] 2023-03-31T07:10:14.9791185Z test_fx.py::TestFunctionalTracing::test_nn_functional_glu PASSED [ 30%] 2023-03-31T07:10:14.9791523Z test_fx.py::TestFunctionalTracing::test_nn_functional_grid_sample PASSED [ 30%] 2023-03-31T07:10:14.9791887Z test_fx.py::TestFunctionalTracing::test_nn_functional_group_norm PASSED [ 30%] 2023-03-31T07:10:14.9792238Z test_fx.py::TestFunctionalTracing::test_nn_functional_gumbel_softmax PASSED [ 30%] 2023-03-31T07:10:14.9792588Z test_fx.py::TestFunctionalTracing::test_nn_functional_hardshrink PASSED [ 30%] 2023-03-31T07:10:14.9792936Z test_fx.py::TestFunctionalTracing::test_nn_functional_hardsigmoid PASSED [ 30%] 2023-03-31T07:10:14.9793264Z test_fx.py::TestFunctionalTracing::test_nn_functional_hardswish PASSED [ 31%] 2023-03-31T07:10:14.9793598Z test_fx.py::TestFunctionalTracing::test_nn_functional_hardtanh PASSED [ 31%] 2023-03-31T07:10:14.9793932Z test_fx.py::TestFunctionalTracing::test_nn_functional_hardtanh_ PASSED [ 31%] 2023-03-31T07:10:14.9794281Z test_fx.py::TestFunctionalTracing::test_nn_functional_hinge_embedding_loss PASSED [ 31%] 2023-03-31T07:10:14.9794637Z test_fx.py::TestFunctionalTracing::test_nn_functional_huber_loss PASSED [ 31%] 2023-03-31T07:10:14.9794985Z test_fx.py::TestFunctionalTracing::test_nn_functional_instance_norm PASSED [ 31%] 2023-03-31T07:10:14.9795334Z test_fx.py::TestFunctionalTracing::test_nn_functional_interpolate PASSED [ 31%] 2023-03-31T07:10:14.9795660Z test_fx.py::TestFunctionalTracing::test_nn_functional_kl_div PASSED [ 31%] 2023-03-31T07:10:14.9795987Z test_fx.py::TestFunctionalTracing::test_nn_functional_l1_loss PASSED [ 31%] 2023-03-31T07:10:14.9796320Z test_fx.py::TestFunctionalTracing::test_nn_functional_layer_norm PASSED [ 31%] 2023-03-31T07:10:14.9796645Z test_fx.py::TestFunctionalTracing::test_nn_functional_leaky_relu PASSED [ 31%] 2023-03-31T07:10:14.9796978Z test_fx.py::TestFunctionalTracing::test_nn_functional_leaky_relu_ PASSED [ 31%] 2023-03-31T07:10:14.9797308Z test_fx.py::TestFunctionalTracing::test_nn_functional_linear PASSED [ 32%] 2023-03-31T07:10:14.9797691Z test_fx.py::TestFunctionalTracing::test_nn_functional_local_response_norm PASSED [ 32%] 2023-03-31T07:10:14.9798063Z test_fx.py::TestFunctionalTracing::test_nn_functional_log_softmax PASSED [ 32%] 2023-03-31T07:10:14.9798407Z test_fx.py::TestFunctionalTracing::test_nn_functional_logsigmoid PASSED [ 32%] 2023-03-31T07:10:14.9798746Z test_fx.py::TestFunctionalTracing::test_nn_functional_lp_pool1d PASSED [ 32%] 2023-03-31T07:10:14.9799070Z test_fx.py::TestFunctionalTracing::test_nn_functional_lp_pool2d PASSED [ 32%] 2023-03-31T07:10:14.9799420Z test_fx.py::TestFunctionalTracing::test_nn_functional_margin_ranking_loss PASSED [ 32%] 2023-03-31T07:10:14.9799770Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool1d PASSED [ 32%] 2023-03-31T07:10:14.9800122Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool1d_with_indices PASSED [ 32%] 2023-03-31T07:10:14.9800466Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool2d PASSED [ 32%] 2023-03-31T07:10:14.9800819Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool2d_with_indices PASSED [ 32%] 2023-03-31T07:10:14.9801176Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool3d PASSED [ 32%] 2023-03-31T07:10:14.9801523Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_pool3d_with_indices PASSED [ 33%] 2023-03-31T07:10:14.9801917Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool1d PASSED [ 33%] 2023-03-31T07:10:14.9802265Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool2d PASSED [ 33%] 2023-03-31T07:10:14.9802612Z test_fx.py::TestFunctionalTracing::test_nn_functional_max_unpool3d PASSED [ 33%] 2023-03-31T07:10:14.9802943Z test_fx.py::TestFunctionalTracing::test_nn_functional_mish PASSED [ 33%] 2023-03-31T07:10:14.9803499Z test_fx.py::TestFunctionalTracing::test_nn_functional_mse_loss PASSED [ 33%] 2023-03-31T07:10:14.9803863Z test_fx.py::TestFunctionalTracing::test_nn_functional_multi_head_attention_forward PASSED [ 33%] 2023-03-31T07:10:14.9804224Z test_fx.py::TestFunctionalTracing::test_nn_functional_multi_margin_loss PASSED [ 33%] 2023-03-31T07:10:14.9804590Z test_fx.py::TestFunctionalTracing::test_nn_functional_multilabel_margin_loss PASSED [ 33%] 2023-03-31T07:10:14.9805036Z test_fx.py::TestFunctionalTracing::test_nn_functional_multilabel_soft_margin_loss PASSED [ 33%] 2023-03-31T07:10:14.9805415Z test_fx.py::TestFunctionalTracing::test_nn_functional_native_channel_shuffle PASSED [ 33%] 2023-03-31T07:10:14.9805758Z test_fx.py::TestFunctionalTracing::test_nn_functional_nll_loss PASSED [ 34%] 2023-03-31T07:10:14.9806099Z test_fx.py::TestFunctionalTracing::test_nn_functional_normalize PASSED [ 34%] 2023-03-31T07:10:14.9806437Z test_fx.py::TestFunctionalTracing::test_nn_functional_one_hot PASSED [ 34%] 2023-03-31T07:10:14.9806755Z test_fx.py::TestFunctionalTracing::test_nn_functional_pad PASSED [ 34%] 2023-03-31T07:10:14.9807108Z test_fx.py::TestFunctionalTracing::test_nn_functional_pairwise_distance PASSED [ 34%] 2023-03-31T07:10:14.9807460Z test_fx.py::TestFunctionalTracing::test_nn_functional_pdist PASSED [ 34%] 2023-03-31T07:10:14.9807806Z test_fx.py::TestFunctionalTracing::test_nn_functional_pixel_shuffle PASSED [ 34%] 2023-03-31T07:10:14.9808152Z test_fx.py::TestFunctionalTracing::test_nn_functional_pixel_unshuffle PASSED [ 34%] 2023-03-31T07:10:14.9808514Z test_fx.py::TestFunctionalTracing::test_nn_functional_poisson_nll_loss PASSED [ 34%] 2023-03-31T07:10:14.9808862Z test_fx.py::TestFunctionalTracing::test_nn_functional_prelu PASSED [ 34%] 2023-03-31T07:10:14.9809183Z test_fx.py::TestFunctionalTracing::test_nn_functional_relu PASSED [ 34%] 2023-03-31T07:10:14.9809514Z test_fx.py::TestFunctionalTracing::test_nn_functional_relu6 PASSED [ 34%] 2023-03-31T07:10:14.9809843Z test_fx.py::TestFunctionalTracing::test_nn_functional_relu_ PASSED [ 35%] 2023-03-31T07:10:14.9810173Z test_fx.py::TestFunctionalTracing::test_nn_functional_rrelu PASSED [ 35%] 2023-03-31T07:10:14.9810540Z test_fx.py::TestFunctionalTracing::test_nn_functional_rrelu_ PASSED [ 35%] 2023-03-31T07:10:14.9810935Z test_fx.py::TestFunctionalTracing::test_nn_functional_scaled_dot_product_attention PASSED [ 35%] 2023-03-31T07:10:14.9811293Z test_fx.py::TestFunctionalTracing::test_nn_functional_selu PASSED [ 35%] 2023-03-31T07:10:14.9811612Z test_fx.py::TestFunctionalTracing::test_nn_functional_selu_ PASSED [ 35%] 2023-03-31T07:10:14.9811939Z test_fx.py::TestFunctionalTracing::test_nn_functional_silu PASSED [ 35%] 2023-03-31T07:10:14.9812278Z test_fx.py::TestFunctionalTracing::test_nn_functional_smooth_l1_loss PASSED [ 35%] 2023-03-31T07:10:14.9812632Z test_fx.py::TestFunctionalTracing::test_nn_functional_soft_margin_loss PASSED [ 35%] 2023-03-31T07:10:14.9812969Z test_fx.py::TestFunctionalTracing::test_nn_functional_softmax PASSED [ 35%] 2023-03-31T07:10:14.9813302Z test_fx.py::TestFunctionalTracing::test_nn_functional_softmin PASSED [ 35%] 2023-03-31T07:10:14.9813640Z test_fx.py::TestFunctionalTracing::test_nn_functional_softplus PASSED [ 35%] 2023-03-31T07:10:14.9813967Z test_fx.py::TestFunctionalTracing::test_nn_functional_softshrink PASSED [ 36%] 2023-03-31T07:10:14.9814305Z test_fx.py::TestFunctionalTracing::test_nn_functional_threshold PASSED [ 36%] 2023-03-31T07:10:14.9814637Z test_fx.py::TestFunctionalTracing::test_nn_functional_threshold_ PASSED [ 36%] 2023-03-31T07:10:14.9815042Z test_fx.py::TestFunctionalTracing::test_nn_functional_triplet_margin_loss PASSED [ 36%] 2023-03-31T07:10:14.9815416Z test_fx.py::TestFunctionalTracing::test_nn_functional_triplet_margin_with_distance_loss PASSED [ 36%] 2023-03-31T07:10:14.9815784Z test_fx.py::TestFunctionalTracing::test_nn_functional_unfold PASSED [ 36%] 2023-03-31T07:10:14.9816121Z test_fx.py::TestFunctionalTracing::test_nn_functional_upsample PASSED [ 36%] 2023-03-31T07:10:14.9816458Z test_fx.py::TestFunctionalTracing::test_nn_functional_upsample_bilinear PASSED [ 36%] 2023-03-31T07:10:14.9816824Z test_fx.py::TestFunctionalTracing::test_nn_functional_upsample_nearest PASSED [ 36%] 2023-03-31T07:10:14.9817499Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_H_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 36%] 2023-03-31T07:10:14.9818216Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_T_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 36%] 2023-03-31T07:10:14.9818850Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___getitem___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 36%] 2023-03-31T07:10:14.9819504Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___radd___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9820139Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rdiv___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9820785Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmatmul___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9821427Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmod___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9822054Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rmul___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9822680Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rpow___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9823306Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive___rsub___cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9823922Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__native_batch_norm_legit_cpu_float32 PASSED [ 37%] 2023-03-31T07:10:14.9824416Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__segment_reduce_lengths_cpu_float32 PASSED [ 37%] 2023-03-31T07:10:14.9824874Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__segment_reduce_offsets_cpu_float32 PASSED [ 37%] 2023-03-31T07:10:14.9825515Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__softmax_backward_data_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9826193Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive__upsample_bilinear2d_aa_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 37%] 2023-03-31T07:10:14.9826667Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_abs_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9827094Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_acos_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9827518Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_acosh_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9827940Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_add_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9828388Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addbmm_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9828820Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addcdiv_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9829246Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addcmul_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9829674Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmm_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9830105Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmm_decomposed_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9830544Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addmv_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9830965Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_addr_cpu_float32 PASSED [ 38%] 2023-03-31T07:10:14.9831420Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_all_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9831836Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_allclose_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9832261Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_amax_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9832675Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_amin_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9833082Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_aminmax_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9833508Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_angle_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9833928Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_any_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9834353Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_arange_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9834764Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argmax_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9835187Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argmin_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9835609Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argsort_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9836034Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_argwhere_cpu_float32 PASSED [ 39%] 2023-03-31T07:10:14.9836483Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9836964Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_partial_views_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9837424Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_as_strided_scatter_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9837859Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_asin_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9838267Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_asinh_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9838687Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atan2_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9839114Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atan_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9839537Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atanh_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9840134Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_1d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 40%] 2023-03-31T07:10:14.9840828Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 40%] 2023-03-31T07:10:14.9841481Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_atleast_3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 40%] 2023-03-31T07:10:14.9841956Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_baddbmm_cpu_float32 PASSED [ 40%] 2023-03-31T07:10:14.9842548Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bernoulli_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9843374Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bfloat16_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9844095Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_block_diag_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9844567Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bmm_cpu_float32 PASSED [ 41%] 2023-03-31T07:10:14.9845141Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bool_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9845791Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_shapes_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9846461Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_tensors_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9846956Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_broadcast_to_cpu_float32 PASSED [ 41%] 2023-03-31T07:10:14.9847385Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_bucketize_cpu_float32 PASSED [ 41%] 2023-03-31T07:10:14.9847981Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_byte_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9848633Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cartesian_prod_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 41%] 2023-03-31T07:10:14.9849111Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cat_cpu_float32 PASSED [ 41%] 2023-03-31T07:10:14.9849740Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cauchy_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9850397Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cdist_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9851036Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cdouble_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9851508Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ceil_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9852094Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cfloat_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9852715Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_chalf_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9853341Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_char_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 42%] 2023-03-31T07:10:14.9853814Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9854256Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_inverse_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9854736Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cholesky_solve_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9855168Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_chunk_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9855590Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_cpu_float32 PASSED [ 42%] 2023-03-31T07:10:14.9856017Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_max_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9856440Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clamp_min_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9856866Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_clone_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9857327Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_column_stack_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9857773Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_combinations_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9858193Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_complex_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9858617Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_conj_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9859053Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_conj_physical_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9859486Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_constant_pad_nd_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9860109Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_contiguous_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 43%] 2023-03-31T07:10:14.9860594Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_copysign_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9861026Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_corrcoef_cpu_float32 PASSED [ 43%] 2023-03-31T07:10:14.9861437Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cos_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9861858Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cosh_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9862289Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_count_nonzero_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9862767Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cov_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9863205Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cross_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9863715Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cummax_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9864140Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cummin_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9864567Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumprod_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9864976Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumsum_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9865416Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_cumulative_trapezoid_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9865859Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_deg2rad_cpu_float32 PASSED [ 44%] 2023-03-31T07:10:14.9866281Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diag_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9866695Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diag_embed_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9867163Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagflat_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9867601Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_copy_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9868035Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9868461Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diagonal_scatter_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9868899Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_diff_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9869323Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_digamma_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9869782Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dist_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9870205Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_floor_rounding_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9870659Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_no_rounding_mode_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9871114Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_div_trunc_rounding_cpu_float32 PASSED [ 45%] 2023-03-31T07:10:14.9871549Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dot_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9872144Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_double_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 46%] 2023-03-31T07:10:14.9872621Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dsplit_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9873050Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_dstack_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9873629Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_einsum_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 46%] 2023-03-31T07:10:14.9874097Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_cpu_float32 XFAIL [ 46%] 2023-03-31T07:10:14.9874527Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_like_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9874999Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_empty_permuted_cpu_float32 XFAIL [ 46%] 2023-03-31T07:10:14.9875440Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_eq_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9875859Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_equal_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9876285Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erf_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9876704Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erfc_cpu_float32 PASSED [ 46%] 2023-03-31T07:10:14.9877112Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_erfinv_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9877534Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exp2_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9877952Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exp_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9878554Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expand_as_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 47%] 2023-03-31T07:10:14.9879188Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expand_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 47%] 2023-03-31T07:10:14.9879698Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_expm1_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9880305Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_exponential_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 47%] 2023-03-31T07:10:14.9880776Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_eye_cpu_float32 XFAIL [ 47%] 2023-03-31T07:10:14.9881187Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fft2_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9881616Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fft_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9882049Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fftn_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9882521Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_fftshift_cpu_float32 PASSED [ 47%] 2023-03-31T07:10:14.9882942Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfft2_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9883585Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfft_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9884013Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_hfftn_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9884438Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifft2_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9884852Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifft_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9885284Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifftn_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9885724Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ifftshift_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9886165Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfft2_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9886583Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfft_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9887012Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_ihfftn_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9887443Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfft2_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9887948Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfft_cpu_float32 PASSED [ 48%] 2023-03-31T07:10:14.9888399Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_irfftn_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9888832Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfft2_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9889262Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfft_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9889674Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fft_rfftn_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9890099Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fill_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9890526Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flatten_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9890950Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flip_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9891363Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fliplr_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9891793Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_flipud_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9892441Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_float_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 49%] 2023-03-31T07:10:14.9892923Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_float_power_cpu_float32 PASSED [ 49%] 2023-03-31T07:10:14.9893338Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_floor_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9893767Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_floor_divide_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9894198Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmax_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9894616Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmin_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9895061Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_fmod_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9895477Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_frac_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9895897Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_frexp_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9896313Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_full_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9896723Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_full_like_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9897151Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gather_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9897575Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ge_cpu_float32 PASSED [ 50%] 2023-03-31T07:10:14.9898178Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_geometric_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 50%] 2023-03-31T07:10:14.9898644Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_geqrf_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9899069Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gradient_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9899505Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_grid_sampler_2d_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9899938Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_gt_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9900554Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_half_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 51%] 2023-03-31T07:10:14.9901058Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_heaviside_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9901491Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histc_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9901903Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histogram_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9902340Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_histogramdd_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9902773Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hsplit_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9903195Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hstack_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9903719Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_hypot_cpu_float32 PASSED [ 51%] 2023-03-31T07:10:14.9904140Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_i0_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9904555Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_igamma_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9905039Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_igammac_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9905459Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_add_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9905898Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_copy_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9906329Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_fill_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9906757Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_put_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9907183Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_reduce_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9907660Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_index_select_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9908093Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_inner_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9908684Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_int_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 52%] 2023-03-31T07:10:14.9909145Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isclose_cpu_float32 PASSED [ 52%] 2023-03-31T07:10:14.9909570Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isfinite_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9909999Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isin_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9910422Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isinf_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9910832Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isnan_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9911256Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isneginf_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9911686Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isposinf_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9912115Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_isreal_cpu_float32 PASSED [ 53%] 2023-03-31T07:10:14.9912568Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_2inputs_2outputs_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T07:10:14.9913119Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_4inputs_with_extra_args_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T07:10:14.9913651Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_binary_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T07:10:14.9914149Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_binary_return_by_ref_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T07:10:14.9914629Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_jiterator_unary_cpu_float32 SKIPPED (Only runs on cuda) [ 53%] 2023-03-31T07:10:14.9915081Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_kron_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9915508Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_kthvalue_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9915936Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ldexp_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9916346Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_le_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9916764Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lerp_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9917248Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lgamma_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9917677Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cholesky_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9932116Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cholesky_ex_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9932735Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cond_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9933202Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_cross_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9933638Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_det_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9934286Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_det_singular_cpu_float32 PASSED [ 54%] 2023-03-31T07:10:14.9934742Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_diagonal_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9935187Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eig_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9935614Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigh_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9936057Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigvals_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9936503Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_eigvalsh_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9936974Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_householder_product_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9937430Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_inv_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9937875Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_inv_ex_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9938327Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_factor_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9938783Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_factor_ex_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9939229Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_ldl_solve_cpu_float32 PASSED [ 55%] 2023-03-31T07:10:14.9939721Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lstsq_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9940516Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lstsq_grad_oriented_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 56%] 2023-03-31T07:10:14.9941024Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9941455Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_factor_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9941907Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_factor_ex_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9942359Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_lu_solve_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9942805Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_norm_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9943253Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_power_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9943822Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_rank_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9944350Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_matrix_rank_hermitian_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9944808Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_multi_dot_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9945235Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_norm_cpu_float32 PASSED [ 56%] 2023-03-31T07:10:14.9945699Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_norm_subgradients_at_zero_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9946160Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9946602Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_hermitian_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9947162Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_pinv_singular_cpu_float32 SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 57%] 2023-03-31T07:10:14.9947671Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_qr_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9948113Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_slogdet_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9948556Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9948985Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_ex_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9949442Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_solve_triangular_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9949896Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_svd_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9950325Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_svdvals_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9950776Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_tensorinv_cpu_float32 PASSED [ 57%] 2023-03-31T07:10:14.9951234Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_tensorsolve_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9951681Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vander_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9952150Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vecdot_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9952626Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linalg_vector_norm_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9953074Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_linspace_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9953503Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log10_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9953914Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log1p_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9954338Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log2_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9954762Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9955388Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_normal_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 58%] 2023-03-31T07:10:14.9955860Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_softmax_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9956345Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_log_softmax_with_dtype_cpu_float32 PASSED [ 58%] 2023-03-31T07:10:14.9956800Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logaddexp2_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9957241Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logaddexp_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9957667Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logcumsumexp_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9958107Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logdet_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9958549Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_and_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9958983Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_not_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9959439Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_or_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9959869Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logical_xor_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9960297Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logit_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9960724Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logspace_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9961150Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_logsumexp_cpu_float32 PASSED [ 59%] 2023-03-31T07:10:14.9961756Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_long_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 59%] 2023-03-31T07:10:14.9962232Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lt_cpu_float32 PASSED [ 60%] 2023-03-31T07:10:14.9962821Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9963555Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_solve_cpu_float32 PASSED [ 60%] 2023-03-31T07:10:14.9963992Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_lu_unpack_cpu_float32 PASSED [ 60%] 2023-03-31T07:10:14.9964594Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mH_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9965313Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mT_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9965976Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_amax_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9966634Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_amin_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9967291Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_argmax_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9967937Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_argmin_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9968581Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_cumprod_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9969239Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_cumsum_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 60%] 2023-03-31T07:10:14.9969773Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_fill_cpu_float32 PASSED [ 61%] 2023-03-31T07:10:14.9970397Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_log_softmax_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9971066Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_logaddexp_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9971718Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_logsumexp_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9972385Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_mean_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9973081Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_median_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9973738Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9974381Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_normalize_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9975035Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_prod_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 61%] 2023-03-31T07:10:14.9975523Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_scatter_cpu_float32 PASSED [ 61%] 2023-03-31T07:10:14.9975970Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_select_cpu_float32 PASSED [ 61%] 2023-03-31T07:10:14.9976576Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_softmax_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9977237Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_softmin_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9977887Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_std_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9978570Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_sum_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9979245Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_masked_var_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9979710Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_matmul_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9980148Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_matrix_exp_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9980586Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_binary_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9981232Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_pool2d_with_indices_backward_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 62%] 2023-03-31T07:10:14.9981737Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_reduction_no_dim_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9982195Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_max_reduction_with_dim_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9982650Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_maximum_cpu_float32 PASSED [ 62%] 2023-03-31T07:10:14.9983115Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mean_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9983602Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_median_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9984236Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_meshgrid_list_of_tensors_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 63%] 2023-03-31T07:10:14.9984764Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_meshgrid_variadic_tensors_cpu_float32 SKIPPED (Skipped!) [ 63%] 2023-03-31T07:10:14.9985233Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_binary_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9985669Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_reduction_no_dim_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9986168Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_min_reduction_with_dim_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9986624Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_minimum_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9987047Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mm_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9987456Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mode_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9987884Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_movedim_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9988313Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_msort_cpu_float32 PASSED [ 63%] 2023-03-31T07:10:14.9988723Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mul_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9989332Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_multinomial_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 64%] 2023-03-31T07:10:14.9989809Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mv_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9990252Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_1_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9990694Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_3_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9991153Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_mvlgamma_mvlgamma_p_5_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9991626Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nan_to_num_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9992085Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanmean_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9992511Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanmedian_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9992948Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nanquantile_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9993381Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nansum_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9993812Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_narrow_copy_cpu_float32 PASSED [ 64%] 2023-03-31T07:10:14.9994230Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_narrow_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:14.9994675Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_batch_norm_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:14.9995315Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_dropout_backward_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:14.9995856Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_native_layer_norm_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:14.9996277Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ne_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:14.9996698Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_neg_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:14.9997296Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_empty_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:14.9997961Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_empty_strided_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:14.9998602Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_full_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:14.9999285Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_ones_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:14.9999928Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_new_zeros_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 65%] 2023-03-31T07:10:15.0000409Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nextafter_cpu_float32 PASSED [ 65%] 2023-03-31T07:10:15.0000859Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool1d_cpu_float32 PASSED [ 66%] 2023-03-31T07:10:15.0001525Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0002246Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_avg_pool3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0002949Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool1d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0004063Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0004838Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_adaptive_max_pool3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0005575Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_alpha_dropout_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0006094Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool1d_cpu_float32 PASSED [ 66%] 2023-03-31T07:10:15.0006561Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool2d_cpu_float32 PASSED [ 66%] 2023-03-31T07:10:15.0007006Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_avg_pool3d_cpu_float32 PASSED [ 66%] 2023-03-31T07:10:15.0007647Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_batch_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 66%] 2023-03-31T07:10:15.0008161Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_bilinear_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0008821Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_binary_cross_entropy_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 67%] 2023-03-31T07:10:15.0009611Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_binary_cross_entropy_with_logits_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 67%] 2023-03-31T07:10:15.0010293Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_celu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 67%] 2023-03-31T07:10:15.0010796Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv1d_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0011252Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv2d_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0011729Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose1d_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0012198Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose2d_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0012722Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_conv_transpose3d_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0013390Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cosine_embedding_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 67%] 2023-03-31T07:10:15.0013922Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cosine_similarity_cpu_float32 PASSED [ 67%] 2023-03-31T07:10:15.0014561Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_cross_entropy_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 67%] 2023-03-31T07:10:15.0015255Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_ctc_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0015938Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0016620Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0017301Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_dropout_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0017986Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_elu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0018700Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_embedding_bag_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0019389Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_embedding_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0020108Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_feature_alpha_dropout_with_train_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0020860Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_feature_alpha_dropout_without_train_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0021575Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_fractional_max_pool2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0022286Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_fractional_max_pool3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0023020Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_gaussian_nll_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 68%] 2023-03-31T07:10:15.0023610Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_gelu_cpu_float32 PASSED [ 69%] 2023-03-31T07:10:15.0024228Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_glu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0024911Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_grid_sample_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0025589Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_group_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0026144Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardshrink_cpu_float32 PASSED [ 69%] 2023-03-31T07:10:15.0026792Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardsigmoid_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0027466Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardswish_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0028144Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hardtanh_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0028849Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_hinge_embedding_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0029547Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_huber_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0030218Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_instance_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0030917Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_area_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 69%] 2023-03-31T07:10:15.0031657Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_bicubic_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0032436Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_bilinear_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0033148Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_linear_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0033836Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_nearest_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0034552Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_interpolate_trilinear_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0035239Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_kl_div_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0035909Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_l1_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0036619Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_layer_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0037288Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_leaky_relu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0037797Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_linear_cpu_float32 PASSED [ 70%] 2023-03-31T07:10:15.0038450Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_local_response_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 70%] 2023-03-31T07:10:15.0038970Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_logsigmoid_cpu_float32 PASSED [ 70%] 2023-03-31T07:10:15.0039645Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_margin_ranking_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0040338Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool1d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0041018Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0041706Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_pool3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0042401Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool1d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0043268Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool1d_grad_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0043964Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool2d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0044656Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool2d_grad_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0045395Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool3d_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0046126Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_max_unpool3d_grad_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0046789Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_mish_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0047453Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_mse_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 71%] 2023-03-31T07:10:15.0048140Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multi_margin_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 72%] 2023-03-31T07:10:15.0048849Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multilabel_margin_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 72%] 2023-03-31T07:10:15.0049558Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_multilabel_soft_margin_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 72%] 2023-03-31T07:10:15.0050292Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_nll_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 72%] 2023-03-31T07:10:15.0050969Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_normalize_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 72%] 2023-03-31T07:10:15.0051481Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_circular_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0051956Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_constant_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0052416Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_reflect_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0052927Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pad_replicate_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0053410Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pairwise_distance_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0053886Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pdist_cpu_float32 PASSED [ 72%] 2023-03-31T07:10:15.0054336Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pixel_shuffle_cpu_float32 PASSED [ 73%] 2023-03-31T07:10:15.0054815Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_pixel_unshuffle_cpu_float32 PASSED [ 73%] 2023-03-31T07:10:15.0055481Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_poisson_nll_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0055999Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_prelu_cpu_float32 PASSED [ 73%] 2023-03-31T07:10:15.0056614Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_relu6_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0057287Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_relu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0057960Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_rrelu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0058717Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_scaled_dot_product_attention_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0059436Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_selu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0060088Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_silu_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0060768Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_smooth_l1_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0061461Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_soft_margin_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 73%] 2023-03-31T07:10:15.0062144Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softmin_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0062820Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softmin_with_dtype_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0063440Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softplus_cpu_float32 PASSED [ 74%] 2023-03-31T07:10:15.0063917Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softshrink_cpu_float32 PASSED [ 74%] 2023-03-31T07:10:15.0064556Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_softsign_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0065239Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_tanhshrink_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0065906Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_threshold_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0066642Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_triplet_margin_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0067373Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_triplet_margin_with_distance_loss_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0068068Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_unfold_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0068746Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_upsample_bilinear_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0069454Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nn_functional_upsample_nearest_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 74%] 2023-03-31T07:10:15.0069952Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_nonzero_cpu_float32 XFAIL [ 75%] 2023-03-31T07:10:15.0070543Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0071178Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_fro_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0071800Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_inf_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0072467Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_norm_nuc_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0073127Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0073780Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_in_place_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0074432Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_normal_number_mean_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 75%] 2023-03-31T07:10:15.0074918Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ones_cpu_float32 PASSED [ 75%] 2023-03-31T07:10:15.0075349Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ones_like_cpu_float32 PASSED [ 75%] 2023-03-31T07:10:15.0075783Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ormqr_cpu_float32 PASSED [ 75%] 2023-03-31T07:10:15.0076199Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_outer_cpu_float32 PASSED [ 75%] 2023-03-31T07:10:15.0076833Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pca_lowrank_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0077316Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_permute_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0077749Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pinverse_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0078167Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polar_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0078795Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_0_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0079481Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_1_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0080207Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_2_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0080874Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_3_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0081552Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_polygamma_polygamma_n_4_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 76%] 2023-03-31T07:10:15.0082043Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_positive_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0082477Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_pow_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0083239Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_prod_cpu_float32 PASSED [ 76%] 2023-03-31T07:10:15.0083682Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_put_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0084105Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_qr_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0084532Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_quantile_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0084949Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rad2deg_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0085648Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rand_like_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 77%] 2023-03-31T07:10:15.0086335Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randint_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 77%] 2023-03-31T07:10:15.0086993Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randint_like_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 77%] 2023-03-31T07:10:15.0087636Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randn_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 77%] 2023-03-31T07:10:15.0088262Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_randn_like_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 77%] 2023-03-31T07:10:15.0088740Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_ravel_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0089170Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_real_cpu_float32 PASSED [ 77%] 2023-03-31T07:10:15.0089612Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reciprocal_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0090038Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_remainder_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0090517Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_renorm_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0091120Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_repeat_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 78%] 2023-03-31T07:10:15.0091611Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_repeat_interleave_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0092214Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reshape_as_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 78%] 2023-03-31T07:10:15.0092691Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_reshape_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0093329Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resize__cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 78%] 2023-03-31T07:10:15.0094300Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resize_as__cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 78%] 2023-03-31T07:10:15.0094840Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resolve_conj_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0095279Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_resolve_neg_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0095711Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_roll_cpu_float32 PASSED [ 78%] 2023-03-31T07:10:15.0096139Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rot90_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0096553Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0096989Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_0_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0097441Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_3_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0097887Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_round_decimals_neg_3_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0098315Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rsqrt_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0098742Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_rsub_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0099222Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scalar_tensor_cpu_float32 XFAIL [ 79%] 2023-03-31T07:10:15.0099670Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_add_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0100107Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0100552Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_amax_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0101004Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_amin_cpu_float32 PASSED [ 79%] 2023-03-31T07:10:15.0101441Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_mean_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0101897Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_prod_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0102350Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_scatter_reduce_sum_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0102797Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_searchsorted_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0103263Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_select_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0103788Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_select_scatter_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0104219Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sgn_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0104817Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_short_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 80%] 2023-03-31T07:10:15.0105284Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sigmoid_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0105714Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sign_cpu_float32 PASSED [ 80%] 2023-03-31T07:10:15.0106377Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_bartlett_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 80%] 2023-03-31T07:10:15.0107070Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_blackman_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 80%] 2023-03-31T07:10:15.0107735Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_cosine_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0108417Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_exponential_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0109106Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_gaussian_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0109792Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_general_cosine_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0110491Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_general_hamming_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0111167Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_hamming_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0111839Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_hann_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0112549Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_kaiser_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0113253Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signal_windows_nuttall_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 81%] 2023-03-31T07:10:15.0113747Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_signbit_cpu_float32 PASSED [ 81%] 2023-03-31T07:10:15.0114159Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sin_cpu_float32 PASSED [ 81%] 2023-03-31T07:10:15.0114580Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sinc_cpu_float32 PASSED [ 81%] 2023-03-31T07:10:15.0115005Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sinh_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0115581Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_slice_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 82%] 2023-03-31T07:10:15.0116068Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_slice_scatter_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0116539Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_softmax_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0116986Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_softmax_with_dtype_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0117407Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sort_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0117844Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sparse_mm_reduce_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0118302Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sparse_sampled_addmm_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0118754Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_airy_ai_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0119192Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_j0_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0119674Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_j1_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0120118Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_y0_cpu_float32 PASSED [ 82%] 2023-03-31T07:10:15.0120557Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_bessel_y1_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0121005Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_t_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0121491Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_u_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0122188Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 83%] 2023-03-31T07:10:15.0122936Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 83%] 2023-03-31T07:10:15.0123636Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_entr_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0124070Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_erfcx_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0124528Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_hermite_polynomial_h_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0125076Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_hermite_polynomial_he_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0125563Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i0e_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0125992Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i1_cpu_float32 PASSED [ 83%] 2023-03-31T07:10:15.0126422Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_i1e_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0126880Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_laguerre_polynomial_l_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0127574Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_legendre_polynomial_p_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 84%] 2023-03-31T07:10:15.0128088Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_log_ndtr_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0128553Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_i0_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0129018Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_i1_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0129523Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_k0_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0129965Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_modified_bessel_k1_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0130419Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_ndtr_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0130858Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_ndtri_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0131532Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_polygamma_special_polygamma_n_0_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 84%] 2023-03-31T07:10:15.0132100Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_scaled_modified_bessel_k0_cpu_float32 PASSED [ 84%] 2023-03-31T07:10:15.0132590Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_scaled_modified_bessel_k1_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0133288Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_t_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 85%] 2023-03-31T07:10:15.0134053Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_u_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 85%] 2023-03-31T07:10:15.0134803Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_v_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 85%] 2023-03-31T07:10:15.0135578Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_shifted_chebyshev_polynomial_w_cpu_float32 SKIPPED (Skipping - testing takes an unreasonably long time, #79528) [ 85%] 2023-03-31T07:10:15.0136116Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_spherical_bessel_j0_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0136579Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_xlog1py_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0137025Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_special_zeta_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0137686Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 85%] 2023-03-31T07:10:15.0138370Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_list_args_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 85%] 2023-03-31T07:10:15.0138594Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_split_with_sizes_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0138801Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sqrt_cpu_float32 PASSED [ 85%] 2023-03-31T07:10:15.0139010Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_square_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0139216Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_squeeze_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0139432Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_squeeze_multiple_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0139628Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_stack_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0139836Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0140079Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_mean_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0140299Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_mean_unbiased_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0140510Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_std_unbiased_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0140881Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_stft_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 86%] 2023-03-31T07:10:15.0141086Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sub_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0141293Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sum_cpu_float32 PASSED [ 86%] 2023-03-31T07:10:15.0141703Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_sum_to_size_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 86%] 2023-03-31T07:10:15.0141913Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_svd_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0142278Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_svd_lowrank_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 87%] 2023-03-31T07:10:15.0142480Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_t_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0142696Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_take_along_dim_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0142906Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_take_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0143109Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tan_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0143313Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tanh_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0143628Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tensor_split_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0143857Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tensordot_cpu_float32 SKIPPED (Skipped!) [ 87%] 2023-03-31T07:10:15.0144058Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tile_cpu_float32 PASSED [ 87%] 2023-03-31T07:10:15.0144430Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_to_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 87%] 2023-03-31T07:10:15.0144852Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_to_sparse_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 87%] 2023-03-31T07:10:15.0145094Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_topk_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0145302Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trace_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0145514Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_transpose_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0145719Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trapezoid_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0145927Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trapz_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0146147Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_triangular_solve_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0146351Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_tril_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0146553Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_triu_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0146799Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_true_divide_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0146989Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_trunc_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0147198Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unbind_cpu_float32 PASSED [ 88%] 2023-03-31T07:10:15.0147412Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unflatten_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0147624Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unfold_copy_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0147999Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unfold_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 89%] 2023-03-31T07:10:15.0148408Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_uniform_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 89%] 2023-03-31T07:10:15.0148806Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unique_consecutive_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 89%] 2023-03-31T07:10:15.0149178Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unique_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 89%] 2023-03-31T07:10:15.0149390Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unsafe_split_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0149606Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_unsqueeze_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0149798Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0150006Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_mean_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0150228Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_mean_unbiased_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0150436Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_var_unbiased_cpu_float32 PASSED [ 89%] 2023-03-31T07:10:15.0150642Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vdot_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0150856Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_as_complex_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0151268Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_as_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 90%] 2023-03-31T07:10:15.0151506Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_copy_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0151879Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_view_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 90%] 2023-03-31T07:10:15.0152091Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vsplit_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0152284Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_vstack_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0152653Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_where_cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 90%] 2023-03-31T07:10:15.0152857Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_xlogy_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0153226Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zero__cpu_float32 SKIPPED (This path doesn't work on Python functions) [ 90%] 2023-03-31T07:10:15.0153432Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zeros_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0153675Z test_fx.py::TestOperatorSignaturesCPU::test_get_torch_func_signature_exhaustive_zeros_like_cpu_float32 PASSED [ 90%] 2023-03-31T07:10:15.0153837Z test_fx.py::TestVisionTracing::test_torchvision_models_alexnet PASSED [ 91%] 2023-03-31T07:10:15.0154006Z test_fx.py::TestVisionTracing::test_torchvision_models_convnext_base PASSED [ 91%] 2023-03-31T07:10:15.0154172Z test_fx.py::TestVisionTracing::test_torchvision_models_convnext_large PASSED [ 91%] 2023-03-31T07:10:15.0154323Z test_fx.py::TestVisionTracing::test_torchvision_models_convnext_small PASSED [ 91%] 2023-03-31T07:10:15.0154487Z test_fx.py::TestVisionTracing::test_torchvision_models_convnext_tiny PASSED [ 91%] 2023-03-31T07:10:15.0154653Z test_fx.py::TestVisionTracing::test_torchvision_models_densenet121 PASSED [ 91%] 2023-03-31T07:10:15.0154818Z test_fx.py::TestVisionTracing::test_torchvision_models_densenet161 PASSED [ 91%] 2023-03-31T07:10:15.0155009Z test_fx.py::TestVisionTracing::test_torchvision_models_densenet169 PASSED [ 91%] 2023-03-31T07:10:15.0155171Z test_fx.py::TestVisionTracing::test_torchvision_models_densenet201 PASSED [ 91%] 2023-03-31T07:10:15.0155375Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_mobilenet_v3_large_320_fpn PASSED [ 91%] 2023-03-31T07:10:15.0155576Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_mobilenet_v3_large_fpn PASSED [ 91%] 2023-03-31T07:10:15.0155767Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_resnet50_fpn PASSED [ 91%] 2023-03-31T07:10:15.0155946Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_fasterrcnn_resnet50_fpn_v2 PASSED [ 92%] 2023-03-31T07:10:15.0156131Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_fcos_resnet50_fpn PASSED [ 92%] 2023-03-31T07:10:15.0156328Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_keypointrcnn_resnet50_fpn PASSED [ 92%] 2023-03-31T07:10:15.0156516Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_maskrcnn_resnet50_fpn PASSED [ 92%] 2023-03-31T07:10:15.0156708Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_maskrcnn_resnet50_fpn_v2 PASSED [ 92%] 2023-03-31T07:10:15.0156899Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_retinanet_resnet50_fpn PASSED [ 92%] 2023-03-31T07:10:15.0157088Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_retinanet_resnet50_fpn_v2 PASSED [ 92%] 2023-03-31T07:10:15.0157262Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_ssd300_vgg16 PASSED [ 92%] 2023-03-31T07:10:15.0157444Z test_fx.py::TestVisionTracing::test_torchvision_models_detection_ssdlite320_mobilenet_v3_large PASSED [ 92%] 2023-03-31T07:10:15.0157642Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b0 PASSED [ 92%] 2023-03-31T07:10:15.0157832Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b1 PASSED [ 92%] 2023-03-31T07:10:15.0157995Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b2 PASSED [ 92%] 2023-03-31T07:10:15.0158164Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b3 PASSED [ 93%] 2023-03-31T07:10:15.0158328Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b4 PASSED [ 93%] 2023-03-31T07:10:15.0158487Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b5 PASSED [ 93%] 2023-03-31T07:10:15.0158645Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b6 PASSED [ 93%] 2023-03-31T07:10:15.0158803Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_b7 PASSED [ 93%] 2023-03-31T07:10:15.0158958Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_l PASSED [ 93%] 2023-03-31T07:10:15.0159125Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_m PASSED [ 93%] 2023-03-31T07:10:15.0159298Z test_fx.py::TestVisionTracing::test_torchvision_models_efficientnet_v2_s PASSED [ 93%] 2023-03-31T07:10:15.0159464Z test_fx.py::TestVisionTracing::test_torchvision_models_googlenet PASSED [ 93%] 2023-03-31T07:10:15.0159662Z test_fx.py::TestVisionTracing::test_torchvision_models_inception_v3 PASSED [ 93%] 2023-03-31T07:10:15.0159824Z test_fx.py::TestVisionTracing::test_torchvision_models_maxvit_t PASSED [ 93%] 2023-03-31T07:10:15.0159985Z test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet0_5 PASSED [ 93%] 2023-03-31T07:10:15.0160146Z test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet0_75 PASSED [ 94%] 2023-03-31T07:10:15.0160288Z test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet1_0 PASSED [ 94%] 2023-03-31T07:10:15.0160446Z test_fx.py::TestVisionTracing::test_torchvision_models_mnasnet1_3 PASSED [ 94%] 2023-03-31T07:10:15.0160611Z test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v2 PASSED [ 94%] 2023-03-31T07:10:15.0160783Z test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v3_large PASSED [ 94%] 2023-03-31T07:10:15.0160982Z test_fx.py::TestVisionTracing::test_torchvision_models_mobilenet_v3_small PASSED [ 94%] 2023-03-31T07:10:15.0161147Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_16gf PASSED [ 94%] 2023-03-31T07:10:15.0161308Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_1_6gf PASSED [ 94%] 2023-03-31T07:10:15.0161467Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_32gf PASSED [ 94%] 2023-03-31T07:10:15.0161612Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_3_2gf PASSED [ 94%] 2023-03-31T07:10:15.0161766Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_400mf PASSED [ 94%] 2023-03-31T07:10:15.0161920Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_800mf PASSED [ 95%] 2023-03-31T07:10:15.0162085Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_x_8gf PASSED [ 95%] 2023-03-31T07:10:15.0162247Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_128gf PASSED [ 95%] 2023-03-31T07:10:15.0162409Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_16gf PASSED [ 95%] 2023-03-31T07:10:15.0162569Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_1_6gf PASSED [ 95%] 2023-03-31T07:10:15.0162726Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_32gf PASSED [ 95%] 2023-03-31T07:10:15.0162869Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_3_2gf PASSED [ 95%] 2023-03-31T07:10:15.0163215Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_400mf PASSED [ 95%] 2023-03-31T07:10:15.0163452Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_800mf PASSED [ 95%] 2023-03-31T07:10:15.0163616Z test_fx.py::TestVisionTracing::test_torchvision_models_regnet_y_8gf PASSED [ 95%] 2023-03-31T07:10:15.0163839Z test_fx.py::TestVisionTracing::test_torchvision_models_resnet101 PASSED [ 95%] 2023-03-31T07:10:15.0164033Z test_fx.py::TestVisionTracing::test_torchvision_models_resnet152 PASSED [ 95%] 2023-03-31T07:10:15.0164189Z test_fx.py::TestVisionTracing::test_torchvision_models_resnet18 PASSED [ 96%] 2023-03-31T07:10:15.0164347Z test_fx.py::TestVisionTracing::test_torchvision_models_resnet34 PASSED [ 96%] 2023-03-31T07:10:15.0164505Z test_fx.py::TestVisionTracing::test_torchvision_models_resnet50 PASSED [ 96%] 2023-03-31T07:10:15.0164659Z test_fx.py::TestVisionTracing::test_torchvision_models_resnext101_32x8d PASSED [ 96%] 2023-03-31T07:10:15.0164821Z test_fx.py::TestVisionTracing::test_torchvision_models_resnext101_64x4d PASSED [ 96%] 2023-03-31T07:10:15.0164982Z test_fx.py::TestVisionTracing::test_torchvision_models_resnext50_32x4d PASSED [ 96%] 2023-03-31T07:10:15.0165183Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_mobilenet_v3_large PASSED [ 96%] 2023-03-31T07:10:15.0165377Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_resnet101 PASSED [ 96%] 2023-03-31T07:10:15.0165569Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_deeplabv3_resnet50 PASSED [ 96%] 2023-03-31T07:10:15.0165753Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_fcn_resnet101 PASSED [ 96%] 2023-03-31T07:10:15.0165984Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_fcn_resnet50 PASSED [ 96%] 2023-03-31T07:10:15.0166168Z test_fx.py::TestVisionTracing::test_torchvision_models_segmentation_lraspp_mobilenet_v3_large PASSED [ 96%] 2023-03-31T07:10:15.0166337Z test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x0_5 PASSED [ 97%] 2023-03-31T07:10:15.0166504Z test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x1_0 PASSED [ 97%] 2023-03-31T07:10:15.0166672Z test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x1_5 PASSED [ 97%] 2023-03-31T07:10:15.0166840Z test_fx.py::TestVisionTracing::test_torchvision_models_shufflenet_v2_x2_0 PASSED [ 97%] 2023-03-31T07:10:15.0167009Z test_fx.py::TestVisionTracing::test_torchvision_models_squeezenet1_0 PASSED [ 97%] 2023-03-31T07:10:15.0167171Z test_fx.py::TestVisionTracing::test_torchvision_models_squeezenet1_1 PASSED [ 97%] 2023-03-31T07:10:15.0167367Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_b PASSED [ 97%] 2023-03-31T07:10:15.0167527Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_s PASSED [ 97%] 2023-03-31T07:10:15.0167668Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_t PASSED [ 97%] 2023-03-31T07:10:15.0167828Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_b PASSED [ 97%] 2023-03-31T07:10:15.0167982Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_s PASSED [ 97%] 2023-03-31T07:10:15.0168133Z test_fx.py::TestVisionTracing::test_torchvision_models_swin_v2_t PASSED [ 97%] 2023-03-31T07:10:15.0168289Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg11 PASSED [ 98%] 2023-03-31T07:10:15.0168444Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg11_bn PASSED [ 98%] 2023-03-31T07:10:15.0168599Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg13 PASSED [ 98%] 2023-03-31T07:10:15.0168753Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg13_bn PASSED [ 98%] 2023-03-31T07:10:15.0168893Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg16 PASSED [ 98%] 2023-03-31T07:10:15.0169048Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg16_bn PASSED [ 98%] 2023-03-31T07:10:15.0169197Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg19 PASSED [ 98%] 2023-03-31T07:10:15.0169349Z test_fx.py::TestVisionTracing::test_torchvision_models_vgg19_bn PASSED [ 98%] 2023-03-31T07:10:15.0169508Z test_fx.py::TestVisionTracing::test_torchvision_models_video_mc3_18 PASSED [ 98%] 2023-03-31T07:10:15.0169675Z test_fx.py::TestVisionTracing::test_torchvision_models_video_mvit_v1_b PASSED [ 98%] 2023-03-31T07:10:15.0169880Z test_fx.py::TestVisionTracing::test_torchvision_models_video_mvit_v2_s PASSED [ 98%] 2023-03-31T07:10:15.0170078Z test_fx.py::TestVisionTracing::test_torchvision_models_video_r2plus1d_18 PASSED [ 98%] 2023-03-31T07:10:15.0170227Z test_fx.py::TestVisionTracing::test_torchvision_models_video_r3d_18 PASSED [ 99%] 2023-03-31T07:10:15.0170385Z test_fx.py::TestVisionTracing::test_torchvision_models_video_s3d PASSED [ 99%] 2023-03-31T07:10:15.0170548Z test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_b PASSED [ 99%] 2023-03-31T07:10:15.0170708Z test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_s PASSED [ 99%] 2023-03-31T07:10:15.0170865Z test_fx.py::TestVisionTracing::test_torchvision_models_video_swin3d_t PASSED [ 99%] 2023-03-31T07:10:15.0171018Z test_fx.py::TestVisionTracing::test_torchvision_models_vit_b_16 PASSED [ 99%] 2023-03-31T07:10:15.0171173Z test_fx.py::TestVisionTracing::test_torchvision_models_vit_b_32 PASSED [ 99%] 2023-03-31T07:10:15.0171327Z test_fx.py::TestVisionTracing::test_torchvision_models_vit_h_14 PASSED [ 99%] 2023-03-31T07:10:15.0171469Z test_fx.py::TestVisionTracing::test_torchvision_models_vit_l_16 PASSED [ 99%] 2023-03-31T07:10:15.0171624Z test_fx.py::TestVisionTracing::test_torchvision_models_vit_l_32 PASSED [ 99%] 2023-03-31T07:10:15.0171790Z test_fx.py::TestVisionTracing::test_torchvision_models_wide_resnet101_2 PASSED [ 99%] 2023-03-31T07:10:15.0171986Z test_fx.py::TestVisionTracing::test_torchvision_models_wide_resnet50_2 PASSED [100%] 2023-03-31T07:10:15.0171998Z 2023-03-31T07:10:15.0172339Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fx/test_fx-b0efee86074bcae7.xml - 2023-03-31T07:10:15.0172472Z =========== 956 passed, 221 skipped, 5 xfailed in 255.04s (0:04:15) ============ 2023-03-31T07:10:15.0172655Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:10:15.0172662Z 2023-03-31T07:10:15.0173059Z ##[endgroup] 2023-03-31T07:10:15.0173323Z FINISHED PRINTING LOG FILE of test_fx (/var/lib/jenkins/workspace/test/test-reports/test_fx_sxdy1_ua.log) 2023-03-31T07:10:15.0173345Z 2023-03-31T07:10:15.0173491Z Running test_nn ... [2023-03-31 07:10:14.933807] 2023-03-31T07:10:15.0174007Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_nn.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:10:14.934161] 2023-03-31T07:12:13.9307476Z 2023-03-31T07:12:13.9307975Z Expand the folded group to see the log file of test_nn 2023-03-31T07:12:13.9311397Z ##[group]PRINTING LOG FILE of test_nn (/var/lib/jenkins/workspace/test/test-reports/test_nn_7anrhld6.log) 2023-03-31T07:12:13.9312325Z Test results will be stored in test-reports/python-pytest/test_nn/test_nn-2e6c9c8f712a39c2.xml 2023-03-31T07:12:13.9312900Z ============================= test session starts ============================== 2023-03-31T07:12:13.9313613Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:12:13.9314042Z cachedir: .pytest_cache 2023-03-31T07:12:13.9314510Z hypothesis profile 'pytorch_ci' -> database=None, deadline=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:12:13.9314905Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:12:13.9315368Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:12:13.9315668Z collecting ... collected 2108 items 2023-03-31T07:12:13.9508057Z Running 2108 items in this shard: test/test_nn.py::TestNN::test_AdaptiveLogSoftmax, test/test_nn.py::TestNN::test_AdaptiveLogSoftmax_cuda, test/test_nn.py::TestNN::test_AvgPool1d, test/test_nn.py::TestNN::test_AvgPool1d_cuda, test/test_nn.py::TestNN::test_AvgPool1d_no_batch_dim, test/test_nn.py::TestNN::test_AvgPool1d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_AvgPool1d_stride, test/test_nn.py::TestNN::test_AvgPool1d_stride_cuda, test/test_nn.py::TestNN::test_AvgPool1d_stride_pad, test/test_nn.py::TestNN::test_AvgPool1d_stride_pad_cuda, test/test_nn.py::TestNN::test_AvgPool2d, test/test_nn.py::TestNN::test_AvgPool2d_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor, test/test_nn.py::TestNN::test_AvgPool2d_divisor_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool2d_divisor_stride_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool2d_divisor_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool2d_divisor_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool2d_no_batch_dim, test/test_nn.py::TestNN::test_AvgPool2d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_AvgPool2d_stride, test/test_nn.py::TestNN::test_AvgPool2d_stride_cuda, test/test_nn.py::TestNN::test_AvgPool2d_stride_pad, test/test_nn.py::TestNN::test_AvgPool2d_stride_pad_cuda, test/test_nn.py::TestNN::test_AvgPool3d, test/test_nn.py::TestNN::test_AvgPool3d_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_stride_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_divisor_with_long_tensor, test/test_nn.py::TestNN::test_AvgPool3d_divisor_with_long_tensor_cuda, test/test_nn.py::TestNN::test_AvgPool3d_no_batch_dim, test/test_nn.py::TestNN::test_AvgPool3d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride, test/test_nn.py::TestNN::test_AvgPool3d_stride1_pad0_gpu_input, test/test_nn.py::TestNN::test_AvgPool3d_stride1_pad0_gpu_input_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_fixedkw_output, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_fixedkw_output_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_general_output, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_general_output_cuda, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_input_nooverlap, test/test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_input_nooverlap_cuda, test/test_nn.py::TestNN::test_BCELoss, test/test_nn.py::TestNN::test_BCELoss_cuda_bfloat16, test/test_nn.py::TestNN::test_BCELoss_cuda_double, test/test_nn.py::TestNN::test_BCELoss_cuda_float, test/test_nn.py::TestNN::test_BCELoss_cuda_half, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_none, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_BCELoss_no_reduce, test/test_nn.py::TestNN::test_BCELoss_no_reduce_cuda, test/test_nn.py::TestNN::test_BCELoss_no_reduce_scalar, test/test_nn.py::TestNN::test_BCELoss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_BCELoss_scalar_weights, test/test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_bfloat16, test/test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_double, test/test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_float, test/test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_half, test/test_nn.py::TestNN::test_BCELoss_weights, test/test_nn.py::TestNN::test_BCELoss_weights_cuda_bfloat16, test/test_nn.py::TestNN::test_BCELoss_weights_cuda_double, test/test_nn.py::TestNN::test_BCELoss_weights_cuda_float, test/test_nn.py::TestNN::test_BCELoss_weights_cuda_half, test/test_nn.py::TestNN::test_BCELoss_weights_no_reduce, test/test_nn.py::TestNN::test_BCELoss_weights_no_reduce_cuda, test/test_nn.py::TestNN::test_BCELoss_weights_no_reduce_scalar, test/test_nn.py::TestNN::test_BCELoss_weights_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_BCEWithLogitsLoss, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_half, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_legacy_enum, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_legacy_enum_cuda, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_scalar, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_half, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_weights, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_double, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_float, test/test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_half, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_eval, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_eval, test/test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_affine, test/test_nn.py::TestNN::test_BatchNorm1d_affine_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_affine_eval, test/test_nn.py::TestNN::test_BatchNorm1d_affine_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average, test/test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_eval, test/test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_not_affine, test/test_nn.py::TestNN::test_BatchNorm1d_not_affine_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_not_affine_eval, test/test_nn.py::TestNN::test_BatchNorm1d_not_affine_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats, test/test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_eval, test/test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_zero_batch, test/test_nn.py::TestNN::test_BatchNorm1d_zero_batch_cuda, test/test_nn.py::TestNN::test_BatchNorm1d_zero_batch_eval, test/test_nn.py::TestNN::test_BatchNorm1d_zero_batch_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d, test/test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average, test/test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_eval, test/test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_eval, test/test_nn.py::TestNN::test_BatchNorm2d_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_momentum, test/test_nn.py::TestNN::test_BatchNorm2d_momentum_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_momentum_eval, test/test_nn.py::TestNN::test_BatchNorm2d_momentum_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_not_affine, test/test_nn.py::TestNN::test_BatchNorm2d_not_affine_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_not_affine_eval, test/test_nn.py::TestNN::test_BatchNorm2d_not_affine_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats, test/test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_eval, test/test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_zero_batch, test/test_nn.py::TestNN::test_BatchNorm2d_zero_batch_cuda, test/test_nn.py::TestNN::test_BatchNorm2d_zero_batch_eval, test/test_nn.py::TestNN::test_BatchNorm2d_zero_batch_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d, test/test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average, test/test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_eval, test/test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_eval, test/test_nn.py::TestNN::test_BatchNorm3d_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_momentum, test/test_nn.py::TestNN::test_BatchNorm3d_momentum_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_momentum_eval, test/test_nn.py::TestNN::test_BatchNorm3d_momentum_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_not_affine, test/test_nn.py::TestNN::test_BatchNorm3d_not_affine_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_not_affine_eval, test/test_nn.py::TestNN::test_BatchNorm3d_not_affine_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats, test/test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_eval, test/test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_zero_batch, test/test_nn.py::TestNN::test_BatchNorm3d_zero_batch_cuda, test/test_nn.py::TestNN::test_BatchNorm3d_zero_batch_eval, test/test_nn.py::TestNN::test_BatchNorm3d_zero_batch_eval_cuda, test/test_nn.py::TestNN::test_CELU_no_batch_dim, test/test_nn.py::TestNN::test_CELU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_critical_target_len, test/test_nn.py::TestNN::test_CTCLoss_lengthchecks_cpu, test/test_nn.py::TestNN::test_CTCLoss_lengthchecks_cuda, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CTCLoss_long_targets, test/test_nn.py::TestNN::test_CTCLoss_typechecks, test/test_nn.py::TestNN::test_CTCLoss_zero_infinity, test/test_nn.py::TestNN::test_Conv1d, test/test_nn.py::TestNN::test_Conv1d_circular_stride2_pad2, test/test_nn.py::TestNN::test_Conv1d_circular_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv1d_cuda, test/test_nn.py::TestNN::test_Conv1d_dilated, test/test_nn.py::TestNN::test_Conv1d_dilated_cuda, test/test_nn.py::TestNN::test_Conv1d_groups, test/test_nn.py::TestNN::test_Conv1d_groups_cuda, test/test_nn.py::TestNN::test_Conv1d_pad1, test/test_nn.py::TestNN::test_Conv1d_pad1_cuda, test/test_nn.py::TestNN::test_Conv1d_pad1size1, test/test_nn.py::TestNN::test_Conv1d_pad1size1_cuda, test/test_nn.py::TestNN::test_Conv1d_pad2, test/test_nn.py::TestNN::test_Conv1d_pad2_cuda, test/test_nn.py::TestNN::test_Conv1d_pad2size1, test/test_nn.py::TestNN::test_Conv1d_pad2size1_cuda, test/test_nn.py::TestNN::test_Conv1d_pad_same, test/test_nn.py::TestNN::test_Conv1d_pad_same2, test/test_nn.py::TestNN::test_Conv1d_pad_same2_cuda, test/test_nn.py::TestNN::test_Conv1d_pad_same_cuda, test/test_nn.py::TestNN::test_Conv1d_pad_same_dilated, test/test_nn.py::TestNN::test_Conv1d_pad_same_dilated_cuda, test/test_nn.py::TestNN::test_Conv1d_pad_valid, test/test_nn.py::TestNN::test_Conv1d_pad_valid_cuda, test/test_nn.py::TestNN::test_Conv1d_reflect_stride2_pad2, test/test_nn.py::TestNN::test_Conv1d_reflect_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv1d_replicate_stride2_pad2, test/test_nn.py::TestNN::test_Conv1d_replicate_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv1d_stride, test/test_nn.py::TestNN::test_Conv1d_stride_cuda, test/test_nn.py::TestNN::test_Conv1d_zero_batch, test/test_nn.py::TestNN::test_Conv1d_zero_batch_cuda, test/test_nn.py::TestNN::test_Conv1d_zeros_stride2_pad2, test/test_nn.py::TestNN::test_Conv1d_zeros_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv2d, test/test_nn.py::TestNN::test_Conv2d_circular_stride2_pad2, test/test_nn.py::TestNN::test_Conv2d_circular_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv2d_cuda, test/test_nn.py::TestNN::test_Conv2d_depthwise, test/test_nn.py::TestNN::test_Conv2d_depthwise_cuda, test/test_nn.py::TestNN::test_Conv2d_depthwise_dilated, test/test_nn.py::TestNN::test_Conv2d_depthwise_dilated_cuda, test/test_nn.py::TestNN::test_Conv2d_depthwise_padded, test/test_nn.py::TestNN::test_Conv2d_depthwise_padded_cuda, test/test_nn.py::TestNN::test_Conv2d_depthwise_strided, test/test_nn.py::TestNN::test_Conv2d_depthwise_strided_cuda, test/test_nn.py::TestNN::test_Conv2d_depthwise_with_multiplier, test/test_nn.py::TestNN::test_Conv2d_depthwise_with_multiplier_cuda, test/test_nn.py::TestNN::test_Conv2d_dilated, test/test_nn.py::TestNN::test_Conv2d_dilated_cuda, test/test_nn.py::TestNN::test_Conv2d_dilated_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_dilated_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_groups, test/test_nn.py::TestNN::test_Conv2d_groups_cuda, test/test_nn.py::TestNN::test_Conv2d_groups_thnn, test/test_nn.py::TestNN::test_Conv2d_groups_thnn_cuda, test/test_nn.py::TestNN::test_Conv2d_groups_thnn_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_groups_thnn_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_groups_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_groups_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_no_bias, test/test_nn.py::TestNN::test_Conv2d_no_bias_cuda, test/test_nn.py::TestNN::test_Conv2d_no_bias_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_no_bias_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_pad_same, test/test_nn.py::TestNN::test_Conv2d_pad_same_cuda, test/test_nn.py::TestNN::test_Conv2d_pad_same_dilated, test/test_nn.py::TestNN::test_Conv2d_pad_same_dilated_cuda, test/test_nn.py::TestNN::test_Conv2d_pad_valid, test/test_nn.py::TestNN::test_Conv2d_pad_valid_cuda, test/test_nn.py::TestNN::test_Conv2d_padding, test/test_nn.py::TestNN::test_Conv2d_padding_cuda, test/test_nn.py::TestNN::test_Conv2d_padding_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_padding_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_reflect_stride2_pad2, test/test_nn.py::TestNN::test_Conv2d_reflect_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv2d_replicate_stride2_pad2, test/test_nn.py::TestNN::test_Conv2d_replicate_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv2d_strided, test/test_nn.py::TestNN::test_Conv2d_strided_cuda, test/test_nn.py::TestNN::test_Conv2d_strided_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_strided_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_zero_batch, test/test_nn.py::TestNN::test_Conv2d_zero_batch_cuda, test/test_nn.py::TestNN::test_Conv2d_zero_batch_with_long_tensor, test/test_nn.py::TestNN::test_Conv2d_zero_batch_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv2d_zeros_stride2_pad2, test/test_nn.py::TestNN::test_Conv2d_zeros_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv3d, test/test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias, test/test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_cuda, test/test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_circular_stride2_pad2, test/test_nn.py::TestNN::test_Conv3d_circular_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv3d_cuda, test/test_nn.py::TestNN::test_Conv3d_dilated, test/test_nn.py::TestNN::test_Conv3d_dilated_cuda, test/test_nn.py::TestNN::test_Conv3d_dilated_strided, test/test_nn.py::TestNN::test_Conv3d_dilated_strided_cuda, test/test_nn.py::TestNN::test_Conv3d_groups, test/test_nn.py::TestNN::test_Conv3d_groups_cuda, test/test_nn.py::TestNN::test_Conv3d_groups_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_groups_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_no_bias, test/test_nn.py::TestNN::test_Conv3d_no_bias_cuda, test/test_nn.py::TestNN::test_Conv3d_no_bias_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_no_bias_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_pad_same, test/test_nn.py::TestNN::test_Conv3d_pad_same_cuda, test/test_nn.py::TestNN::test_Conv3d_pad_same_dilated, test/test_nn.py::TestNN::test_Conv3d_pad_same_dilated_cuda, test/test_nn.py::TestNN::test_Conv3d_pad_valid, test/test_nn.py::TestNN::test_Conv3d_pad_valid_cuda, test/test_nn.py::TestNN::test_Conv3d_replicate_stride2_pad2, test/test_nn.py::TestNN::test_Conv3d_replicate_stride2_pad2_cuda, test/test_nn.py::TestNN::test_Conv3d_stride, test/test_nn.py::TestNN::test_Conv3d_stride_cuda, test/test_nn.py::TestNN::test_Conv3d_stride_padding, test/test_nn.py::TestNN::test_Conv3d_stride_padding_cuda, test/test_nn.py::TestNN::test_Conv3d_stride_padding_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_stride_padding_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_stride_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_stride_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_zero_batch, test/test_nn.py::TestNN::test_Conv3d_zero_batch_cuda, test/test_nn.py::TestNN::test_Conv3d_zero_batch_with_long_tensor, test/test_nn.py::TestNN::test_Conv3d_zero_batch_with_long_tensor_cuda, test/test_nn.py::TestNN::test_Conv3d_zeros_stride2_pad2, test/test_nn.py::TestNN::test_Conv3d_zeros_stride2_pad2_cuda, test/test_nn.py::TestNN::test_ConvTranspose1d, test/test_nn.py::TestNN::test_ConvTranspose1d_cuda, test/test_nn.py::TestNN::test_ConvTranspose1d_dilated, test/test_nn.py::TestNN::test_ConvTranspose1d_dilated_cuda, test/test_nn.py::TestNN::test_ConvTranspose1d_groups, test/test_nn.py::TestNN::test_ConvTranspose1d_groups_cuda, test/test_nn.py::TestNN::test_ConvTranspose1d_no_bias, test/test_nn.py::TestNN::test_ConvTranspose1d_no_bias_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d, test/test_nn.py::TestNN::test_ConvTranspose2d_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_dilated, test/test_nn.py::TestNN::test_ConvTranspose2d_dilated_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_dilated_with_long_tensor, test/test_nn.py::TestNN::test_ConvTranspose2d_dilated_with_long_tensor_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_groups, test/test_nn.py::TestNN::test_ConvTranspose2d_groups_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_groups_with_long_tensor, test/test_nn.py::TestNN::test_ConvTranspose2d_groups_with_long_tensor_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_no_bias, test/test_nn.py::TestNN::test_ConvTranspose2d_no_bias_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_no_bias_with_long_tensor, test/test_nn.py::TestNN::test_ConvTranspose2d_no_bias_with_long_tensor_cuda, test/test_nn.py::TestNN::test_ConvTranspose2d_with_long_tensor, test/test_nn.py::TestNN::test_ConvTranspose2d_with_long_tensor_cuda, test/test_nn.py::TestNN::test_ConvTranspose3d, test/test_nn.py::TestNN::test_ConvTranspose3d_cuda, test/test_nn.py::TestNN::test_ConvTranspose3d_dilated, test/test_nn.py::TestNN::test_ConvTranspose3d_dilated_cuda, test/test_nn.py::TestNN::test_CosineEmbeddingLoss, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_CrossEntropyLoss_weights, test/test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_double, test/test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_float, test/test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_half, test/test_nn.py::TestNN::test_CrossMapLRN2d, test/test_nn.py::TestNN::test_CrossMapLRN2d_cuda, test/test_nn.py::TestNN::test_ELU_no_batch_dim, test/test_nn.py::TestNN::test_ELU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Embedding, test/test_nn.py::TestNN::test_EmbeddingBag_discontiguous, test/test_nn.py::TestNN::test_EmbeddingBag_discontiguous_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_max, test/test_nn.py::TestNN::test_EmbeddingBag_max_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_max_padding_idx, test/test_nn.py::TestNN::test_EmbeddingBag_max_padding_idx_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_mean, test/test_nn.py::TestNN::test_EmbeddingBag_mean_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_mean_padding_idx, test/test_nn.py::TestNN::test_EmbeddingBag_mean_padding_idx_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_sparse, test/test_nn.py::TestNN::test_EmbeddingBag_sparse_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_sum, test/test_nn.py::TestNN::test_EmbeddingBag_sum_cuda, test/test_nn.py::TestNN::test_EmbeddingBag_sum_padding_idx, test/test_nn.py::TestNN::test_EmbeddingBag_sum_padding_idx_cuda, test/test_nn.py::TestNN::test_Embedding_cuda, test/test_nn.py::TestNN::test_Embedding_discontiguous, test/test_nn.py::TestNN::test_Embedding_discontiguous_cuda, test/test_nn.py::TestNN::test_Embedding_sparse, test/test_nn.py::TestNN::test_Embedding_sparse_cuda, test/test_nn.py::TestNN::test_Flatten, test/test_nn.py::TestNN::test_Flatten_cuda, test/test_nn.py::TestNN::test_Flatten_no_batch_dim, test/test_nn.py::TestNN::test_Flatten_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Fold, test/test_nn.py::TestNN::test_Fold_cuda, test/test_nn.py::TestNN::test_Fold_int_input, test/test_nn.py::TestNN::test_Fold_int_input_cuda, test/test_nn.py::TestNN::test_Fold_no_batch_dim_input, test/test_nn.py::TestNN::test_Fold_no_batch_dim_input_cuda, test/test_nn.py::TestNN::test_Fold_no_batch_dim_int_input, test/test_nn.py::TestNN::test_Fold_no_batch_dim_int_input_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_return_indices, test/test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_return_indices_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_no_random_samples, test/test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_no_random_samples_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_asymsize, test/test_nn.py::TestNN::test_FractionalMaxPool3d_asymsize_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_return_indices, test/test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_return_indices_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_cuda, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_no_random_samples, test/test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_no_random_samples_cuda, test/test_nn.py::TestNN::test_GELU_no_batch_dim, test/test_nn.py::TestNN::test_GELU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_GLU_no_batch_dim, test/test_nn.py::TestNN::test_GLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_GN, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_eval, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_eval, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_large_batch_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_affine_large_batch_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_eval, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_cuda, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_eval, test/test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_affine, test/test_nn.py::TestNN::test_GroupNorm_2d_affine_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_affine_eval, test/test_nn.py::TestNN::test_GroupNorm_2d_affine_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_affine_large_feature_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_affine_large_feature_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_eval, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_eval, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_eval_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_large_feature_cuda, test/test_nn.py::TestNN::test_GroupNorm_2d_no_affine_large_feature_eval_cuda, test/test_nn.py::TestNN::test_Hardshrink_no_batch_dim, test/test_nn.py::TestNN::test_Hardshrink_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Hardsigmoid_no_batch_dim, test/test_nn.py::TestNN::test_Hardsigmoid_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Hardswish_no_batch_dim, test/test_nn.py::TestNN::test_Hardswish_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Hardtanh_no_batch_dim, test/test_nn.py::TestNN::test_Hardtanh_no_batch_dim_cuda, test/test_nn.py::TestNN::test_HingeEmbeddingLoss, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_no_reduce, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_no_reduce_cuda, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_reduce, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_HuberLoss, test/test_nn.py::TestNN::test_HuberLoss_cuda_bfloat16, test/test_nn.py::TestNN::test_HuberLoss_cuda_double, test/test_nn.py::TestNN::test_HuberLoss_cuda_float, test/test_nn.py::TestNN::test_HuberLoss_cuda_half, test/test_nn.py::TestNN::test_HuberLoss_delta, test/test_nn.py::TestNN::test_HuberLoss_delta_cuda, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_HuberLoss_sum_reduction, test/test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_InstanceNorm1d, test/test_nn.py::TestNN::test_InstanceNorm1d_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_eval, test/test_nn.py::TestNN::test_InstanceNorm1d_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_eval, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d, test/test_nn.py::TestNN::test_InstanceNorm2d_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_eval, test/test_nn.py::TestNN::test_InstanceNorm2d_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_eval, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d, test/test_nn.py::TestNN::test_InstanceNorm3d_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_eval, test/test_nn.py::TestNN::test_InstanceNorm3d_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_eval, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_eval_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_cuda, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_eval, test/test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_eval_cuda, test/test_nn.py::TestNN::test_KLDivLoss, test/test_nn.py::TestNN::test_KLDivLoss_batch_mean, test/test_nn.py::TestNN::test_KLDivLoss_batch_mean_log_target, test/test_nn.py::TestNN::test_KLDivLoss_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_log_target, test/test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction, test/test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_log_target, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_log_target_cuda, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_log_target, test/test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_log_target_cuda, test/test_nn.py::TestNN::test_KLDivLoss_scalar, test/test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction, test/test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_sum_reduction, test/test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_KLDivLoss_with_log_target_no_reduce, test/test_nn.py::TestNN::test_KLDivLoss_with_log_target_no_reduce_cuda, test/test_nn.py::TestNN::test_KLDivLoss_with_target_no_reduce, test/test_nn.py::TestNN::test_KLDivLoss_with_target_no_reduce_cuda, test/test_nn.py::TestNN::test_L1Loss, test/test_nn.py::TestNN::test_L1Loss_cuda_cdouble, test/test_nn.py::TestNN::test_L1Loss_cuda_cfloat, test/test_nn.py::TestNN::test_L1Loss_cuda_double, test/test_nn.py::TestNN::test_L1Loss_cuda_float, test/test_nn.py::TestNN::test_L1Loss_cuda_half, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_none, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_L1Loss_no_reduce, test/test_nn.py::TestNN::test_L1Loss_no_reduce_complex, test/test_nn.py::TestNN::test_L1Loss_no_reduce_complex_cuda, test/test_nn.py::TestNN::test_L1Loss_no_reduce_cuda, test/test_nn.py::TestNN::test_L1Loss_no_reduce_scalar, test/test_nn.py::TestNN::test_L1Loss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_L1Loss_scalar, test/test_nn.py::TestNN::test_L1Loss_scalar_cuda_cdouble, test/test_nn.py::TestNN::test_L1Loss_scalar_cuda_cfloat, test/test_nn.py::TestNN::test_L1Loss_scalar_cuda_double, test/test_nn.py::TestNN::test_L1Loss_scalar_cuda_float, test/test_nn.py::TestNN::test_L1Loss_scalar_cuda_half, test/test_nn.py::TestNN::test_LPPool1d, test/test_nn.py::TestNN::test_LPPool1d_cuda, test/test_nn.py::TestNN::test_LPPool1d_no_batch_dim, test/test_nn.py::TestNN::test_LPPool1d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_LPPool1d_norm, test/test_nn.py::TestNN::test_LPPool1d_norm_cuda, test/test_nn.py::TestNN::test_LPPool2d, test/test_nn.py::TestNN::test_LPPool2d_cuda, test/test_nn.py::TestNN::test_LPPool2d_norm, test/test_nn.py::TestNN::test_LPPool2d_norm_cuda, test/test_nn.py::TestNN::test_LSTM_cell, test/test_nn.py::TestNN::test_LSTM_cell_forward_hidden_size, test/test_nn.py::TestNN::test_LSTM_cell_forward_input_size, test/test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine, test/test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_cuda, test/test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_eval, test/test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_eval_cuda, test/test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine, test/test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_cuda, test/test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_eval, test/test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_eval_cuda, test/test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine, test/test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_cuda, test/test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_eval, test/test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_eval_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine, test/test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_eval, test/test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_eval_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature, test/test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_eval, test/test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_eval_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine, test/test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_cuda, test/test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_eval, test/test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_eval_cuda, test/test_nn.py::TestNN::test_LeakyReLU_no_batch_dim, test/test_nn.py::TestNN::test_LeakyReLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Linear, test/test_nn.py::TestNN::test_Linear_cuda, test/test_nn.py::TestNN::test_Linear_no_batch_dim, test/test_nn.py::TestNN::test_Linear_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Linear_no_bias, test/test_nn.py::TestNN::test_Linear_no_bias_cuda, test/test_nn.py::TestNN::test_LocalResponseNorm_1d, test/test_nn.py::TestNN::test_LocalResponseNorm_1d_cuda, test/test_nn.py::TestNN::test_LocalResponseNorm_2d_uneven_pad, test/test_nn.py::TestNN::test_LocalResponseNorm_2d_uneven_pad_cuda, test/test_nn.py::TestNN::test_LocalResponseNorm_3d_custom_params, test/test_nn.py::TestNN::test_LocalResponseNorm_3d_custom_params_cuda, test/test_nn.py::TestNN::test_LogSigmoid_no_batch_dim, test/test_nn.py::TestNN::test_LogSigmoid_no_batch_dim_cuda, test/test_nn.py::TestNN::test_MSELoss, test/test_nn.py::TestNN::test_MSELoss_cuda_double, test/test_nn.py::TestNN::test_MSELoss_cuda_float, test/test_nn.py::TestNN::test_MSELoss_cuda_half, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_none, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_MSELoss_no_reduce, test/test_nn.py::TestNN::test_MSELoss_no_reduce_cuda, test/test_nn.py::TestNN::test_MSELoss_no_reduce_scalar, test/test_nn.py::TestNN::test_MSELoss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_MSELoss_prec, test/test_nn.py::TestNN::test_MSELoss_prec_cuda_bfloat16, test/test_nn.py::TestNN::test_MSELoss_prec_cuda_double, test/test_nn.py::TestNN::test_MSELoss_prec_cuda_float, test/test_nn.py::TestNN::test_MSELoss_prec_cuda_half, test/test_nn.py::TestNN::test_MSELoss_scalar, test/test_nn.py::TestNN::test_MSELoss_scalar_cuda_bfloat16, test/test_nn.py::TestNN::test_MSELoss_scalar_cuda_double, test/test_nn.py::TestNN::test_MSELoss_scalar_cuda_float, test/test_nn.py::TestNN::test_MSELoss_scalar_cuda_half, test/test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction, test/test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MSELoss_sum_reduction, test/test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss, test/test_nn.py::TestNN::test_MarginRankingLoss_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_margin, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction, test/test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MaxPool1d, test/test_nn.py::TestNN::test_MaxPool1d_cuda, test/test_nn.py::TestNN::test_MaxPool1d_return_indices, test/test_nn.py::TestNN::test_MaxPool1d_return_indices_cuda, test/test_nn.py::TestNN::test_MaxPool1d_stride, test/test_nn.py::TestNN::test_MaxPool1d_stride_cuda, test/test_nn.py::TestNN::test_MaxPool2d_3d_input, test/test_nn.py::TestNN::test_MaxPool2d_3d_input_cuda, test/test_nn.py::TestNN::test_MaxPool2d_4d_input, test/test_nn.py::TestNN::test_MaxPool2d_4d_input_cuda, test/test_nn.py::TestNN::test_MaxPool2d_return_indices, test/test_nn.py::TestNN::test_MaxPool2d_return_indices_cuda, test/test_nn.py::TestNN::test_MaxPool3d, test/test_nn.py::TestNN::test_MaxPool3d_cuda, test/test_nn.py::TestNN::test_MaxPool3d_return_indices, test/test_nn.py::TestNN::test_MaxPool3d_return_indices_cuda, test/test_nn.py::TestNN::test_MaxPool3d_stride, test/test_nn.py::TestNN::test_MaxPool3d_stride_cuda, test/test_nn.py::TestNN::test_MaxPool3d_stride_padding, test/test_nn.py::TestNN::test_MaxPool3d_stride_padding_cuda, test/test_nn.py::TestNN::test_MaxUnpool1d_net, test/test_nn.py::TestNN::test_MaxUnpool1d_net_cuda, test/test_nn.py::TestNN::test_MaxUnpool1d_net_no_batch_dim, test/test_nn.py::TestNN::test_MaxUnpool1d_net_no_batch_dim_cuda, test/test_nn.py::TestNN::test_MaxUnpool2d_net, test/test_nn.py::TestNN::test_MaxUnpool2d_net_cuda, test/test_nn.py::TestNN::test_MaxUnpool2d_net_no_batch_dim, test/test_nn.py::TestNN::test_MaxUnpool2d_net_no_batch_dim_cuda, test/test_nn.py::TestNN::test_MaxUnpool3d_net, test/test_nn.py::TestNN::test_MaxUnpool3d_net_cuda, test/test_nn.py::TestNN::test_MaxUnpool3d_net_no_batch_dim, test/test_nn.py::TestNN::test_MaxUnpool3d_net_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Mish_no_batch_dim, test/test_nn.py::TestNN::test_Mish_no_batch_dim_cuda, test/test_nn.py::TestNN::test_ModuleDict, test/test_nn.py::TestNN::test_ModuleList, test/test_nn.py::TestNN::test_MultiLabelMarginLoss, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_0d_no_reduce, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_0d_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_bfloat16, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_no_reduce, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_bfloat16, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_index_neg, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_index_neg_cuda, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_reduce, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_reduce, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_half, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_no_reduce, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss, test/test_nn.py::TestNN::test_MultiMarginLoss_1d, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_no_reduce, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_margin, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_no_reduce, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_no_reduce, test/test_nn.py::TestNN::test_MultiMarginLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiMarginLoss_p, test/test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_p_no_reduce, test/test_nn.py::TestNN::test_MultiMarginLoss_p_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction, test/test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction, test/test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_weights, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_half, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_no_reduce, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_no_reduce_cuda, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_NLLLoss, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce_cuda, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce_ignore_index, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce_ignore_index_cuda, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce_weights, test/test_nn.py::TestNN::test_NLLLoss2d_no_reduce_weights_cuda, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce_cuda, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce_ignore_index, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce_ignore_index_cuda, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce_weights, test/test_nn.py::TestNN::test_NLLLossNd_no_reduce_weights_cuda, test/test_nn.py::TestNN::test_NLLLoss_2d, test/test_nn.py::TestNN::test_NLLLoss_2d_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_2d_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_2d_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_2d_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_2d_ignore_index, test/test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction, test/test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_2d_weights, test/test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_higher_dim, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_ignore_index, test/test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_no_reduce, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_ignore_index, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_ignore_index_cuda, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_cuda, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_cuda, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_neg, test/test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_neg_cuda, test/test_nn.py::TestNN::test_NLLLoss_sum_reduction, test/test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_weights, test/test_nn.py::TestNN::test_NLLLoss_weights_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_weights_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_weights_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_weights_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_half, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_bfloat16, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_double, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_float, test/test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_half, test/test_nn.py::TestNN::test_PReLU_backward_requires_grad_false, test/test_nn.py::TestNN::test_PReLU_no_batch_dim, test/test_nn.py::TestNN::test_PReLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Padding122112_3dcircular, test/test_nn.py::TestNN::test_Padding122112_3dcircular_cuda, test/test_nn.py::TestNN::test_Padding1221_2dcircular, test/test_nn.py::TestNN::test_Padding1221_2dcircular_cuda, test/test_nn.py::TestNN::test_Padding12_1dcircular, test/test_nn.py::TestNN::test_Padding12_1dcircular_cuda, test/test_nn.py::TestNN::test_Padding2322_2dcircular, test/test_nn.py::TestNN::test_Padding2322_2dcircular_cuda, test/test_nn.py::TestNN::test_Padding31_1dcircular, test/test_nn.py::TestNN::test_Padding31_1dcircular_cuda, test/test_nn.py::TestNN::test_Padding322112_3dcircular, test/test_nn.py::TestNN::test_Padding322112_3dcircular_cuda, test/test_nn.py::TestNN::test_Padding332122_3dcircular, test/test_nn.py::TestNN::test_Padding332122_3dcircular_cuda, test/test_nn.py::TestNN::test_Padding3331_2dcircular, test/test_nn.py::TestNN::test_Padding3331_2dcircular_cuda, test/test_nn.py::TestNN::test_Padding33_1dcircular, test/test_nn.py::TestNN::test_Padding33_1dcircular_cuda, test/test_nn.py::TestNN::test_PairwiseDistance, test/test_nn.py::TestNN::test_PairwiseDistance_broadcast_lhs, test/test_nn.py::TestNN::test_PairwiseDistance_broadcast_lhs_cuda, test/test_nn.py::TestNN::test_PairwiseDistance_broadcast_rhs, test/test_nn.py::TestNN::test_PairwiseDistance_broadcast_rhs_cuda, test/test_nn.py::TestNN::test_PairwiseDistance_cuda, test/test_nn.py::TestNN::test_PairwiseDistance_no_batch_dim, test/test_nn.py::TestNN::test_PairwiseDistance_no_batch_dim_cuda, test/test_nn.py::TestNN::test_PairwiseDistance_with_non_default_args, test/test_nn.py::TestNN::test_PairwiseDistance_with_non_default_args_cuda, test/test_nn.py::TestNN::test_ParameterDict, test/test_nn.py::TestNN::test_ParameterDict_replication, test/test_nn.py::TestNN::test_ParameterList, test/test_nn.py::TestNN::test_ParameterList_meta, test/test_nn.py::TestNN::test_ParameterList_replication, test/test_nn.py::TestNN::test_PixelShuffle, test/test_nn.py::TestNN::test_PixelShuffle_cuda, test/test_nn.py::TestNN::test_PixelUnshuffle, test/test_nn.py::TestNN::test_PixelUnshuffle_cuda, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_double, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_float, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_half, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_reduce, test/test_nn.py::TestNN::test_PoissonNLLLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_RNN_cell, test/test_nn.py::TestNN::test_RNN_cell_forward_hidden_size, test/test_nn.py::TestNN::test_RNN_cell_forward_input_size, test/test_nn.py::TestNN::test_RNN_cell_forward_zero_hidden_size, test/test_nn.py::TestNN::test_RNN_cell_no_broadcasting, test/test_nn.py::TestNN::test_RNN_change_dropout, test/test_nn.py::TestNN::test_RNN_cpu_vs_cudnn_no_dropout, test/test_nn.py::TestNN::test_RNN_cpu_vs_cudnn_with_dropout, test/test_nn.py::TestNN::test_RNN_cudnn_weight_norm, test/test_nn.py::TestNN::test_RNN_dropout, test/test_nn.py::TestNN::test_RNN_dropout_state, test/test_nn.py::TestNN::test_RNN_input_size_zero, test/test_nn.py::TestNN::test_RNN_nonlinearity, test/test_nn.py::TestNN::test_RReLU, test/test_nn.py::TestNN::test_RReLU_cuda, test/test_nn.py::TestNN::test_RReLU_no_batch_dim, test/test_nn.py::TestNN::test_RReLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_RReLU_with_up_down, test/test_nn.py::TestNN::test_RReLU_with_up_down_cuda, test/test_nn.py::TestNN::test_RReLU_with_up_down_scalar, test/test_nn.py::TestNN::test_RReLU_with_up_down_scalar_cuda, test/test_nn.py::TestNN::test_ReLU6_no_batch_dim, test/test_nn.py::TestNN::test_ReLU6_no_batch_dim_cuda, test/test_nn.py::TestNN::test_ReLU_no_batch_dim, test/test_nn.py::TestNN::test_ReLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_ReplicationPad3d, test/test_nn.py::TestNN::test_ReplicationPad3d_complex, test/test_nn.py::TestNN::test_ReplicationPad3d_complex_cuda, test/test_nn.py::TestNN::test_ReplicationPad3d_cuda, test/test_nn.py::TestNN::test_ReplicationPad3d_no_batch_dim, test/test_nn.py::TestNN::test_ReplicationPad3d_no_batch_dim_cuda, test/test_nn.py::TestNN::test_SELU_no_batch_dim, test/test_nn.py::TestNN::test_SELU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Sequential_add, test/test_nn.py::TestNN::test_Sequential_append, test/test_nn.py::TestNN::test_Sequential_delitem, test/test_nn.py::TestNN::test_Sequential_extend, test/test_nn.py::TestNN::test_Sequential_getitem, test/test_nn.py::TestNN::test_Sequential_iadd, test/test_nn.py::TestNN::test_Sequential_imul, test/test_nn.py::TestNN::test_Sequential_insert, test/test_nn.py::TestNN::test_Sequential_insert_fail_case, test/test_nn.py::TestNN::test_Sequential_mul, test/test_nn.py::TestNN::test_Sequential_pop, test/test_nn.py::TestNN::test_Sequential_rmul, test/test_nn.py::TestNN::test_Sequential_setitem, test/test_nn.py::TestNN::test_Sequential_setitem_named, test/test_nn.py::TestNN::test_SiLU_no_batch_dim, test/test_nn.py::TestNN::test_SiLU_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Sigmoid_no_batch_dim, test/test_nn.py::TestNN::test_Sigmoid_no_batch_dim_cuda, test/test_nn.py::TestNN::test_SmoothL1Loss, test/test_nn.py::TestNN::test_SmoothL1Loss_beta, test/test_nn.py::TestNN::test_SmoothL1Loss_beta_cuda, test/test_nn.py::TestNN::test_SmoothL1Loss_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_no_reduce, test/test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_cuda, test/test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_scalar, test/test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_scalar_cuda, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction, test/test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_SmoothL1Loss_zero_beta, test/test_nn.py::TestNN::test_SmoothL1Loss_zero_beta_cuda, test/test_nn.py::TestNN::test_SoftMarginLoss, test/test_nn.py::TestNN::test_SoftMarginLoss_cuda_double, test/test_nn.py::TestNN::test_SoftMarginLoss_cuda_float, test/test_nn.py::TestNN::test_SoftMarginLoss_cuda_half, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_SoftMarginLoss_no_reduce, test/test_nn.py::TestNN::test_SoftMarginLoss_no_reduce_cuda, test/test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction, test/test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_double, test/test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_float, test/test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_half, test/test_nn.py::TestNN::test_Softplus_no_batch_dim, test/test_nn.py::TestNN::test_Softplus_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Softshrink_no_batch_dim, test/test_nn.py::TestNN::test_Softshrink_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Softsign_no_batch_dim, test/test_nn.py::TestNN::test_Softsign_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Tanh_no_batch_dim, test/test_nn.py::TestNN::test_Tanh_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Tanhshrink_no_batch_dim, test/test_nn.py::TestNN::test_Tanhshrink_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Threshold_no_batch_dim, test/test_nn.py::TestNN::test_Threshold_no_batch_dim_cuda, test/test_nn.py::TestNN::test_TransformerDecoderLayer_gelu_activation, test/test_nn.py::TestNN::test_TransformerDecoderLayer_gelu_activation_cuda, test/test_nn.py::TestNN::test_TransformerDecoderLayer_relu_activation, test/test_nn.py::TestNN::test_TransformerDecoderLayer_relu_activation_cuda, test/test_nn.py::TestNN::test_TransformerEncoderLayer_gelu_activation, test/test_nn.py::TestNN::test_TransformerEncoderLayer_gelu_activation_cuda, test/test_nn.py::TestNN::test_TransformerEncoderLayer_relu_activation, test/test_nn.py::TestNN::test_TransformerEncoderLayer_relu_activation_cuda, test/test_nn.py::TestNN::test_Transformer_cell, test/test_nn.py::TestNN::test_Transformer_multilayer_coder, test/test_nn.py::TestNN::test_Transformer_multilayer_coder_cuda, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_double, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_float, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_half, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_double, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_float, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_half, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_double, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_float, test/test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_half, test/test_nn.py::TestNN::test_Unflatten_no_batch_dim, test/test_nn.py::TestNN::test_Unflatten_no_batch_dim_cuda, test/test_nn.py::TestNN::test_Unfold, test/test_nn.py::TestNN::test_Unfold_cuda, test/test_nn.py::TestNN::test_Unfold_int_input, test/test_nn.py::TestNN::test_Unfold_int_input_cuda, test/test_nn.py::TestNN::test_adaptive_log_softmax, test/test_nn.py::TestNN::test_add_module, test/test_nn.py::TestNN::test_add_module_raises_error_if_attr_exists, test/test_nn.py::TestNN::test_affine_grid, test/test_nn.py::TestNN::test_affine_grid_3d, test/test_nn.py::TestNN::test_affine_grid_error_checking, test/test_nn.py::TestNN::test_assignment, test/test_nn.py::TestNN::test_batchnorm_buffer_update_when_stats_are_not_tracked, test/test_nn.py::TestNN::test_batchnorm_cudnn_half, test/test_nn.py::TestNN::test_batchnorm_cudnn_nhwc, test/test_nn.py::TestNN::test_batchnorm_nhwc_cpu, test/test_nn.py::TestNN::test_batchnorm_nhwc_cuda, test/test_nn.py::TestNN::test_batchnorm_non_contig_cpu_BatchNorm2d, test/test_nn.py::TestNN::test_batchnorm_non_contig_cpu_SyncBatchNorm, test/test_nn.py::TestNN::test_batchnorm_nonaffine_cuda_half_input, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_bias_is_not_same_size_as_input, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_less_than_one_value_per_channel, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_running_mean_is_not_same_size_as_input, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_running_var_is_not_same_size_as_input, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_running_var_or_running_mean_have_forward_grad, test/test_nn.py::TestNN::test_batchnorm_raises_error_if_weight_is_not_same_size_as_input, test/test_nn.py::TestNN::test_bce_loss_always_nonnegative, test/test_nn.py::TestNN::test_bce_loss_broadcasts_weights, test/test_nn.py::TestNN::test_bce_loss_input_range, test/test_nn.py::TestNN::test_bce_loss_size_mismatch, test/test_nn.py::TestNN::test_bce_with_logits_broadcasts_pos_weights, test/test_nn.py::TestNN::test_bce_with_logits_broadcasts_weights, test/test_nn.py::TestNN::test_bce_with_logits_gives_same_result_as_sigmoid_and_bce_loss, test/test_nn.py::TestNN::test_bce_with_logits_gives_same_result_as_sigmoid_and_bce_loss_large_tensors_with_grad, test/test_nn.py::TestNN::test_bce_with_logits_has_correct_forward_grad, test/test_nn.py::TestNN::test_bce_with_logits_has_correct_grad_at_zero, test/test_nn.py::TestNN::test_bce_with_logits_ones_in_pos_weights_are_the_same_as_none, test/test_nn.py::TestNN::test_bce_with_logits_raises_if_target_and_input_are_different_size, test/test_nn.py::TestNN::test_bce_with_logits_stability, test/test_nn.py::TestNN::test_bce_with_logits_with_pos_weight_has_correct_grad_at_zero, test/test_nn.py::TestNN::test_bilinear, test/test_nn.py::TestNN::test_bilinear_broadcasting, test/test_nn.py::TestNN::test_bilinear_no_bias, test/test_nn.py::TestNN::test_bilinear_non_contiguous, test/test_nn.py::TestNN::test_broadcast_double_backwards_gpu, test/test_nn.py::TestNN::test_broadcast_no_grad, test/test_nn.py::TestNN::test_broadcast_not_requiring_grad, test/test_nn.py::TestNN::test_buffer_not_persistent, test/test_nn.py::TestNN::test_buffer_not_persistent_assign, test/test_nn.py::TestNN::test_buffer_not_persistent_del, test/test_nn.py::TestNN::test_buffer_not_persistent_load, test/test_nn.py::TestNN::test_buffer_not_persistent_overwrite, test/test_nn.py::TestNN::test_buffers_and_named_buffers, test/test_nn.py::TestNN::test_call_supports_python_dict_output, test/test_nn.py::TestNN::test_channel_shuffle, test/test_nn.py::TestNN::test_channel_shuffle_return_self, test/test_nn.py::TestNN::test_children, test/test_nn.py::TestNN::test_container_copy, test/test_nn.py::TestNN::test_convert_sync_batchnorm, test/test_nn.py::TestNN::test_cosine_embedding_loss_invalid_shape, test/test_nn.py::TestNN::test_cosine_embedding_loss_margin_no_reduce, test/test_nn.py::TestNN::test_cosine_embedding_loss_no_reduce, test/test_nn.py::TestNN::test_cosine_embedding_loss_with_diff_type, test/test_nn.py::TestNN::test_cosine_similarity, test/test_nn.py::TestNN::test_cross_entropy_loss, test/test_nn.py::TestNN::test_cross_entropy_loss_precision, test/test_nn.py::TestNN::test_cross_entropy_loss_zero_div, test/test_nn.py::TestNN::test_cudnn_rnn_dropout_states_device, test/test_nn.py::TestNN::test_cudnn_weight_format, test/test_nn.py::TestNN::test_cudnn_weight_tying, test/test_nn.py::TestNN::test_dir, test/test_nn.py::TestNN::test_dir_digit, test/test_nn.py::TestNN::test_elu_inplace_gradgrad, test/test_nn.py::TestNN::test_elu_inplace_on_view, test/test_nn.py::TestNN::test_error_RNN_seq_len_zero, test/test_nn.py::TestNN::test_extra_state, test/test_nn.py::TestNN::test_extra_state_missing_get_extra_state, test/test_nn.py::TestNN::test_extra_state_missing_set_extra_state, test/test_nn.py::TestNN::test_extra_state_non_dict, test/test_nn.py::TestNN::test_fb_fc_packed, test/test_nn.py::TestNN::test_flatten, test/test_nn.py::TestNN::test_fold_invalid_arg, test/test_nn.py::TestNN::test_gaussian_nll_loss_args, test/test_nn.py::TestNN::test_gaussian_nll_loss_broadcasting, test/test_nn.py::TestNN::test_get_buffer, test/test_nn.py::TestNN::test_get_buffer_from_submodules, test/test_nn.py::TestNN::test_getattr_with_property, test/test_nn.py::TestNN::test_grid_sample, test/test_nn.py::TestNN::test_grid_sample_3d, test/test_nn.py::TestNN::test_grid_sample_error_checking, test/test_nn.py::TestNN::test_hardtanh_backward, test/test_nn.py::TestNN::test_hardtanh_inplace_gradgrad, test/test_nn.py::TestNN::test_huber_loss_invalid_delta, test/test_nn.py::TestNN::test_inplace_thnn, test/test_nn.py::TestNN::test_interpolate, test/test_nn.py::TestNN::test_interpolate_bicubic_2d, test/test_nn.py::TestNN::test_interpolate_bicubic_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_2d_zero_dim, test/test_nn.py::TestNN::test_interpolate_bicubic_2d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_2d, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_shared_2d, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_shared_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_align_corners, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d, test/test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_align_corners, test/test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_2d, test/test_nn.py::TestNN::test_interpolate_bilinear_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_2d_zero_dim, test/test_nn.py::TestNN::test_interpolate_bilinear_2d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_2d, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_shared_2d, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_shared_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_align_corners, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d, test/test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_align_corners, test/test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_cuda, test/test_nn.py::TestNN::test_interpolate_buffer_overflow, test/test_nn.py::TestNN::test_interpolate_illegal_memory_access, test/test_nn.py::TestNN::test_interpolate_linear_1d, test/test_nn.py::TestNN::test_interpolate_linear_1d_align_corners, test/test_nn.py::TestNN::test_interpolate_linear_1d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_linear_1d_cuda, test/test_nn.py::TestNN::test_interpolate_linear_1d_zero_dim, test/test_nn.py::TestNN::test_interpolate_linear_1d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_linear_scale_1d, test/test_nn.py::TestNN::test_interpolate_linear_scale_1d_align_corners, test/test_nn.py::TestNN::test_interpolate_linear_scale_1d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_linear_scale_1d_cuda, test/test_nn.py::TestNN::test_interpolate_linear_tuple_1d, test/test_nn.py::TestNN::test_interpolate_linear_tuple_1d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_1d, test/test_nn.py::TestNN::test_interpolate_nearest_1d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_1d_zero_dim, test/test_nn.py::TestNN::test_interpolate_nearest_1d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_2d, test/test_nn.py::TestNN::test_interpolate_nearest_2d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_2d_launch_configs, test/test_nn.py::TestNN::test_interpolate_nearest_2d_launch_configs_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_2d_zero_dim, test/test_nn.py::TestNN::test_interpolate_nearest_2d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_3d, test/test_nn.py::TestNN::test_interpolate_nearest_3d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_3d_zero_dim, test/test_nn.py::TestNN::test_interpolate_nearest_3d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_scale_1d, test/test_nn.py::TestNN::test_interpolate_nearest_scale_1d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_scale_2d, test/test_nn.py::TestNN::test_interpolate_nearest_scale_2d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_scale_3d, test/test_nn.py::TestNN::test_interpolate_nearest_scale_3d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_1d, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_1d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_2d, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_2d_cuda, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_3d, test/test_nn.py::TestNN::test_interpolate_nearest_tuple_3d_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_3d, test/test_nn.py::TestNN::test_interpolate_trilinear_3d_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_3d_zero_dim, test/test_nn.py::TestNN::test_interpolate_trilinear_3d_zero_dim_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_scale_3d, test/test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_align_corners, test/test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d, test/test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_align_corners, test/test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_align_corners_cuda, test/test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_cuda, test/test_nn.py::TestNN::test_interpolate_undefined_behavior_casting, test/test_nn.py::TestNN::test_kl_div_log_softmax_target, test/test_nn.py::TestNN::test_kl_div_with_diff_type, test/test_nn.py::TestNN::test_kl_div_with_diff_type_log_target, test/test_nn.py::TestNN::test_l1_loss_correct, test/test_nn.py::TestNN::test_layer_norm_grads_with_create_graph_flag, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCOO, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCSC, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCSR, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightStrided, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCOO, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCSC, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCSR, test/test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightStrided, test/test_nn.py::TestNN::test_linear_broadcasting, test/test_nn.py::TestNN::test_load_state_dict, test/test_nn.py::TestNN::test_load_state_dict_BC, test/test_nn.py::TestNN::test_load_state_dict_child, test/test_nn.py::TestNN::test_load_state_dict_custom, test/test_nn.py::TestNN::test_load_state_dict_invalid, test/test_nn.py::TestNN::test_load_state_dict_ref_cycle, test/test_nn.py::TestNN::test_load_state_dict_type, test/test_nn.py::TestNN::test_log_softmax_cpu, test/test_nn.py::TestNN::test_log_softmax_dim0, test/test_nn.py::TestNN::test_log_softmax_dim0_cuda, test/test_nn.py::TestNN::test_log_softmax_dim3, test/test_nn.py::TestNN::test_log_softmax_dim3_cuda, test/test_nn.py::TestNN::test_log_softmax_lastdim, test/test_nn.py::TestNN::test_log_softmax_lastdim_cuda, test/test_nn.py::TestNN::test_log_softmax_scalar, test/test_nn.py::TestNN::test_log_softmax_scalar_cuda, test/test_nn.py::TestNN::test_log_softmax_spatial, test/test_nn.py::TestNN::test_log_softmax_spatial_cuda, test/test_nn.py::TestNN::test_log_softmax_spatial_special, test/test_nn.py::TestNN::test_log_softmax_spatial_special_cuda, test/test_nn.py::TestNN::test_loss_equal_input_target_shape, test/test_nn.py::TestNN::test_margin_ranking_loss_margin_no_reduce, test/test_nn.py::TestNN::test_margin_ranking_loss_no_reduce, test/test_nn.py::TestNN::test_module_apply_inplace_op, test/test_nn.py::TestNN::test_module_backcompat, test/test_nn.py::TestNN::test_module_super_init, test/test_nn.py::TestNN::test_module_to_argparse, test/test_nn.py::TestNN::test_modules, test/test_nn.py::TestNN::test_mse_loss_size_warning, test/test_nn.py::TestNN::test_multimarginloss_1d_input_0d_target_no_reduce, test/test_nn.py::TestNN::test_multimarginloss_1d_input_0d_target_no_reduce_cuda, test/test_nn.py::TestNN::test_named_children, test/test_nn.py::TestNN::test_named_modules, test/test_nn.py::TestNN::test_named_parameters_remove_duplicate, test/test_nn.py::TestNN::test_nested_tensor_from_mask, test/test_nn.py::TestNN::test_nested_tensor_from_mask_error, test/test_nn.py::TestNN::test_no_grad, test/test_nn.py::TestNN::test_non_leaf_parameters, test/test_nn.py::TestNN::test_normalize, test/test_nn.py::TestNN::test_overwrite_module_params_on_conversion, test/test_nn.py::TestNN::test_pack_sequence_batch_sizes_throw, test/test_nn.py::TestNN::test_pad_scalar_error, test/test_nn.py::TestNN::test_padding_list, test/test_nn.py::TestNN::test_pairwise_distance, test/test_nn.py::TestNN::test_parameter_assignment, test/test_nn.py::TestNN::test_parameterlistdict_pickle, test/test_nn.py::TestNN::test_parameterlistdict_setting_attributes, test/test_nn.py::TestNN::test_parameters_and_named_parameters, test/test_nn.py::TestNN::test_parameters_to_vector, test/test_nn.py::TestNN::test_parse_to, test/test_nn.py::TestNN::test_partial_flat_weights, test/test_nn.py::TestNN::test_pdist, test/test_nn.py::TestNN::test_pdist_cpu_gradgrad_unimplemented, test/test_nn.py::TestNN::test_pdist_cuda_gradgrad_unimplemented, test/test_nn.py::TestNN::test_pdist_empty_col, test/test_nn.py::TestNN::test_pdist_empty_row, test/test_nn.py::TestNN::test_pdist_large, test/test_nn.py::TestNN::test_pdist_zeros, test/test_nn.py::TestNN::test_pixel_shuffle_nhwc_cpu, test/test_nn.py::TestNN::test_pixel_shuffle_unshuffle, test/test_nn.py::TestNN::test_pointwise_loss_broadcast, test/test_nn.py::TestNN::test_pointwise_loss_target_grad_none_reduction, test/test_nn.py::TestNN::test_projections_errors_on_gru_and_rnn, test/test_nn.py::TestNN::test_projections_lstm_args_check, test/test_nn.py::TestNN::test_projections_lstm_check_device, test/test_nn.py::TestNN::test_projections_lstm_initial_hidden_state, test/test_nn.py::TestNN::test_register_buffer_allows_overwriting_with_same_name, test/test_nn.py::TestNN::test_register_buffer_raises_error_if_attr_exists, test/test_nn.py::TestNN::test_register_buffer_raises_error_if_name_is_not_string, test/test_nn.py::TestNN::test_register_buffer_raises_error_if_not_tensor, test/test_nn.py::TestNN::test_register_parameter_allows_overwriting_with_same_name, test/test_nn.py::TestNN::test_register_parameter_raises_error_if_attr_exists, test/test_nn.py::TestNN::test_register_parameter_raises_error_if_name_is_not_string, test/test_nn.py::TestNN::test_register_state_dict_pre_hook, test/test_nn.py::TestNN::test_register_state_dict_pre_hook_backward_compat, test/test_nn.py::TestNN::test_relu_inplace_on_view, test/test_nn.py::TestNN::test_repr, test/test_nn.py::TestNN::test_requires_grad_, test/test_nn.py::TestNN::test_rnn_args_check, test/test_nn.py::TestNN::test_rnn_check_device, test/test_nn.py::TestNN::test_rnn_initial_hidden_state, test/test_nn.py::TestNN::test_rnn_weight_norm, test/test_nn.py::TestNN::test_share_memory, test/test_nn.py::TestNN::test_smoothl1loss_intergral_target, test/test_nn.py::TestNN::test_smoothl1loss_negative_beta_not_supported, test/test_nn.py::TestNN::test_softmax_cpu, test/test_nn.py::TestNN::test_softmax_functional_dim0, test/test_nn.py::TestNN::test_softmax_functional_dim0_cuda, test/test_nn.py::TestNN::test_softmax_functional_dim3, test/test_nn.py::TestNN::test_softmax_functional_dim3_cuda, test/test_nn.py::TestNN::test_softmax_functional_scalar, test/test_nn.py::TestNN::test_softmax_functional_scalar_cuda, test/test_nn.py::TestNN::test_softmax_lastdim, test/test_nn.py::TestNN::test_softmax_lastdim_cuda, test/test_nn.py::TestNN::test_softmax_lastdim_dtype, test/test_nn.py::TestNN::test_softmax_lastdim_dtype_cuda, test/test_nn.py::TestNN::test_softmax_spatial, test/test_nn.py::TestNN::test_softmax_spatial_cuda, test/test_nn.py::TestNN::test_softmax_spatial_dtype, test/test_nn.py::TestNN::test_softmax_spatial_dtype_cuda, test/test_nn.py::TestNN::test_softmax_spatial_special, test/test_nn.py::TestNN::test_softmax_spatial_special_cuda, test/test_nn.py::TestNN::test_softmin, test/test_nn.py::TestNN::test_spectral_norm, test/test_nn.py::TestNN::test_spectral_norm_dim, test/test_nn.py::TestNN::test_spectral_norm_forward, test/test_nn.py::TestNN::test_spectral_norm_load_state_dict, test/test_nn.py::TestNN::test_spectral_norm_pickle, test/test_nn.py::TestNN::test_state_dict, test/test_nn.py::TestNN::test_sync_batchnorm_accuracy_cuda, test/test_nn.py::TestNN::test_sync_batchnorm_backward_elemt, test/test_nn.py::TestNN::test_threshold_bfloat16, test/test_nn.py::TestNN::test_threshold_int, test/test_nn.py::TestNN::test_to, test/test_nn.py::TestNN::test_train_errors_for_invalid_mode, test/test_nn.py::TestNN::test_transformer_args_check, test/test_nn.py::TestNN::test_transformer_layer_args_check, test/test_nn.py::TestNN::test_transformerdecoder, test/test_nn.py::TestNN::test_transformerdecoderlayer, test/test_nn.py::TestNN::test_transformerdecoderlayer_gelu, test/test_nn.py::TestNN::test_triplet_margin_loss, test/test_nn.py::TestNN::test_triplet_margin_loss_no_reduce, test/test_nn.py::TestNN::test_triplet_margin_loss_swap, test/test_nn.py::TestNN::test_triplet_margin_loss_swap_no_reduce, test/test_nn.py::TestNN::test_type, test/test_nn.py::TestNN::test_unflatten, test/test_nn.py::TestNN::test_unflatten_invalid_arg, test/test_nn.py::TestNN::test_unfold_invalid_arg, test/test_nn.py::TestNN::test_upsamplingBilinear2d_spatial_invariance, test/test_nn.py::TestNN::test_upsamplingLinear1d, test/test_nn.py::TestNN::test_upsamplingLinear1d_spatial_invariance, test/test_nn.py::TestNN::test_upsamplingTrilinear3d_spatial_invariance, test/test_nn.py::TestNN::test_upsampling_bfloat16, test/test_nn.py::TestNN::test_upsampling_not_recompute_scale_factor, test/test_nn.py::TestNN::test_upsampling_small_scale, test/test_nn.py::TestNN::test_vector_to_parameters, test/test_nn.py::TestNN::test_weight_norm, test/test_nn.py::TestNN::test_weight_norm_pickle, test/test_nn.py::TestNN::test_zero_grad, test/test_nn.py::TestFusionEval::test_fuse_module_eval_numerics, test/test_nn.py::TestConstantPadNd::test_constant_pad_nd, test/test_nn.py::TestConstantPadNd::test_preserves_memory_format, test/test_nn.py::TestAddRelu::test_add_relu, test/test_nn.py::TestAddRelu::test_add_relu_broadcasting, test/test_nn.py::TestFunctionalPickle::test_pickle_softsign, test/test_nn.py::TestFusionUtils::test_fuse_conv_bn_requires_grad, test/test_nn.py::TestFusionUtils::test_fuse_linear_bn_requires_grad, test/test_nn.py::TestNNDeviceTypeCPU::test_BatchNorm_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_Bilinear_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_cudnn_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_empty_target_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_mean_use_module_form_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_mean_use_module_form_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_none_use_module_form_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_none_use_module_form_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_sum_use_module_form_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_sum_use_module_form_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_GRU_grad_and_gradgrad_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_general_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_memory_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_numeric_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_raises_error_if_one_value_per_group_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm1d_general_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm2d_general_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm3d_general_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_LSTM_grad_and_gradgrad_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_LayerNorm_general_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_LayerNorm_numeric_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_LocalResponseNorm_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_empty_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_empty_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_warnings_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad2d_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad3d_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad_empty_cpu_complex64, test/test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad_empty_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad1d_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad2d_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad3d_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad_empty_cpu_complex128, test/test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad_empty_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_TransformerDecoderLayer_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_TransformerDecoder_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_TransformerEncoderLayer_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_TransformerEncoder_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_Transformer_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_Unfold_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_activations_bfloat16_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_activations_bfloat16_cpu_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate0_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate45_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate90_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotateRandom_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_affine_3d_rotateRandom_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_mixed_cpu_bfloat16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_mixed_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_mixed_cpu_bfloat16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_mixed_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_grad_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_large_batch_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_large_batch_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_mixed_cpu_bfloat16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_mixed_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_update_stats_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_error_if_nonfinite_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_0_5_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_1_5_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_2_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_4_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_inf_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_0_5_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_1_5_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_2_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_4_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_inf_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_multi_device_foreach_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_multi_device_foreach_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_value_foreach_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_value_foreach_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_bfloat16, test/test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_complex128, test/test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_consistent_index_target_and_probs_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_errors_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_weight_ignore_indices_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_with_probs_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_mean_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_none_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_sum_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_index_target_unit_weights_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_one_hot_target_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_all_reductions_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_mean_weighted_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_mean_weighted_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_none_weighted_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_none_weighted_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_sum_weighted_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_sum_weighted_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_unit_weights_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ctc_loss_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_ctc_loss_cudnn_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_device_mask_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_elu_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_elu_inplace_with_neg_alpha_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_fold_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_glu_bfloat16_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_half_precision_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_2d_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_2d_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_3d_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_3d_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_nan_inf_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_nan_inf_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_bfloat16, test/test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_gumbel_softmax_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_gumbel_softmax_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_hardsigmoid_grad_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_hardswish_grad_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_hardswish_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_for_single_spatial_element_during_training_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_False_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_False_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_True_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_True_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_False_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_False_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_True_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_True_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_False_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_False_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_True_affine_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_True_affine_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_less_than_one_value_per_channel_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_invalid_reduction_strings_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_layernorm_half_precision_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_layernorm_weight_bias_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_with_neg_slope_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_with_zero_slope_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_linear_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_log_softmax_big_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_logsigmoid_out_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_lstmcell_backward_only_one_output_grad_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_TxT_layout_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_devices_parity_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_forward_with_nans_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_grad_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_mask_types_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_transformer_layout_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_mish_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_module_to_empty_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_module_to_empty_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_multi_margin_loss_errors_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_all_ignored_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_byte_target_matches_long_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_mean_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_none_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_sum_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_invalid_target_dim_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_invalid_weights_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_mean_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_none_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_sum_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_mismatched_batch_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_out_of_bounds_ignore_index_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_total_weight_is_zero_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nn_empty_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nn_scalars_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nn_scalars_reductions_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_nonlinearity_propagate_nan_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_one_hot_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_overwrite_module_params_on_conversion_cpu_device_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_pad_cpu_complex128, test/test_nn.py::TestNNDeviceTypeCPU::test_pad_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_prelu_backward_32bit_indexing_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_rnn_fused_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_rnn_fused_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_rnn_retain_variables_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_save_lstm_compatibility_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_silu_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_skip_init_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_smooth_l1_loss_bfloat16_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_smooth_l1_loss_vs_huber_loss_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_smoothl1loss_backward_zero_beta_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softmax_64bit_indexing_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_softmax_bfloat16_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softmax_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_softmax_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_softmax_results_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_softplus_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softplus_low_threshold_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softshrink_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softshrink_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_softshrink_negative_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_threshold_inplace_overlap_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_to_complex_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_fast_path_cpu_float64, test/test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_gelu_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_triplet_margin_with_distance_loss_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_triplet_margin_with_distance_loss_default_parity_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_3_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_3_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_5_output_size_32_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_5_output_size_600_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bicubic_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bicubic_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bilinear_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bilinear_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bicubic_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bicubic_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bilinear_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bilinear_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bicubic_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bicubic_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bilinear_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bilinear_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bicubic_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bicubic_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bilinear_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bilinear_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_aa_correctness_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_aa_correctness_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBilinear2d_aa_correctness_memory_format_torch_channels_last_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBilinear2d_aa_correctness_memory_format_torch_contiguous_format_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_launch_config_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_config_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_fail_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_rocm_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_launch_config_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact1d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact1d_rescale_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact2d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact3d_correctness_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingTrilinear3d_align_corners_False_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsamplingTrilinear3d_align_corners_True_cpu, test/test_nn.py::TestNNDeviceTypeCPU::test_upsampling_64bit_indexing_channels_last_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_variable_sequence_cpu_float32, test/test_nn.py::TestNNDeviceTypeCPU::test_warp_softmax_64bit_indexing_cpu_float16, test/test_nn.py::TestNNDeviceTypeCPU::test_warp_softmax_64bit_indexing_cpu_float32 2023-03-31T07:12:13.9671280Z 2023-03-31T07:12:13.9671500Z test_nn.py::TestNN::test_AdaptiveLogSoftmax PASSED [ 0%] 2023-03-31T07:12:13.9671875Z test_nn.py::TestNN::test_AdaptiveLogSoftmax_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9672200Z test_nn.py::TestNN::test_AvgPool1d PASSED [ 0%] 2023-03-31T07:12:13.9672533Z test_nn.py::TestNN::test_AvgPool1d_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9672860Z test_nn.py::TestNN::test_AvgPool1d_no_batch_dim PASSED [ 0%] 2023-03-31T07:12:13.9673198Z test_nn.py::TestNN::test_AvgPool1d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9673521Z test_nn.py::TestNN::test_AvgPool1d_stride PASSED [ 0%] 2023-03-31T07:12:13.9674077Z test_nn.py::TestNN::test_AvgPool1d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9674408Z test_nn.py::TestNN::test_AvgPool1d_stride_pad PASSED [ 0%] 2023-03-31T07:12:13.9674738Z test_nn.py::TestNN::test_AvgPool1d_stride_pad_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9675046Z test_nn.py::TestNN::test_AvgPool2d PASSED [ 0%] 2023-03-31T07:12:13.9675355Z test_nn.py::TestNN::test_AvgPool2d_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9675672Z test_nn.py::TestNN::test_AvgPool2d_divisor PASSED [ 0%] 2023-03-31T07:12:13.9675989Z test_nn.py::TestNN::test_AvgPool2d_divisor_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9676320Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride PASSED [ 0%] 2023-03-31T07:12:13.9676722Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9677063Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad PASSED [ 0%] 2023-03-31T07:12:13.9677400Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9677759Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_with_long_tensor PASSED [ 0%] 2023-03-31T07:12:13.9678143Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_pad_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 0%] 2023-03-31T07:12:13.9678502Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_with_long_tensor PASSED [ 0%] 2023-03-31T07:12:13.9678876Z test_nn.py::TestNN::test_AvgPool2d_divisor_stride_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9679236Z test_nn.py::TestNN::test_AvgPool2d_divisor_with_long_tensor PASSED [ 1%] 2023-03-31T07:12:13.9679597Z test_nn.py::TestNN::test_AvgPool2d_divisor_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9679928Z test_nn.py::TestNN::test_AvgPool2d_no_batch_dim PASSED [ 1%] 2023-03-31T07:12:13.9680259Z test_nn.py::TestNN::test_AvgPool2d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9680586Z test_nn.py::TestNN::test_AvgPool2d_stride PASSED [ 1%] 2023-03-31T07:12:13.9680896Z test_nn.py::TestNN::test_AvgPool2d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9681221Z test_nn.py::TestNN::test_AvgPool2d_stride_pad PASSED [ 1%] 2023-03-31T07:12:13.9681601Z test_nn.py::TestNN::test_AvgPool2d_stride_pad_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9681977Z test_nn.py::TestNN::test_AvgPool3d PASSED [ 1%] 2023-03-31T07:12:13.9682276Z test_nn.py::TestNN::test_AvgPool3d_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9682593Z test_nn.py::TestNN::test_AvgPool3d_divisor PASSED [ 1%] 2023-03-31T07:12:13.9682921Z test_nn.py::TestNN::test_AvgPool3d_divisor_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9683405Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride PASSED [ 1%] 2023-03-31T07:12:13.9683724Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input PASSED [ 1%] 2023-03-31T07:12:13.9684089Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9684470Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_with_long_tensor PASSED [ 1%] 2023-03-31T07:12:13.9684862Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride1_pad0_gpu_input_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9685260Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9685599Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad PASSED [ 1%] 2023-03-31T07:12:13.9685992Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_cuda SKIPPED (Excluded from CUDA tests) [ 1%] 2023-03-31T07:12:13.9686343Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output PASSED [ 2%] 2023-03-31T07:12:13.9686729Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9687131Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9687539Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9687934Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output PASSED [ 2%] 2023-03-31T07:12:13.9688323Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9688764Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9689171Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_general_output_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9689570Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap PASSED [ 2%] 2023-03-31T07:12:13.9689954Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9690355Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9690764Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9704848Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9705258Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_pad_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9705640Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9706016Z test_nn.py::TestNN::test_AvgPool3d_divisor_stride_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9706384Z test_nn.py::TestNN::test_AvgPool3d_divisor_with_long_tensor PASSED [ 2%] 2023-03-31T07:12:13.9706735Z test_nn.py::TestNN::test_AvgPool3d_divisor_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9707085Z test_nn.py::TestNN::test_AvgPool3d_no_batch_dim PASSED [ 2%] 2023-03-31T07:12:13.9707600Z test_nn.py::TestNN::test_AvgPool3d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 2%] 2023-03-31T07:12:13.9707984Z test_nn.py::TestNN::test_AvgPool3d_stride PASSED [ 2%] 2023-03-31T07:12:13.9708290Z test_nn.py::TestNN::test_AvgPool3d_stride1_pad0_gpu_input PASSED [ 3%] 2023-03-31T07:12:13.9708650Z test_nn.py::TestNN::test_AvgPool3d_stride1_pad0_gpu_input_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9709018Z test_nn.py::TestNN::test_AvgPool3d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9709335Z test_nn.py::TestNN::test_AvgPool3d_stride_pad PASSED [ 3%] 2023-03-31T07:12:13.9709672Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9710018Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_fixedkw_output PASSED [ 3%] 2023-03-31T07:12:13.9710391Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_fixedkw_output_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9710743Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_general_output PASSED [ 3%] 2023-03-31T07:12:13.9711110Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_general_output_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9711484Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_input_nooverlap PASSED [ 3%] 2023-03-31T07:12:13.9711886Z test_nn.py::TestNN::test_AvgPool3d_stride_pad_gpu_input_nooverlap_cuda SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9712235Z test_nn.py::TestNN::test_BCELoss PASSED [ 3%] 2023-03-31T07:12:13.9712548Z test_nn.py::TestNN::test_BCELoss_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9712894Z test_nn.py::TestNN::test_BCELoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9713222Z test_nn.py::TestNN::test_BCELoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9713566Z test_nn.py::TestNN::test_BCELoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9713888Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean PASSED [ 3%] 2023-03-31T07:12:13.9714220Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9714645Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9715022Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 3%] 2023-03-31T07:12:13.9715364Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_none PASSED [ 3%] 2023-03-31T07:12:13.9715697Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9716074Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9716447Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9716787Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum PASSED [ 4%] 2023-03-31T07:12:13.9717118Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9717493Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9717866Z test_nn.py::TestNN::test_BCELoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9718188Z test_nn.py::TestNN::test_BCELoss_no_reduce PASSED [ 4%] 2023-03-31T07:12:13.9718516Z test_nn.py::TestNN::test_BCELoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9718849Z test_nn.py::TestNN::test_BCELoss_no_reduce_scalar PASSED [ 4%] 2023-03-31T07:12:13.9719188Z test_nn.py::TestNN::test_BCELoss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9719556Z test_nn.py::TestNN::test_BCELoss_scalar_weights PASSED [ 4%] 2023-03-31T07:12:13.9719948Z test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9720331Z test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9720710Z test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9721075Z test_nn.py::TestNN::test_BCELoss_scalar_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9721410Z test_nn.py::TestNN::test_BCELoss_weights PASSED [ 4%] 2023-03-31T07:12:13.9721747Z test_nn.py::TestNN::test_BCELoss_weights_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9722098Z test_nn.py::TestNN::test_BCELoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9722464Z test_nn.py::TestNN::test_BCELoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9722829Z test_nn.py::TestNN::test_BCELoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 4%] 2023-03-31T07:12:13.9723370Z test_nn.py::TestNN::test_BCELoss_weights_no_reduce PASSED [ 5%] 2023-03-31T07:12:13.9723757Z test_nn.py::TestNN::test_BCELoss_weights_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9724103Z test_nn.py::TestNN::test_BCELoss_weights_no_reduce_scalar PASSED [ 5%] 2023-03-31T07:12:13.9724460Z test_nn.py::TestNN::test_BCELoss_weights_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9724793Z test_nn.py::TestNN::test_BCEWithLogitsLoss PASSED [ 5%] 2023-03-31T07:12:13.9725145Z test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9725525Z test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9725900Z test_nn.py::TestNN::test_BCEWithLogitsLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9726231Z test_nn.py::TestNN::test_BCEWithLogitsLoss_legacy_enum PASSED [ 5%] 2023-03-31T07:12:13.9726642Z test_nn.py::TestNN::test_BCEWithLogitsLoss_legacy_enum_cuda SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9727002Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean PASSED [ 5%] 2023-03-31T07:12:13.9727379Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9727775Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9728184Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9728548Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none PASSED [ 5%] 2023-03-31T07:12:13.9728910Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9729319Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9729734Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9730104Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum PASSED [ 5%] 2023-03-31T07:12:13.9730465Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9730869Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 5%] 2023-03-31T07:12:13.9731269Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9731682Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce PASSED [ 6%] 2023-03-31T07:12:13.9732057Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9732413Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_scalar PASSED [ 6%] 2023-03-31T07:12:13.9732784Z test_nn.py::TestNN::test_BCEWithLogitsLoss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9733131Z test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights PASSED [ 6%] 2023-03-31T07:12:13.9733507Z test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9733914Z test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9734318Z test_nn.py::TestNN::test_BCEWithLogitsLoss_scalar_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9734665Z test_nn.py::TestNN::test_BCEWithLogitsLoss_weights PASSED [ 6%] 2023-03-31T07:12:13.9735026Z test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9735421Z test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9735842Z test_nn.py::TestNN::test_BCEWithLogitsLoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9736175Z test_nn.py::TestNN::test_BatchNorm1d_3d_input PASSED [ 6%] 2023-03-31T07:12:13.9736507Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9736843Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_eval PASSED [ 6%] 2023-03-31T07:12:13.9737170Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_eval_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9737520Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine PASSED [ 6%] 2023-03-31T07:12:13.9737874Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9738227Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_eval PASSED [ 6%] 2023-03-31T07:12:13.9738612Z test_nn.py::TestNN::test_BatchNorm1d_3d_input_not_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 6%] 2023-03-31T07:12:13.9738959Z test_nn.py::TestNN::test_BatchNorm1d_affine PASSED [ 7%] 2023-03-31T07:12:13.9739288Z test_nn.py::TestNN::test_BatchNorm1d_affine_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9739606Z test_nn.py::TestNN::test_BatchNorm1d_affine_eval PASSED [ 7%] 2023-03-31T07:12:13.9739943Z test_nn.py::TestNN::test_BatchNorm1d_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9740283Z test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average PASSED [ 7%] 2023-03-31T07:12:13.9740655Z test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9740999Z test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_eval PASSED [ 7%] 2023-03-31T07:12:13.9741366Z test_nn.py::TestNN::test_BatchNorm1d_affine_simple_average_eval_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9741720Z test_nn.py::TestNN::test_BatchNorm1d_not_affine PASSED [ 7%] 2023-03-31T07:12:13.9742059Z test_nn.py::TestNN::test_BatchNorm1d_not_affine_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9742384Z test_nn.py::TestNN::test_BatchNorm1d_not_affine_eval PASSED [ 7%] 2023-03-31T07:12:13.9742725Z test_nn.py::TestNN::test_BatchNorm1d_not_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9743067Z test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats PASSED [ 7%] 2023-03-31T07:12:13.9743489Z test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9743905Z test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_eval PASSED [ 7%] 2023-03-31T07:12:13.9744296Z test_nn.py::TestNN::test_BatchNorm1d_not_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9744645Z test_nn.py::TestNN::test_BatchNorm1d_zero_batch PASSED [ 7%] 2023-03-31T07:12:13.9744968Z test_nn.py::TestNN::test_BatchNorm1d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9745304Z test_nn.py::TestNN::test_BatchNorm1d_zero_batch_eval PASSED [ 7%] 2023-03-31T07:12:13.9745647Z test_nn.py::TestNN::test_BatchNorm1d_zero_batch_eval_cuda SKIPPED (Excluded from CUDA tests) [ 7%] 2023-03-31T07:12:13.9745960Z test_nn.py::TestNN::test_BatchNorm2d PASSED [ 7%] 2023-03-31T07:12:13.9746264Z test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average PASSED [ 8%] 2023-03-31T07:12:13.9746609Z test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9746958Z test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_eval PASSED [ 8%] 2023-03-31T07:12:13.9747301Z test_nn.py::TestNN::test_BatchNorm2d_2d_simple_average_eval_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9747668Z test_nn.py::TestNN::test_BatchNorm2d_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9748023Z test_nn.py::TestNN::test_BatchNorm2d_eval PASSED [ 8%] 2023-03-31T07:12:13.9748351Z test_nn.py::TestNN::test_BatchNorm2d_eval_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9748666Z test_nn.py::TestNN::test_BatchNorm2d_momentum PASSED [ 8%] 2023-03-31T07:12:13.9748999Z test_nn.py::TestNN::test_BatchNorm2d_momentum_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9749335Z test_nn.py::TestNN::test_BatchNorm2d_momentum_eval PASSED [ 8%] 2023-03-31T07:12:13.9749666Z test_nn.py::TestNN::test_BatchNorm2d_momentum_eval_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9750006Z test_nn.py::TestNN::test_BatchNorm2d_not_affine PASSED [ 8%] 2023-03-31T07:12:13.9750381Z test_nn.py::TestNN::test_BatchNorm2d_not_affine_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9750718Z test_nn.py::TestNN::test_BatchNorm2d_not_affine_eval PASSED [ 8%] 2023-03-31T07:12:13.9751047Z test_nn.py::TestNN::test_BatchNorm2d_not_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9751390Z test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats PASSED [ 8%] 2023-03-31T07:12:13.9751742Z test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9752082Z test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_eval PASSED [ 8%] 2023-03-31T07:12:13.9752444Z test_nn.py::TestNN::test_BatchNorm2d_not_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9752790Z test_nn.py::TestNN::test_BatchNorm2d_zero_batch PASSED [ 8%] 2023-03-31T07:12:13.9753126Z test_nn.py::TestNN::test_BatchNorm2d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 8%] 2023-03-31T07:12:13.9753447Z test_nn.py::TestNN::test_BatchNorm2d_zero_batch_eval PASSED [ 9%] 2023-03-31T07:12:13.9753793Z test_nn.py::TestNN::test_BatchNorm2d_zero_batch_eval_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9754125Z test_nn.py::TestNN::test_BatchNorm3d PASSED [ 9%] 2023-03-31T07:12:13.9754416Z test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average PASSED [ 9%] 2023-03-31T07:12:13.9754765Z test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9755114Z test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_eval PASSED [ 9%] 2023-03-31T07:12:13.9755507Z test_nn.py::TestNN::test_BatchNorm3d_3d_simple_average_eval_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9755886Z test_nn.py::TestNN::test_BatchNorm3d_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9756208Z test_nn.py::TestNN::test_BatchNorm3d_eval PASSED [ 9%] 2023-03-31T07:12:13.9756535Z test_nn.py::TestNN::test_BatchNorm3d_eval_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9756848Z test_nn.py::TestNN::test_BatchNorm3d_momentum PASSED [ 9%] 2023-03-31T07:12:13.9757178Z test_nn.py::TestNN::test_BatchNorm3d_momentum_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9757511Z test_nn.py::TestNN::test_BatchNorm3d_momentum_eval PASSED [ 9%] 2023-03-31T07:12:13.9757847Z test_nn.py::TestNN::test_BatchNorm3d_momentum_eval_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9758169Z test_nn.py::TestNN::test_BatchNorm3d_not_affine PASSED [ 9%] 2023-03-31T07:12:13.9758504Z test_nn.py::TestNN::test_BatchNorm3d_not_affine_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9758843Z test_nn.py::TestNN::test_BatchNorm3d_not_affine_eval PASSED [ 9%] 2023-03-31T07:12:13.9759185Z test_nn.py::TestNN::test_BatchNorm3d_not_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9759514Z test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats PASSED [ 9%] 2023-03-31T07:12:13.9759942Z test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 9%] 2023-03-31T07:12:13.9760293Z test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_eval PASSED [ 9%] 2023-03-31T07:12:13.9760633Z test_nn.py::TestNN::test_BatchNorm3d_not_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9760980Z test_nn.py::TestNN::test_BatchNorm3d_zero_batch PASSED [ 10%] 2023-03-31T07:12:13.9761314Z test_nn.py::TestNN::test_BatchNorm3d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9761657Z test_nn.py::TestNN::test_BatchNorm3d_zero_batch_eval PASSED [ 10%] 2023-03-31T07:12:13.9761989Z test_nn.py::TestNN::test_BatchNorm3d_zero_batch_eval_cuda SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9762371Z test_nn.py::TestNN::test_CELU_no_batch_dim PASSED [ 10%] 2023-03-31T07:12:13.9762692Z test_nn.py::TestNN::test_CELU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9763163Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists PASSED [ 10%] 2023-03-31T07:12:13.9763597Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_cuda_double SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9764006Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_cuda_float SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9764385Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction PASSED [ 10%] 2023-03-31T07:12:13.9764773Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9765217Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_intlists_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9765598Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors PASSED [ 10%] 2023-03-31T07:12:13.9765976Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_cuda_double SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9766365Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_cuda_float SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9766741Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction PASSED [ 10%] 2023-03-31T07:12:13.9767139Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9767644Z test_nn.py::TestNN::test_CTCLoss_2d_int_target_lengths_tensors_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9768036Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors PASSED [ 10%] 2023-03-31T07:12:13.9768389Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_cuda_double SKIPPED (Excluded from CUDA tests) [ 10%] 2023-03-31T07:12:13.9768774Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9769113Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction PASSED [ 11%] 2023-03-31T07:12:13.9769490Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9769898Z test_nn.py::TestNN::test_CTCLoss_2d_lengths_tensors_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9770275Z test_nn.py::TestNN::test_CTCLoss_critical_target_len SKIPPED (CUDA not available) [ 11%] 2023-03-31T07:12:13.9770600Z test_nn.py::TestNN::test_CTCLoss_lengthchecks_cpu PASSED [ 11%] 2023-03-31T07:12:13.9770934Z test_nn.py::TestNN::test_CTCLoss_lengthchecks_cuda SKIPPED (CUDA not available) [ 11%] 2023-03-31T07:12:13.9771265Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists PASSED [ 11%] 2023-03-31T07:12:13.9771653Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists_cuda_double SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9772023Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9772371Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction PASSED [ 11%] 2023-03-31T07:12:13.9772741Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9773134Z test_nn.py::TestNN::test_CTCLoss_lengths_intlists_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9773492Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors PASSED [ 11%] 2023-03-31T07:12:13.9773838Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors_cuda_double SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9774270Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9774605Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction PASSED [ 11%] 2023-03-31T07:12:13.9774977Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9775382Z test_nn.py::TestNN::test_CTCLoss_lengths_tensors_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 11%] 2023-03-31T07:12:13.9775746Z test_nn.py::TestNN::test_CTCLoss_long_targets SKIPPED (CUDA not available) [ 11%] 2023-03-31T07:12:13.9776050Z test_nn.py::TestNN::test_CTCLoss_typechecks PASSED [ 11%] 2023-03-31T07:12:13.9776373Z test_nn.py::TestNN::test_CTCLoss_zero_infinity SKIPPED (CUDA not available) [ 12%] 2023-03-31T07:12:13.9776685Z test_nn.py::TestNN::test_Conv1d PASSED [ 12%] 2023-03-31T07:12:13.9776973Z test_nn.py::TestNN::test_Conv1d_circular_stride2_pad2 PASSED [ 12%] 2023-03-31T07:12:13.9777323Z test_nn.py::TestNN::test_Conv1d_circular_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9777674Z test_nn.py::TestNN::test_Conv1d_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9777985Z test_nn.py::TestNN::test_Conv1d_dilated PASSED [ 12%] 2023-03-31T07:12:13.9778291Z test_nn.py::TestNN::test_Conv1d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9778614Z test_nn.py::TestNN::test_Conv1d_groups PASSED [ 12%] 2023-03-31T07:12:13.9778930Z test_nn.py::TestNN::test_Conv1d_groups_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9779274Z test_nn.py::TestNN::test_Conv1d_pad1 PASSED [ 12%] 2023-03-31T07:12:13.9779613Z test_nn.py::TestNN::test_Conv1d_pad1_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9779930Z test_nn.py::TestNN::test_Conv1d_pad1size1 PASSED [ 12%] 2023-03-31T07:12:13.9780257Z test_nn.py::TestNN::test_Conv1d_pad1size1_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9780563Z test_nn.py::TestNN::test_Conv1d_pad2 PASSED [ 12%] 2023-03-31T07:12:13.9780876Z test_nn.py::TestNN::test_Conv1d_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9781189Z test_nn.py::TestNN::test_Conv1d_pad2size1 PASSED [ 12%] 2023-03-31T07:12:13.9781495Z test_nn.py::TestNN::test_Conv1d_pad2size1_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9781817Z test_nn.py::TestNN::test_Conv1d_pad_same PASSED [ 12%] 2023-03-31T07:12:13.9782107Z test_nn.py::TestNN::test_Conv1d_pad_same2 PASSED [ 12%] 2023-03-31T07:12:13.9782428Z test_nn.py::TestNN::test_Conv1d_pad_same2_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9782759Z test_nn.py::TestNN::test_Conv1d_pad_same_cuda SKIPPED (Excluded from CUDA tests) [ 12%] 2023-03-31T07:12:13.9783087Z test_nn.py::TestNN::test_Conv1d_pad_same_dilated PASSED [ 12%] 2023-03-31T07:12:13.9783593Z test_nn.py::TestNN::test_Conv1d_pad_same_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9783912Z test_nn.py::TestNN::test_Conv1d_pad_valid PASSED [ 13%] 2023-03-31T07:12:13.9784233Z test_nn.py::TestNN::test_Conv1d_pad_valid_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9784560Z test_nn.py::TestNN::test_Conv1d_reflect_stride2_pad2 PASSED [ 13%] 2023-03-31T07:12:13.9784899Z test_nn.py::TestNN::test_Conv1d_reflect_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9785230Z test_nn.py::TestNN::test_Conv1d_replicate_stride2_pad2 PASSED [ 13%] 2023-03-31T07:12:13.9785579Z test_nn.py::TestNN::test_Conv1d_replicate_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9785914Z test_nn.py::TestNN::test_Conv1d_stride PASSED [ 13%] 2023-03-31T07:12:13.9786257Z test_nn.py::TestNN::test_Conv1d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9786576Z test_nn.py::TestNN::test_Conv1d_zero_batch PASSED [ 13%] 2023-03-31T07:12:13.9786898Z test_nn.py::TestNN::test_Conv1d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9787227Z test_nn.py::TestNN::test_Conv1d_zeros_stride2_pad2 PASSED [ 13%] 2023-03-31T07:12:13.9787550Z test_nn.py::TestNN::test_Conv1d_zeros_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9787875Z test_nn.py::TestNN::test_Conv2d PASSED [ 13%] 2023-03-31T07:12:13.9788179Z test_nn.py::TestNN::test_Conv2d_circular_stride2_pad2 PASSED [ 13%] 2023-03-31T07:12:13.9788508Z test_nn.py::TestNN::test_Conv2d_circular_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9788856Z test_nn.py::TestNN::test_Conv2d_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9789170Z test_nn.py::TestNN::test_Conv2d_depthwise PASSED [ 13%] 2023-03-31T07:12:13.9789497Z test_nn.py::TestNN::test_Conv2d_depthwise_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9789812Z test_nn.py::TestNN::test_Conv2d_depthwise_dilated PASSED [ 13%] 2023-03-31T07:12:13.9790154Z test_nn.py::TestNN::test_Conv2d_depthwise_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 13%] 2023-03-31T07:12:13.9790488Z test_nn.py::TestNN::test_Conv2d_depthwise_padded PASSED [ 14%] 2023-03-31T07:12:13.9790831Z test_nn.py::TestNN::test_Conv2d_depthwise_padded_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9791195Z test_nn.py::TestNN::test_Conv2d_depthwise_strided PASSED [ 14%] 2023-03-31T07:12:13.9791569Z test_nn.py::TestNN::test_Conv2d_depthwise_strided_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9791918Z test_nn.py::TestNN::test_Conv2d_depthwise_with_multiplier PASSED [ 14%] 2023-03-31T07:12:13.9792264Z test_nn.py::TestNN::test_Conv2d_depthwise_with_multiplier_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9792603Z test_nn.py::TestNN::test_Conv2d_dilated PASSED [ 14%] 2023-03-31T07:12:13.9792922Z test_nn.py::TestNN::test_Conv2d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9793253Z test_nn.py::TestNN::test_Conv2d_dilated_with_long_tensor PASSED [ 14%] 2023-03-31T07:12:13.9793590Z test_nn.py::TestNN::test_Conv2d_dilated_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9793922Z test_nn.py::TestNN::test_Conv2d_groups PASSED [ 14%] 2023-03-31T07:12:13.9794244Z test_nn.py::TestNN::test_Conv2d_groups_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9794553Z test_nn.py::TestNN::test_Conv2d_groups_thnn PASSED [ 14%] 2023-03-31T07:12:13.9794877Z test_nn.py::TestNN::test_Conv2d_groups_thnn_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9795260Z test_nn.py::TestNN::test_Conv2d_groups_thnn_with_long_tensor PASSED [ 14%] 2023-03-31T07:12:13.9795619Z test_nn.py::TestNN::test_Conv2d_groups_thnn_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9795950Z test_nn.py::TestNN::test_Conv2d_groups_with_long_tensor PASSED [ 14%] 2023-03-31T07:12:13.9796296Z test_nn.py::TestNN::test_Conv2d_groups_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9796629Z test_nn.py::TestNN::test_Conv2d_no_bias PASSED [ 14%] 2023-03-31T07:12:13.9796929Z test_nn.py::TestNN::test_Conv2d_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 14%] 2023-03-31T07:12:13.9797253Z test_nn.py::TestNN::test_Conv2d_no_bias_with_long_tensor PASSED [ 14%] 2023-03-31T07:12:13.9797598Z test_nn.py::TestNN::test_Conv2d_no_bias_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9797961Z test_nn.py::TestNN::test_Conv2d_pad_same PASSED [ 15%] 2023-03-31T07:12:13.9798268Z test_nn.py::TestNN::test_Conv2d_pad_same_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9798592Z test_nn.py::TestNN::test_Conv2d_pad_same_dilated PASSED [ 15%] 2023-03-31T07:12:13.9798923Z test_nn.py::TestNN::test_Conv2d_pad_same_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9799240Z test_nn.py::TestNN::test_Conv2d_pad_valid PASSED [ 15%] 2023-03-31T07:12:13.9799561Z test_nn.py::TestNN::test_Conv2d_pad_valid_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9799882Z test_nn.py::TestNN::test_Conv2d_padding PASSED [ 15%] 2023-03-31T07:12:13.9800204Z test_nn.py::TestNN::test_Conv2d_padding_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9800521Z test_nn.py::TestNN::test_Conv2d_padding_with_long_tensor PASSED [ 15%] 2023-03-31T07:12:13.9800873Z test_nn.py::TestNN::test_Conv2d_padding_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9801223Z test_nn.py::TestNN::test_Conv2d_reflect_stride2_pad2 PASSED [ 15%] 2023-03-31T07:12:13.9801565Z test_nn.py::TestNN::test_Conv2d_reflect_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9801891Z test_nn.py::TestNN::test_Conv2d_replicate_stride2_pad2 PASSED [ 15%] 2023-03-31T07:12:13.9802239Z test_nn.py::TestNN::test_Conv2d_replicate_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9802569Z test_nn.py::TestNN::test_Conv2d_strided PASSED [ 15%] 2023-03-31T07:12:13.9802914Z test_nn.py::TestNN::test_Conv2d_strided_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9803488Z test_nn.py::TestNN::test_Conv2d_strided_with_long_tensor PASSED [ 15%] 2023-03-31T07:12:13.9803836Z test_nn.py::TestNN::test_Conv2d_strided_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9804177Z test_nn.py::TestNN::test_Conv2d_with_long_tensor PASSED [ 15%] 2023-03-31T07:12:13.9804495Z test_nn.py::TestNN::test_Conv2d_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 15%] 2023-03-31T07:12:13.9804825Z test_nn.py::TestNN::test_Conv2d_zero_batch PASSED [ 16%] 2023-03-31T07:12:13.9805147Z test_nn.py::TestNN::test_Conv2d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9805466Z test_nn.py::TestNN::test_Conv2d_zero_batch_with_long_tensor PASSED [ 16%] 2023-03-31T07:12:13.9805820Z test_nn.py::TestNN::test_Conv2d_zero_batch_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9806164Z test_nn.py::TestNN::test_Conv2d_zeros_stride2_pad2 PASSED [ 16%] 2023-03-31T07:12:13.9806503Z test_nn.py::TestNN::test_Conv2d_zeros_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9806814Z test_nn.py::TestNN::test_Conv3d PASSED [ 16%] 2023-03-31T07:12:13.9807170Z test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias PASSED [ 16%] 2023-03-31T07:12:13.9807491Z test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9807809Z test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_with_long_tensor PASSED [ 16%] 2023-03-31T07:12:13.9808159Z test_nn.py::TestNN::test_Conv3d_1x1x1_no_bias_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9808503Z test_nn.py::TestNN::test_Conv3d_circular_stride2_pad2 PASSED [ 16%] 2023-03-31T07:12:13.9808854Z test_nn.py::TestNN::test_Conv3d_circular_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9809190Z test_nn.py::TestNN::test_Conv3d_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9809496Z test_nn.py::TestNN::test_Conv3d_dilated PASSED [ 16%] 2023-03-31T07:12:13.9809867Z test_nn.py::TestNN::test_Conv3d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9810198Z test_nn.py::TestNN::test_Conv3d_dilated_strided PASSED [ 16%] 2023-03-31T07:12:13.9810516Z test_nn.py::TestNN::test_Conv3d_dilated_strided_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9810841Z test_nn.py::TestNN::test_Conv3d_groups PASSED [ 16%] 2023-03-31T07:12:13.9811159Z test_nn.py::TestNN::test_Conv3d_groups_cuda SKIPPED (Excluded from CUDA tests) [ 16%] 2023-03-31T07:12:13.9811473Z test_nn.py::TestNN::test_Conv3d_groups_with_long_tensor PASSED [ 16%] 2023-03-31T07:12:13.9811819Z test_nn.py::TestNN::test_Conv3d_groups_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9812154Z test_nn.py::TestNN::test_Conv3d_no_bias PASSED [ 17%] 2023-03-31T07:12:13.9812466Z test_nn.py::TestNN::test_Conv3d_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9812783Z test_nn.py::TestNN::test_Conv3d_no_bias_with_long_tensor PASSED [ 17%] 2023-03-31T07:12:13.9813132Z test_nn.py::TestNN::test_Conv3d_no_bias_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9813464Z test_nn.py::TestNN::test_Conv3d_pad_same PASSED [ 17%] 2023-03-31T07:12:13.9813767Z test_nn.py::TestNN::test_Conv3d_pad_same_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9814094Z test_nn.py::TestNN::test_Conv3d_pad_same_dilated PASSED [ 17%] 2023-03-31T07:12:13.9814426Z test_nn.py::TestNN::test_Conv3d_pad_same_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9814797Z test_nn.py::TestNN::test_Conv3d_pad_valid PASSED [ 17%] 2023-03-31T07:12:13.9815146Z test_nn.py::TestNN::test_Conv3d_pad_valid_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9815476Z test_nn.py::TestNN::test_Conv3d_replicate_stride2_pad2 PASSED [ 17%] 2023-03-31T07:12:13.9815822Z test_nn.py::TestNN::test_Conv3d_replicate_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9816147Z test_nn.py::TestNN::test_Conv3d_stride PASSED [ 17%] 2023-03-31T07:12:13.9816462Z test_nn.py::TestNN::test_Conv3d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9816781Z test_nn.py::TestNN::test_Conv3d_stride_padding PASSED [ 17%] 2023-03-31T07:12:13.9817115Z test_nn.py::TestNN::test_Conv3d_stride_padding_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9817445Z test_nn.py::TestNN::test_Conv3d_stride_padding_with_long_tensor PASSED [ 17%] 2023-03-31T07:12:13.9817807Z test_nn.py::TestNN::test_Conv3d_stride_padding_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9818164Z test_nn.py::TestNN::test_Conv3d_stride_with_long_tensor PASSED [ 17%] 2023-03-31T07:12:13.9818497Z test_nn.py::TestNN::test_Conv3d_stride_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 17%] 2023-03-31T07:12:13.9818865Z test_nn.py::TestNN::test_Conv3d_with_long_tensor PASSED [ 18%] 2023-03-31T07:12:13.9819193Z test_nn.py::TestNN::test_Conv3d_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9819517Z test_nn.py::TestNN::test_Conv3d_zero_batch PASSED [ 18%] 2023-03-31T07:12:13.9819828Z test_nn.py::TestNN::test_Conv3d_zero_batch_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9820158Z test_nn.py::TestNN::test_Conv3d_zero_batch_with_long_tensor PASSED [ 18%] 2023-03-31T07:12:13.9820514Z test_nn.py::TestNN::test_Conv3d_zero_batch_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9820855Z test_nn.py::TestNN::test_Conv3d_zeros_stride2_pad2 PASSED [ 18%] 2023-03-31T07:12:13.9821178Z test_nn.py::TestNN::test_Conv3d_zeros_stride2_pad2_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9821545Z test_nn.py::TestNN::test_ConvTranspose1d PASSED [ 18%] 2023-03-31T07:12:13.9821875Z test_nn.py::TestNN::test_ConvTranspose1d_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9822199Z test_nn.py::TestNN::test_ConvTranspose1d_dilated PASSED [ 18%] 2023-03-31T07:12:13.9822542Z test_nn.py::TestNN::test_ConvTranspose1d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9822882Z test_nn.py::TestNN::test_ConvTranspose1d_groups PASSED [ 18%] 2023-03-31T07:12:13.9823221Z test_nn.py::TestNN::test_ConvTranspose1d_groups_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9823642Z test_nn.py::TestNN::test_ConvTranspose1d_no_bias PASSED [ 18%] 2023-03-31T07:12:13.9823986Z test_nn.py::TestNN::test_ConvTranspose1d_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9824316Z test_nn.py::TestNN::test_ConvTranspose2d PASSED [ 18%] 2023-03-31T07:12:13.9824631Z test_nn.py::TestNN::test_ConvTranspose2d_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9824968Z test_nn.py::TestNN::test_ConvTranspose2d_dilated PASSED [ 18%] 2023-03-31T07:12:13.9825305Z test_nn.py::TestNN::test_ConvTranspose2d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 18%] 2023-03-31T07:12:13.9825652Z test_nn.py::TestNN::test_ConvTranspose2d_dilated_with_long_tensor PASSED [ 18%] 2023-03-31T07:12:13.9826011Z test_nn.py::TestNN::test_ConvTranspose2d_dilated_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9826369Z test_nn.py::TestNN::test_ConvTranspose2d_groups PASSED [ 19%] 2023-03-31T07:12:13.9826769Z test_nn.py::TestNN::test_ConvTranspose2d_groups_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9827136Z test_nn.py::TestNN::test_ConvTranspose2d_groups_with_long_tensor PASSED [ 19%] 2023-03-31T07:12:13.9827506Z test_nn.py::TestNN::test_ConvTranspose2d_groups_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9827869Z test_nn.py::TestNN::test_ConvTranspose2d_no_bias PASSED [ 19%] 2023-03-31T07:12:13.9828209Z test_nn.py::TestNN::test_ConvTranspose2d_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9828544Z test_nn.py::TestNN::test_ConvTranspose2d_no_bias_with_long_tensor PASSED [ 19%] 2023-03-31T07:12:13.9828912Z test_nn.py::TestNN::test_ConvTranspose2d_no_bias_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9829272Z test_nn.py::TestNN::test_ConvTranspose2d_with_long_tensor PASSED [ 19%] 2023-03-31T07:12:13.9829636Z test_nn.py::TestNN::test_ConvTranspose2d_with_long_tensor_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9829967Z test_nn.py::TestNN::test_ConvTranspose3d PASSED [ 19%] 2023-03-31T07:12:13.9830293Z test_nn.py::TestNN::test_ConvTranspose3d_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9830631Z test_nn.py::TestNN::test_ConvTranspose3d_dilated PASSED [ 19%] 2023-03-31T07:12:13.9830993Z test_nn.py::TestNN::test_ConvTranspose3d_dilated_cuda SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9831340Z test_nn.py::TestNN::test_CosineEmbeddingLoss PASSED [ 19%] 2023-03-31T07:12:13.9831691Z test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9832073Z test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9832438Z test_nn.py::TestNN::test_CosineEmbeddingLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9832790Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin PASSED [ 19%] 2023-03-31T07:12:13.9833159Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_double SKIPPED (Excluded from CUDA tests) [ 19%] 2023-03-31T07:12:13.9833576Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9833973Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_cuda_half SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9834339Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction PASSED [ 20%] 2023-03-31T07:12:13.9834730Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9835143Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9835572Z test_nn.py::TestNN::test_CosineEmbeddingLoss_margin_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9835956Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean PASSED [ 20%] 2023-03-31T07:12:13.9836338Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9836744Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9837161Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9837534Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none PASSED [ 20%] 2023-03-31T07:12:13.9837902Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9838321Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9838769Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9839175Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum PASSED [ 20%] 2023-03-31T07:12:13.9839537Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9839954Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9840365Z test_nn.py::TestNN::test_CosineEmbeddingLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9840737Z test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction PASSED [ 20%] 2023-03-31T07:12:13.9841103Z test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9841515Z test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 20%] 2023-03-31T07:12:13.9841926Z test_nn.py::TestNN::test_CosineEmbeddingLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9842288Z test_nn.py::TestNN::test_CrossEntropyLoss PASSED [ 21%] 2023-03-31T07:12:13.9842583Z test_nn.py::TestNN::test_CrossEntropyLoss_2d PASSED [ 21%] 2023-03-31T07:12:13.9842957Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_double SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9843559Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_float SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9843917Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_cuda_half SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9844264Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index PASSED [ 21%] 2023-03-31T07:12:13.9844633Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9845040Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9845429Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9845878Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing PASSED [ 21%] 2023-03-31T07:12:13.9846276Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_double SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9846707Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_float SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9847119Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_cuda_half SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9847518Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index PASSED [ 21%] 2023-03-31T07:12:13.9847948Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9848405Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9848849Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9849264Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction PASSED [ 21%] 2023-03-31T07:12:13.9849689Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9850142Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 21%] 2023-03-31T07:12:13.9850580Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9851028Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight PASSED [ 22%] 2023-03-31T07:12:13.9851504Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_double SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9851949Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_float SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9852377Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9852761Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target PASSED [ 22%] 2023-03-31T07:12:13.9853133Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_double SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9853519Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_float SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9853921Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9854292Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing PASSED [ 22%] 2023-03-31T07:12:13.9854683Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_double SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9855132Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_float SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9855555Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9855946Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction PASSED [ 22%] 2023-03-31T07:12:13.9856360Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9856796Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9857251Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9857681Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight PASSED [ 22%] 2023-03-31T07:12:13.9858086Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_double SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9858514Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_float SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9858950Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda_half SKIPPED (Excluded from CUDA tests) [ 22%] 2023-03-31T07:12:13.9859335Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction PASSED [ 23%] 2023-03-31T07:12:13.9859726Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9860142Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9860574Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9860953Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights PASSED [ 23%] 2023-03-31T07:12:13.9861323Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9861737Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9862154Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9862544Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction PASSED [ 23%] 2023-03-31T07:12:13.9862980Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9863531Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9863983Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_prob_target_weights_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9864370Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction PASSED [ 23%] 2023-03-31T07:12:13.9864729Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9865129Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9865528Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9865893Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights PASSED [ 23%] 2023-03-31T07:12:13.9866245Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9866641Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9867075Z test_nn.py::TestNN::test_CrossEntropyLoss_2d_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 23%] 2023-03-31T07:12:13.9867431Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing PASSED [ 23%] 2023-03-31T07:12:13.9867827Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9868258Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_float SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9868683Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_cuda_half SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9869071Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index PASSED [ 24%] 2023-03-31T07:12:13.9869489Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9869984Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9870432Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9870832Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction PASSED [ 24%] 2023-03-31T07:12:13.9871252Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9871710Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9872164Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9872583Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index PASSED [ 24%] 2023-03-31T07:12:13.9873031Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9873510Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9873988Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9874406Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target PASSED [ 24%] 2023-03-31T07:12:13.9874804Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9875203Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_float SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9875607Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_cuda_half SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9875960Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing PASSED [ 24%] 2023-03-31T07:12:13.9876349Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_double SKIPPED (Excluded from CUDA tests) [ 24%] 2023-03-31T07:12:13.9876769Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9877186Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9877569Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction PASSED [ 25%] 2023-03-31T07:12:13.9877986Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9878475Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9878921Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9879301Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction PASSED [ 25%] 2023-03-31T07:12:13.9879696Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9880122Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9880549Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9880914Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights PASSED [ 25%] 2023-03-31T07:12:13.9881335Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9881758Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9882156Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9882542Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction PASSED [ 25%] 2023-03-31T07:12:13.9882951Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9883649Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9884084Z test_nn.py::TestNN::test_CrossEntropyLoss_3d_prob_target_weights_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9884470Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target PASSED [ 25%] 2023-03-31T07:12:13.9884846Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_double SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9885249Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_float SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9885632Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_cuda_half SKIPPED (Excluded from CUDA tests) [ 25%] 2023-03-31T07:12:13.9886004Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction PASSED [ 26%] 2023-03-31T07:12:13.9886476Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9886950Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9887369Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9887751Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights PASSED [ 26%] 2023-03-31T07:12:13.9888138Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9888541Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9888956Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9889340Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction PASSED [ 26%] 2023-03-31T07:12:13.9889757Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9890193Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9890683Z test_nn.py::TestNN::test_CrossEntropyLoss_4d_prob_target_weights_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9891093Z test_nn.py::TestNN::test_CrossEntropyLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9891463Z test_nn.py::TestNN::test_CrossEntropyLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9891818Z test_nn.py::TestNN::test_CrossEntropyLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9892158Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3 PASSED [ 26%] 2023-03-31T07:12:13.9892513Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9892902Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_float SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9893314Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_cuda_half SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9893675Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction PASSED [ 26%] 2023-03-31T07:12:13.9894053Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 26%] 2023-03-31T07:12:13.9894452Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9894868Z test_nn.py::TestNN::test_CrossEntropyLoss_dim_is_3_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9895226Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim PASSED [ 27%] 2023-03-31T07:12:13.9895592Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_double SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9895980Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_float SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9896375Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_cuda_half SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9896743Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction PASSED [ 27%] 2023-03-31T07:12:13.9897132Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9897542Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9897960Z test_nn.py::TestNN::test_CrossEntropyLoss_higher_dim_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9898356Z test_nn.py::TestNN::test_CrossEntropyLoss_weights PASSED [ 27%] 2023-03-31T07:12:13.9898747Z test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9899122Z test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9899512Z test_nn.py::TestNN::test_CrossEntropyLoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9899855Z test_nn.py::TestNN::test_CrossMapLRN2d PASSED [ 27%] 2023-03-31T07:12:13.9900166Z test_nn.py::TestNN::test_CrossMapLRN2d_cuda SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9900487Z test_nn.py::TestNN::test_ELU_no_batch_dim PASSED [ 27%] 2023-03-31T07:12:13.9900806Z test_nn.py::TestNN::test_ELU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9901133Z test_nn.py::TestNN::test_Embedding PASSED [ 27%] 2023-03-31T07:12:13.9901428Z test_nn.py::TestNN::test_EmbeddingBag_discontiguous PASSED [ 27%] 2023-03-31T07:12:13.9901779Z test_nn.py::TestNN::test_EmbeddingBag_discontiguous_cuda SKIPPED (Excluded from CUDA tests) [ 27%] 2023-03-31T07:12:13.9902126Z test_nn.py::TestNN::test_EmbeddingBag_max PASSED [ 28%] 2023-03-31T07:12:13.9902476Z test_nn.py::TestNN::test_EmbeddingBag_max_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9902816Z test_nn.py::TestNN::test_EmbeddingBag_max_padding_idx PASSED [ 28%] 2023-03-31T07:12:13.9903174Z test_nn.py::TestNN::test_EmbeddingBag_max_padding_idx_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9903586Z test_nn.py::TestNN::test_EmbeddingBag_mean PASSED [ 28%] 2023-03-31T07:12:13.9903912Z test_nn.py::TestNN::test_EmbeddingBag_mean_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9904245Z test_nn.py::TestNN::test_EmbeddingBag_mean_padding_idx PASSED [ 28%] 2023-03-31T07:12:13.9904587Z test_nn.py::TestNN::test_EmbeddingBag_mean_padding_idx_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9904933Z test_nn.py::TestNN::test_EmbeddingBag_sparse PASSED [ 28%] 2023-03-31T07:12:13.9905311Z test_nn.py::TestNN::test_EmbeddingBag_sparse_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9905644Z test_nn.py::TestNN::test_EmbeddingBag_sum PASSED [ 28%] 2023-03-31T07:12:13.9905960Z test_nn.py::TestNN::test_EmbeddingBag_sum_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9906292Z test_nn.py::TestNN::test_EmbeddingBag_sum_padding_idx PASSED [ 28%] 2023-03-31T07:12:13.9906640Z test_nn.py::TestNN::test_EmbeddingBag_sum_padding_idx_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9906979Z test_nn.py::TestNN::test_Embedding_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9907302Z test_nn.py::TestNN::test_Embedding_discontiguous PASSED [ 28%] 2023-03-31T07:12:13.9907646Z test_nn.py::TestNN::test_Embedding_discontiguous_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9907982Z test_nn.py::TestNN::test_Embedding_sparse PASSED [ 28%] 2023-03-31T07:12:13.9908297Z test_nn.py::TestNN::test_Embedding_sparse_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9908613Z test_nn.py::TestNN::test_Flatten PASSED [ 28%] 2023-03-31T07:12:13.9908917Z test_nn.py::TestNN::test_Flatten_cuda SKIPPED (Excluded from CUDA tests) [ 28%] 2023-03-31T07:12:13.9909227Z test_nn.py::TestNN::test_Flatten_no_batch_dim PASSED [ 29%] 2023-03-31T07:12:13.9909539Z test_nn.py::TestNN::test_Flatten_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9909851Z test_nn.py::TestNN::test_Fold PASSED [ 29%] 2023-03-31T07:12:13.9910147Z test_nn.py::TestNN::test_Fold_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9910477Z test_nn.py::TestNN::test_Fold_int_input PASSED [ 29%] 2023-03-31T07:12:13.9910820Z test_nn.py::TestNN::test_Fold_int_input_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9911138Z test_nn.py::TestNN::test_Fold_no_batch_dim_input PASSED [ 29%] 2023-03-31T07:12:13.9911469Z test_nn.py::TestNN::test_Fold_no_batch_dim_input_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9911787Z test_nn.py::TestNN::test_Fold_no_batch_dim_int_input PASSED [ 29%] 2023-03-31T07:12:13.9912126Z test_nn.py::TestNN::test_Fold_no_batch_dim_int_input_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9912468Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio PASSED [ 29%] 2023-03-31T07:12:13.9912807Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9913161Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim PASSED [ 29%] 2023-03-31T07:12:13.9913536Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9913927Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples PASSED [ 29%] 2023-03-31T07:12:13.9914361Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9914750Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_return_indices PASSED [ 29%] 2023-03-31T07:12:13.9915130Z test_nn.py::TestNN::test_FractionalMaxPool2d_ratio_return_indices_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9915483Z test_nn.py::TestNN::test_FractionalMaxPool2d_size PASSED [ 29%] 2023-03-31T07:12:13.9915833Z test_nn.py::TestNN::test_FractionalMaxPool2d_size_cuda SKIPPED (Excluded from CUDA tests) [ 29%] 2023-03-31T07:12:13.9916185Z test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim PASSED [ 29%] 2023-03-31T07:12:13.9916555Z test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9916932Z test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_no_random_samples PASSED [ 30%] 2023-03-31T07:12:13.9917380Z test_nn.py::TestNN::test_FractionalMaxPool2d_size_no_batch_dim_no_random_samples_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9917761Z test_nn.py::TestNN::test_FractionalMaxPool3d_asymsize PASSED [ 30%] 2023-03-31T07:12:13.9918119Z test_nn.py::TestNN::test_FractionalMaxPool3d_asymsize_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9918461Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio PASSED [ 30%] 2023-03-31T07:12:13.9918812Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9919165Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim PASSED [ 30%] 2023-03-31T07:12:13.9919525Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9919916Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples PASSED [ 30%] 2023-03-31T07:12:13.9920329Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9920718Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_return_indices PASSED [ 30%] 2023-03-31T07:12:13.9921086Z test_nn.py::TestNN::test_FractionalMaxPool3d_ratio_return_indices_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9921449Z test_nn.py::TestNN::test_FractionalMaxPool3d_size PASSED [ 30%] 2023-03-31T07:12:13.9921800Z test_nn.py::TestNN::test_FractionalMaxPool3d_size_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9922153Z test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim PASSED [ 30%] 2023-03-31T07:12:13.9922543Z test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9922957Z test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_no_random_samples PASSED [ 30%] 2023-03-31T07:12:13.9923620Z test_nn.py::TestNN::test_FractionalMaxPool3d_size_no_batch_dim_no_random_samples_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9923977Z test_nn.py::TestNN::test_GELU_no_batch_dim PASSED [ 30%] 2023-03-31T07:12:13.9924297Z test_nn.py::TestNN::test_GELU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 30%] 2023-03-31T07:12:13.9924615Z test_nn.py::TestNN::test_GLU_no_batch_dim PASSED [ 31%] 2023-03-31T07:12:13.9924932Z test_nn.py::TestNN::test_GLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9925244Z test_nn.py::TestNN::test_GroupNorm_1d_affine PASSED [ 31%] 2023-03-31T07:12:13.9925546Z test_nn.py::TestNN::test_GroupNorm_1d_affine_GN PASSED [ 31%] 2023-03-31T07:12:13.9925880Z test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9926194Z test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_eval PASSED [ 31%] 2023-03-31T07:12:13.9926534Z test_nn.py::TestNN::test_GroupNorm_1d_affine_GN_eval_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9926974Z test_nn.py::TestNN::test_GroupNorm_1d_affine_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9927304Z test_nn.py::TestNN::test_GroupNorm_1d_affine_eval PASSED [ 31%] 2023-03-31T07:12:13.9927628Z test_nn.py::TestNN::test_GroupNorm_1d_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9927997Z test_nn.py::TestNN::test_GroupNorm_1d_affine_large_batch_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9928381Z test_nn.py::TestNN::test_GroupNorm_1d_affine_large_batch_eval_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9928727Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN PASSED [ 31%] 2023-03-31T07:12:13.9929052Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9929440Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_eval PASSED [ 31%] 2023-03-31T07:12:13.9929786Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_IN_eval_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9930108Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN PASSED [ 31%] 2023-03-31T07:12:13.9930443Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9930777Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_eval PASSED [ 31%] 2023-03-31T07:12:13.9931119Z test_nn.py::TestNN::test_GroupNorm_1d_no_affine_LN_eval_cuda SKIPPED (Excluded from CUDA tests) [ 31%] 2023-03-31T07:12:13.9931443Z test_nn.py::TestNN::test_GroupNorm_2d_affine PASSED [ 31%] 2023-03-31T07:12:13.9931770Z test_nn.py::TestNN::test_GroupNorm_2d_affine_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9932096Z test_nn.py::TestNN::test_GroupNorm_2d_affine_eval PASSED [ 32%] 2023-03-31T07:12:13.9932417Z test_nn.py::TestNN::test_GroupNorm_2d_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9932793Z test_nn.py::TestNN::test_GroupNorm_2d_affine_large_feature_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9933181Z test_nn.py::TestNN::test_GroupNorm_2d_affine_large_feature_eval_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9933527Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN PASSED [ 32%] 2023-03-31T07:12:13.9933851Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9934186Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_eval PASSED [ 32%] 2023-03-31T07:12:13.9934577Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_IN_eval_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9934946Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN PASSED [ 32%] 2023-03-31T07:12:13.9935269Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9935606Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_eval PASSED [ 32%] 2023-03-31T07:12:13.9935948Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_LN_eval_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9936315Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_large_feature_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9936710Z test_nn.py::TestNN::test_GroupNorm_2d_no_affine_large_feature_eval_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9937057Z test_nn.py::TestNN::test_Hardshrink_no_batch_dim PASSED [ 32%] 2023-03-31T07:12:13.9937395Z test_nn.py::TestNN::test_Hardshrink_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9937720Z test_nn.py::TestNN::test_Hardsigmoid_no_batch_dim PASSED [ 32%] 2023-03-31T07:12:13.9938068Z test_nn.py::TestNN::test_Hardsigmoid_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9938434Z test_nn.py::TestNN::test_Hardswish_no_batch_dim PASSED [ 32%] 2023-03-31T07:12:13.9938755Z test_nn.py::TestNN::test_Hardswish_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 32%] 2023-03-31T07:12:13.9939086Z test_nn.py::TestNN::test_Hardtanh_no_batch_dim PASSED [ 33%] 2023-03-31T07:12:13.9939415Z test_nn.py::TestNN::test_Hardtanh_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9939749Z test_nn.py::TestNN::test_HingeEmbeddingLoss PASSED [ 33%] 2023-03-31T07:12:13.9940083Z test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9940463Z test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9940841Z test_nn.py::TestNN::test_HingeEmbeddingLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9941229Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin PASSED [ 33%] 2023-03-31T07:12:13.9941579Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_double SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9941971Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_float SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9942361Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_cuda_half SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9942708Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_no_reduce PASSED [ 33%] 2023-03-31T07:12:13.9943077Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9943529Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction PASSED [ 33%] 2023-03-31T07:12:13.9943918Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9944330Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9944752Z test_nn.py::TestNN::test_HingeEmbeddingLoss_margin_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9945124Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean PASSED [ 33%] 2023-03-31T07:12:13.9945504Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9945906Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9946387Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 33%] 2023-03-31T07:12:13.9946781Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none PASSED [ 33%] 2023-03-31T07:12:13.9947144Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9947561Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9947968Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9948333Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum PASSED [ 34%] 2023-03-31T07:12:13.9948695Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9949106Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9949512Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9949879Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_reduce PASSED [ 34%] 2023-03-31T07:12:13.9950227Z test_nn.py::TestNN::test_HingeEmbeddingLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9950619Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin PASSED [ 34%] 2023-03-31T07:12:13.9950992Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_double SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9951391Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_float SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9951794Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_cuda_half SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9952168Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction PASSED [ 34%] 2023-03-31T07:12:13.9952572Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9952995Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9953458Z test_nn.py::TestNN::test_HingeEmbeddingLoss_scalar_margin_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9953837Z test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction PASSED [ 34%] 2023-03-31T07:12:13.9954211Z test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9954605Z test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9955006Z test_nn.py::TestNN::test_HingeEmbeddingLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 34%] 2023-03-31T07:12:13.9955353Z test_nn.py::TestNN::test_HuberLoss PASSED [ 35%] 2023-03-31T07:12:13.9955680Z test_nn.py::TestNN::test_HuberLoss_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9956019Z test_nn.py::TestNN::test_HuberLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9956370Z test_nn.py::TestNN::test_HuberLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9956715Z test_nn.py::TestNN::test_HuberLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9957023Z test_nn.py::TestNN::test_HuberLoss_delta PASSED [ 35%] 2023-03-31T07:12:13.9957343Z test_nn.py::TestNN::test_HuberLoss_delta_cuda SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9957667Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean PASSED [ 35%] 2023-03-31T07:12:13.9958014Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9958415Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9958821Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9959164Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none PASSED [ 35%] 2023-03-31T07:12:13.9959504Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9959885Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9960259Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9960598Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum PASSED [ 35%] 2023-03-31T07:12:13.9960932Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9961307Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9961683Z test_nn.py::TestNN::test_HuberLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9962024Z test_nn.py::TestNN::test_HuberLoss_sum_reduction PASSED [ 35%] 2023-03-31T07:12:13.9962397Z test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 35%] 2023-03-31T07:12:13.9962776Z test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9963279Z test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9963652Z test_nn.py::TestNN::test_HuberLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9963977Z test_nn.py::TestNN::test_InstanceNorm1d PASSED [ 36%] 2023-03-31T07:12:13.9964306Z test_nn.py::TestNN::test_InstanceNorm1d_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9964633Z test_nn.py::TestNN::test_InstanceNorm1d_eval PASSED [ 36%] 2023-03-31T07:12:13.9965006Z test_nn.py::TestNN::test_InstanceNorm1d_eval_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9965342Z test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim PASSED [ 36%] 2023-03-31T07:12:13.9965689Z test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9966033Z test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_eval PASSED [ 36%] 2023-03-31T07:12:13.9966369Z test_nn.py::TestNN::test_InstanceNorm1d_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9966715Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats PASSED [ 36%] 2023-03-31T07:12:13.9967066Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9967404Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_eval PASSED [ 36%] 2023-03-31T07:12:13.9967763Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9968128Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim PASSED [ 36%] 2023-03-31T07:12:13.9968501Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9968863Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_eval PASSED [ 36%] 2023-03-31T07:12:13.9969247Z test_nn.py::TestNN::test_InstanceNorm1d_tracking_stats_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9969602Z test_nn.py::TestNN::test_InstanceNorm2d PASSED [ 36%] 2023-03-31T07:12:13.9969913Z test_nn.py::TestNN::test_InstanceNorm2d_cuda SKIPPED (Excluded from CUDA tests) [ 36%] 2023-03-31T07:12:13.9970281Z test_nn.py::TestNN::test_InstanceNorm2d_eval PASSED [ 37%] 2023-03-31T07:12:13.9970654Z test_nn.py::TestNN::test_InstanceNorm2d_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9970988Z test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim PASSED [ 37%] 2023-03-31T07:12:13.9971322Z test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9971667Z test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_eval PASSED [ 37%] 2023-03-31T07:12:13.9972022Z test_nn.py::TestNN::test_InstanceNorm2d_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9972372Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats PASSED [ 37%] 2023-03-31T07:12:13.9972708Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9973057Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_eval PASSED [ 37%] 2023-03-31T07:12:13.9973415Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9973767Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim PASSED [ 37%] 2023-03-31T07:12:13.9974142Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9974560Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_eval PASSED [ 37%] 2023-03-31T07:12:13.9974949Z test_nn.py::TestNN::test_InstanceNorm2d_tracking_stats_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9975290Z test_nn.py::TestNN::test_InstanceNorm3d PASSED [ 37%] 2023-03-31T07:12:13.9975610Z test_nn.py::TestNN::test_InstanceNorm3d_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9975935Z test_nn.py::TestNN::test_InstanceNorm3d_eval PASSED [ 37%] 2023-03-31T07:12:13.9976252Z test_nn.py::TestNN::test_InstanceNorm3d_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9976592Z test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim PASSED [ 37%] 2023-03-31T07:12:13.9976967Z test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9977313Z test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_eval PASSED [ 37%] 2023-03-31T07:12:13.9977651Z test_nn.py::TestNN::test_InstanceNorm3d_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 37%] 2023-03-31T07:12:13.9977998Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats PASSED [ 38%] 2023-03-31T07:12:13.9978344Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_cuda SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9978694Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_eval PASSED [ 38%] 2023-03-31T07:12:13.9979040Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_eval_cuda SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9979403Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim PASSED [ 38%] 2023-03-31T07:12:13.9979777Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9980136Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_eval PASSED [ 38%] 2023-03-31T07:12:13.9980521Z test_nn.py::TestNN::test_InstanceNorm3d_tracking_stats_no_batch_dim_eval_cuda SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9980867Z test_nn.py::TestNN::test_KLDivLoss PASSED [ 38%] 2023-03-31T07:12:13.9981161Z test_nn.py::TestNN::test_KLDivLoss_batch_mean PASSED [ 38%] 2023-03-31T07:12:13.9981454Z test_nn.py::TestNN::test_KLDivLoss_batch_mean_log_target PASSED [ 38%] 2023-03-31T07:12:13.9981784Z test_nn.py::TestNN::test_KLDivLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9982165Z test_nn.py::TestNN::test_KLDivLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9982519Z test_nn.py::TestNN::test_KLDivLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9982844Z test_nn.py::TestNN::test_KLDivLoss_log_target PASSED [ 38%] 2023-03-31T07:12:13.9983188Z test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_double SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9983649Z test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_float SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9984007Z test_nn.py::TestNN::test_KLDivLoss_log_target_cuda_half SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9984350Z test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction PASSED [ 38%] 2023-03-31T07:12:13.9984715Z test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9985122Z test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 38%] 2023-03-31T07:12:13.9985509Z test_nn.py::TestNN::test_KLDivLoss_log_target_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9985859Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean PASSED [ 39%] 2023-03-31T07:12:13.9986248Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9986619Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9986996Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9987341Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none PASSED [ 39%] 2023-03-31T07:12:13.9987689Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9988058Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9988442Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9988818Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum PASSED [ 39%] 2023-03-31T07:12:13.9989165Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9989533Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9989906Z test_nn.py::TestNN::test_KLDivLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9990244Z test_nn.py::TestNN::test_KLDivLoss_no_reduce PASSED [ 39%] 2023-03-31T07:12:13.9990562Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9990903Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_log_target PASSED [ 39%] 2023-03-31T07:12:13.9991253Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_log_target_cuda SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9991597Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar PASSED [ 39%] 2023-03-31T07:12:13.9991925Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9992272Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_log_target PASSED [ 39%] 2023-03-31T07:12:13.9992633Z test_nn.py::TestNN::test_KLDivLoss_no_reduce_scalar_log_target_cuda SKIPPED (Excluded from CUDA tests) [ 39%] 2023-03-31T07:12:13.9992977Z test_nn.py::TestNN::test_KLDivLoss_scalar PASSED [ 40%] 2023-03-31T07:12:13.9993297Z test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_double SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9993661Z test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_float SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9994052Z test_nn.py::TestNN::test_KLDivLoss_scalar_cuda_half SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9994399Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target PASSED [ 40%] 2023-03-31T07:12:13.9994755Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_double SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9995150Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_float SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9995532Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_cuda_half SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9995876Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction PASSED [ 40%] 2023-03-31T07:12:13.9996251Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9996672Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9997090Z test_nn.py::TestNN::test_KLDivLoss_scalar_log_target_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9997443Z test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction PASSED [ 40%] 2023-03-31T07:12:13.9997803Z test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9998235Z test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9998612Z test_nn.py::TestNN::test_KLDivLoss_scalar_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9998960Z test_nn.py::TestNN::test_KLDivLoss_sum_reduction PASSED [ 40%] 2023-03-31T07:12:13.9999305Z test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:13.9999685Z test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:14.0000052Z test_nn.py::TestNN::test_KLDivLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:12:14.0000398Z test_nn.py::TestNN::test_KLDivLoss_with_log_target_no_reduce PASSED [ 40%] 2023-03-31T07:12:14.0000788Z test_nn.py::TestNN::test_KLDivLoss_with_log_target_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0001139Z test_nn.py::TestNN::test_KLDivLoss_with_target_no_reduce PASSED [ 41%] 2023-03-31T07:12:14.0001474Z test_nn.py::TestNN::test_KLDivLoss_with_target_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0001802Z test_nn.py::TestNN::test_L1Loss PASSED [ 41%] 2023-03-31T07:12:14.0002115Z test_nn.py::TestNN::test_L1Loss_cuda_cdouble SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0002443Z test_nn.py::TestNN::test_L1Loss_cuda_cfloat SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0002784Z test_nn.py::TestNN::test_L1Loss_cuda_double SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0003246Z test_nn.py::TestNN::test_L1Loss_cuda_float SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0003584Z test_nn.py::TestNN::test_L1Loss_cuda_half SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0003892Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean PASSED [ 41%] 2023-03-31T07:12:14.0004233Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0004606Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0004964Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0005295Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_none PASSED [ 41%] 2023-03-31T07:12:14.0005687Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0006095Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0006446Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0006778Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum PASSED [ 41%] 2023-03-31T07:12:14.0007119Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0007486Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0007838Z test_nn.py::TestNN::test_L1Loss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:12:14.0008166Z test_nn.py::TestNN::test_L1Loss_no_reduce PASSED [ 42%] 2023-03-31T07:12:14.0008466Z test_nn.py::TestNN::test_L1Loss_no_reduce_complex PASSED [ 42%] 2023-03-31T07:12:14.0008790Z test_nn.py::TestNN::test_L1Loss_no_reduce_complex_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0009147Z test_nn.py::TestNN::test_L1Loss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0009473Z test_nn.py::TestNN::test_L1Loss_no_reduce_scalar PASSED [ 42%] 2023-03-31T07:12:14.0009842Z test_nn.py::TestNN::test_L1Loss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0010153Z test_nn.py::TestNN::test_L1Loss_scalar PASSED [ 42%] 2023-03-31T07:12:14.0010478Z test_nn.py::TestNN::test_L1Loss_scalar_cuda_cdouble SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0010836Z test_nn.py::TestNN::test_L1Loss_scalar_cuda_cfloat SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0011179Z test_nn.py::TestNN::test_L1Loss_scalar_cuda_double SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0011527Z test_nn.py::TestNN::test_L1Loss_scalar_cuda_float SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0011878Z test_nn.py::TestNN::test_L1Loss_scalar_cuda_half SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0012192Z test_nn.py::TestNN::test_LPPool1d PASSED [ 42%] 2023-03-31T07:12:14.0012547Z test_nn.py::TestNN::test_LPPool1d_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0012864Z test_nn.py::TestNN::test_LPPool1d_no_batch_dim PASSED [ 42%] 2023-03-31T07:12:14.0013195Z test_nn.py::TestNN::test_LPPool1d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0013507Z test_nn.py::TestNN::test_LPPool1d_norm PASSED [ 42%] 2023-03-31T07:12:14.0013825Z test_nn.py::TestNN::test_LPPool1d_norm_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0014138Z test_nn.py::TestNN::test_LPPool2d PASSED [ 42%] 2023-03-31T07:12:14.0014448Z test_nn.py::TestNN::test_LPPool2d_cuda SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:12:14.0014751Z test_nn.py::TestNN::test_LPPool2d_norm PASSED [ 42%] 2023-03-31T07:12:14.0015064Z test_nn.py::TestNN::test_LPPool2d_norm_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0015383Z test_nn.py::TestNN::test_LSTM_cell PASSED [ 43%] 2023-03-31T07:12:14.0015669Z test_nn.py::TestNN::test_LSTM_cell_forward_hidden_size PASSED [ 43%] 2023-03-31T07:12:14.0015982Z test_nn.py::TestNN::test_LSTM_cell_forward_input_size PASSED [ 43%] 2023-03-31T07:12:14.0016291Z test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine PASSED [ 43%] 2023-03-31T07:12:14.0016642Z test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0016978Z test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_eval PASSED [ 43%] 2023-03-31T07:12:14.0017375Z test_nn.py::TestNN::test_LayerNorm_1d_elementwise_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0017760Z test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine PASSED [ 43%] 2023-03-31T07:12:14.0018110Z test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0018479Z test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_eval PASSED [ 43%] 2023-03-31T07:12:14.0018856Z test_nn.py::TestNN::test_LayerNorm_1d_empty_elementwise_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0019215Z test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine PASSED [ 43%] 2023-03-31T07:12:14.0019558Z test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0019904Z test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_eval PASSED [ 43%] 2023-03-31T07:12:14.0020268Z test_nn.py::TestNN::test_LayerNorm_1d_no_elementwise_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0020624Z test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine PASSED [ 43%] 2023-03-31T07:12:14.0020964Z test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0021317Z test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_eval PASSED [ 43%] 2023-03-31T07:12:14.0021713Z test_nn.py::TestNN::test_LayerNorm_3d_elementwise_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 43%] 2023-03-31T07:12:14.0022055Z test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature PASSED [ 43%] 2023-03-31T07:12:14.0022412Z test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0022766Z test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_eval PASSED [ 44%] 2023-03-31T07:12:14.0023132Z test_nn.py::TestNN::test_LayerNorm_3d_no_affine_large_feature_eval_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0023575Z test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine PASSED [ 44%] 2023-03-31T07:12:14.0023937Z test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0024335Z test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_eval PASSED [ 44%] 2023-03-31T07:12:14.0024697Z test_nn.py::TestNN::test_LayerNorm_3d_no_elementwise_affine_eval_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0025030Z test_nn.py::TestNN::test_LeakyReLU_no_batch_dim PASSED [ 44%] 2023-03-31T07:12:14.0025362Z test_nn.py::TestNN::test_LeakyReLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0025680Z test_nn.py::TestNN::test_Linear PASSED [ 44%] 2023-03-31T07:12:14.0025971Z test_nn.py::TestNN::test_Linear_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0026276Z test_nn.py::TestNN::test_Linear_no_batch_dim PASSED [ 44%] 2023-03-31T07:12:14.0026599Z test_nn.py::TestNN::test_Linear_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0026917Z test_nn.py::TestNN::test_Linear_no_bias PASSED [ 44%] 2023-03-31T07:12:14.0027221Z test_nn.py::TestNN::test_Linear_no_bias_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0027549Z test_nn.py::TestNN::test_LocalResponseNorm_1d PASSED [ 44%] 2023-03-31T07:12:14.0027891Z test_nn.py::TestNN::test_LocalResponseNorm_1d_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0028220Z test_nn.py::TestNN::test_LocalResponseNorm_2d_uneven_pad PASSED [ 44%] 2023-03-31T07:12:14.0028574Z test_nn.py::TestNN::test_LocalResponseNorm_2d_uneven_pad_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0028927Z test_nn.py::TestNN::test_LocalResponseNorm_3d_custom_params PASSED [ 44%] 2023-03-31T07:12:14.0029329Z test_nn.py::TestNN::test_LocalResponseNorm_3d_custom_params_cuda SKIPPED (Excluded from CUDA tests) [ 44%] 2023-03-31T07:12:14.0029702Z test_nn.py::TestNN::test_LogSigmoid_no_batch_dim PASSED [ 45%] 2023-03-31T07:12:14.0030035Z test_nn.py::TestNN::test_LogSigmoid_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0030358Z test_nn.py::TestNN::test_MSELoss PASSED [ 45%] 2023-03-31T07:12:14.0030659Z test_nn.py::TestNN::test_MSELoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0030998Z test_nn.py::TestNN::test_MSELoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0031335Z test_nn.py::TestNN::test_MSELoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0031655Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean PASSED [ 45%] 2023-03-31T07:12:14.0031983Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0032363Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0032737Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0033078Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_none PASSED [ 45%] 2023-03-31T07:12:14.0033449Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0033828Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0034198Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0034518Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum PASSED [ 45%] 2023-03-31T07:12:14.0034856Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0035229Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0035602Z test_nn.py::TestNN::test_MSELoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0035995Z test_nn.py::TestNN::test_MSELoss_no_reduce PASSED [ 45%] 2023-03-31T07:12:14.0036320Z test_nn.py::TestNN::test_MSELoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 45%] 2023-03-31T07:12:14.0036642Z test_nn.py::TestNN::test_MSELoss_no_reduce_scalar PASSED [ 45%] 2023-03-31T07:12:14.0036961Z test_nn.py::TestNN::test_MSELoss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0037287Z test_nn.py::TestNN::test_MSELoss_prec PASSED [ 46%] 2023-03-31T07:12:14.0037610Z test_nn.py::TestNN::test_MSELoss_prec_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0051106Z test_nn.py::TestNN::test_MSELoss_prec_cuda_double SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0051493Z test_nn.py::TestNN::test_MSELoss_prec_cuda_float SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0051839Z test_nn.py::TestNN::test_MSELoss_prec_cuda_half SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0052172Z test_nn.py::TestNN::test_MSELoss_scalar PASSED [ 46%] 2023-03-31T07:12:14.0052511Z test_nn.py::TestNN::test_MSELoss_scalar_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0052874Z test_nn.py::TestNN::test_MSELoss_scalar_cuda_double SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0053220Z test_nn.py::TestNN::test_MSELoss_scalar_cuda_float SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0053569Z test_nn.py::TestNN::test_MSELoss_scalar_cuda_half SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0053899Z test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction PASSED [ 46%] 2023-03-31T07:12:14.0054403Z test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0054841Z test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0055234Z test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0055620Z test_nn.py::TestNN::test_MSELoss_scalar_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0055947Z test_nn.py::TestNN::test_MSELoss_sum_reduction PASSED [ 46%] 2023-03-31T07:12:14.0056286Z test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0056661Z test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0057031Z test_nn.py::TestNN::test_MSELoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 46%] 2023-03-31T07:12:14.0057365Z test_nn.py::TestNN::test_MarginRankingLoss PASSED [ 46%] 2023-03-31T07:12:14.0057716Z test_nn.py::TestNN::test_MarginRankingLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0058094Z test_nn.py::TestNN::test_MarginRankingLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0058513Z test_nn.py::TestNN::test_MarginRankingLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0058856Z test_nn.py::TestNN::test_MarginRankingLoss_margin PASSED [ 47%] 2023-03-31T07:12:14.0059216Z test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0059602Z test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_float SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0059979Z test_nn.py::TestNN::test_MarginRankingLoss_margin_cuda_half SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0060339Z test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction PASSED [ 47%] 2023-03-31T07:12:14.0060726Z test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0061133Z test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0061601Z test_nn.py::TestNN::test_MarginRankingLoss_margin_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0061971Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean PASSED [ 47%] 2023-03-31T07:12:14.0062345Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0062741Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0063141Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0063608Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none PASSED [ 47%] 2023-03-31T07:12:14.0063985Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0064384Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0064791Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0065155Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum PASSED [ 47%] 2023-03-31T07:12:14.0065525Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 47%] 2023-03-31T07:12:14.0065916Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0066351Z test_nn.py::TestNN::test_MarginRankingLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0066764Z test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction PASSED [ 48%] 2023-03-31T07:12:14.0066960Z test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0067154Z test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0067335Z test_nn.py::TestNN::test_MarginRankingLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0067469Z test_nn.py::TestNN::test_MaxPool1d PASSED [ 48%] 2023-03-31T07:12:14.0067625Z test_nn.py::TestNN::test_MaxPool1d_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0067770Z test_nn.py::TestNN::test_MaxPool1d_return_indices PASSED [ 48%] 2023-03-31T07:12:14.0067944Z test_nn.py::TestNN::test_MaxPool1d_return_indices_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0068087Z test_nn.py::TestNN::test_MaxPool1d_stride PASSED [ 48%] 2023-03-31T07:12:14.0068254Z test_nn.py::TestNN::test_MaxPool1d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0068396Z test_nn.py::TestNN::test_MaxPool2d_3d_input PASSED [ 48%] 2023-03-31T07:12:14.0068582Z test_nn.py::TestNN::test_MaxPool2d_3d_input_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0068721Z test_nn.py::TestNN::test_MaxPool2d_4d_input PASSED [ 48%] 2023-03-31T07:12:14.0068885Z test_nn.py::TestNN::test_MaxPool2d_4d_input_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0069030Z test_nn.py::TestNN::test_MaxPool2d_return_indices PASSED [ 48%] 2023-03-31T07:12:14.0069204Z test_nn.py::TestNN::test_MaxPool2d_return_indices_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0069338Z test_nn.py::TestNN::test_MaxPool3d PASSED [ 48%] 2023-03-31T07:12:14.0069495Z test_nn.py::TestNN::test_MaxPool3d_cuda SKIPPED (Excluded from CUDA tests) [ 48%] 2023-03-31T07:12:14.0069638Z test_nn.py::TestNN::test_MaxPool3d_return_indices PASSED [ 48%] 2023-03-31T07:12:14.0069835Z test_nn.py::TestNN::test_MaxPool3d_return_indices_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0069973Z test_nn.py::TestNN::test_MaxPool3d_stride PASSED [ 49%] 2023-03-31T07:12:14.0070136Z test_nn.py::TestNN::test_MaxPool3d_stride_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0070280Z test_nn.py::TestNN::test_MaxPool3d_stride_padding PASSED [ 49%] 2023-03-31T07:12:14.0070453Z test_nn.py::TestNN::test_MaxPool3d_stride_padding_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0070590Z test_nn.py::TestNN::test_MaxUnpool1d_net PASSED [ 49%] 2023-03-31T07:12:14.0070752Z test_nn.py::TestNN::test_MaxUnpool1d_net_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0070898Z test_nn.py::TestNN::test_MaxUnpool1d_net_no_batch_dim PASSED [ 49%] 2023-03-31T07:12:14.0071077Z test_nn.py::TestNN::test_MaxUnpool1d_net_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0071205Z test_nn.py::TestNN::test_MaxUnpool2d_net PASSED [ 49%] 2023-03-31T07:12:14.0071367Z test_nn.py::TestNN::test_MaxUnpool2d_net_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0071506Z test_nn.py::TestNN::test_MaxUnpool2d_net_no_batch_dim PASSED [ 49%] 2023-03-31T07:12:14.0071683Z test_nn.py::TestNN::test_MaxUnpool2d_net_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0071817Z test_nn.py::TestNN::test_MaxUnpool3d_net PASSED [ 49%] 2023-03-31T07:12:14.0071976Z test_nn.py::TestNN::test_MaxUnpool3d_net_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0072116Z test_nn.py::TestNN::test_MaxUnpool3d_net_no_batch_dim PASSED [ 49%] 2023-03-31T07:12:14.0072326Z test_nn.py::TestNN::test_MaxUnpool3d_net_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0072474Z test_nn.py::TestNN::test_Mish_no_batch_dim PASSED [ 49%] 2023-03-31T07:12:14.0072638Z test_nn.py::TestNN::test_Mish_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 49%] 2023-03-31T07:12:14.0072770Z test_nn.py::TestNN::test_ModuleDict PASSED [ 49%] 2023-03-31T07:12:14.0072903Z test_nn.py::TestNN::test_ModuleList PASSED [ 49%] 2023-03-31T07:12:14.0073050Z test_nn.py::TestNN::test_MultiLabelMarginLoss PASSED [ 50%] 2023-03-31T07:12:14.0073202Z test_nn.py::TestNN::test_MultiLabelMarginLoss_0d_no_reduce PASSED [ 50%] 2023-03-31T07:12:14.0073392Z test_nn.py::TestNN::test_MultiLabelMarginLoss_0d_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0073540Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d PASSED [ 50%] 2023-03-31T07:12:14.0073720Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0073906Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_double SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0074127Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_float SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0074311Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_cuda_half SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0074463Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_no_reduce PASSED [ 50%] 2023-03-31T07:12:14.0074652Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0074808Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction PASSED [ 50%] 2023-03-31T07:12:14.0075015Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0075220Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0075407Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0075641Z test_nn.py::TestNN::test_MultiLabelMarginLoss_1d_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0075827Z test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0076009Z test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0076191Z test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0076371Z test_nn.py::TestNN::test_MultiLabelMarginLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0076526Z test_nn.py::TestNN::test_MultiLabelMarginLoss_index_neg PASSED [ 50%] 2023-03-31T07:12:14.0076714Z test_nn.py::TestNN::test_MultiLabelMarginLoss_index_neg_cuda SKIPPED (Excluded from CUDA tests) [ 50%] 2023-03-31T07:12:14.0076873Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean PASSED [ 50%] 2023-03-31T07:12:14.0077066Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0077266Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0077465Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0077621Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none PASSED [ 51%] 2023-03-31T07:12:14.0077826Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0078053Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0078278Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0078436Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum PASSED [ 51%] 2023-03-31T07:12:14.0078636Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0078820Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0079015Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0079165Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_reduce PASSED [ 51%] 2023-03-31T07:12:14.0079353Z test_nn.py::TestNN::test_MultiLabelMarginLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0079511Z test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction PASSED [ 51%] 2023-03-31T07:12:14.0079710Z test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0079942Z test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0080146Z test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0080346Z test_nn.py::TestNN::test_MultiLabelMarginLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0080488Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss PASSED [ 51%] 2023-03-31T07:12:14.0080682Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0080875Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 51%] 2023-03-31T07:12:14.0081066Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0081264Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean PASSED [ 52%] 2023-03-31T07:12:14.0081479Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0081688Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0081895Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0082059Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none PASSED [ 52%] 2023-03-31T07:12:14.0082255Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0082465Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0082674Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0082840Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum PASSED [ 52%] 2023-03-31T07:12:14.0083216Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0083426Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0083634Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0083796Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_reduce PASSED [ 52%] 2023-03-31T07:12:14.0084040Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0084222Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights PASSED [ 52%] 2023-03-31T07:12:14.0084428Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0084633Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0084832Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0084999Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_no_reduce PASSED [ 52%] 2023-03-31T07:12:14.0085204Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 52%] 2023-03-31T07:12:14.0085376Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction PASSED [ 53%] 2023-03-31T07:12:14.0085595Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0085813Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0086016Z test_nn.py::TestNN::test_MultiLabelSoftMarginLoss_weights_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0086210Z test_nn.py::TestNN::test_MultiMarginLoss PASSED [ 53%] 2023-03-31T07:12:14.0086353Z test_nn.py::TestNN::test_MultiMarginLoss_1d PASSED [ 53%] 2023-03-31T07:12:14.0086531Z test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_double SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0086708Z test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_float SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0086879Z test_nn.py::TestNN::test_MultiMarginLoss_1d_cuda_half SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0087028Z test_nn.py::TestNN::test_MultiMarginLoss_1d_no_reduce PASSED [ 53%] 2023-03-31T07:12:14.0087212Z test_nn.py::TestNN::test_MultiMarginLoss_1d_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0087417Z test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction PASSED [ 53%] 2023-03-31T07:12:14.0087598Z test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0087788Z test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0087978Z test_nn.py::TestNN::test_MultiMarginLoss_1d_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0088151Z test_nn.py::TestNN::test_MultiMarginLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0088321Z test_nn.py::TestNN::test_MultiMarginLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0088492Z test_nn.py::TestNN::test_MultiMarginLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0088638Z test_nn.py::TestNN::test_MultiMarginLoss_margin PASSED [ 53%] 2023-03-31T07:12:14.0088822Z test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_double SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0089004Z test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_float SKIPPED (Excluded from CUDA tests) [ 53%] 2023-03-31T07:12:14.0089168Z test_nn.py::TestNN::test_MultiMarginLoss_margin_cuda_half SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0089318Z test_nn.py::TestNN::test_MultiMarginLoss_margin_no_reduce PASSED [ 54%] 2023-03-31T07:12:14.0089502Z test_nn.py::TestNN::test_MultiMarginLoss_margin_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0089653Z test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction PASSED [ 54%] 2023-03-31T07:12:14.0089884Z test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0090106Z test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0090303Z test_nn.py::TestNN::test_MultiMarginLoss_margin_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0090449Z test_nn.py::TestNN::test_MultiMarginLoss_no_reduce PASSED [ 54%] 2023-03-31T07:12:14.0090612Z test_nn.py::TestNN::test_MultiMarginLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0090755Z test_nn.py::TestNN::test_MultiMarginLoss_p PASSED [ 54%] 2023-03-31T07:12:14.0090932Z test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_double SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0091104Z test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_float SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0091276Z test_nn.py::TestNN::test_MultiMarginLoss_p_cuda_half SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0091424Z test_nn.py::TestNN::test_MultiMarginLoss_p_no_reduce PASSED [ 54%] 2023-03-31T07:12:14.0091600Z test_nn.py::TestNN::test_MultiMarginLoss_p_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0091756Z test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction PASSED [ 54%] 2023-03-31T07:12:14.0091986Z test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0092163Z test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0092352Z test_nn.py::TestNN::test_MultiMarginLoss_p_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0092500Z test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction PASSED [ 54%] 2023-03-31T07:12:14.0092692Z test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 54%] 2023-03-31T07:12:14.0092885Z test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0093072Z test_nn.py::TestNN::test_MultiMarginLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0093253Z test_nn.py::TestNN::test_MultiMarginLoss_weights PASSED [ 55%] 2023-03-31T07:12:14.0093439Z test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0093624Z test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0093790Z test_nn.py::TestNN::test_MultiMarginLoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0093941Z test_nn.py::TestNN::test_MultiMarginLoss_weights_no_reduce PASSED [ 55%] 2023-03-31T07:12:14.0094127Z test_nn.py::TestNN::test_MultiMarginLoss_weights_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0094280Z test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction PASSED [ 55%] 2023-03-31T07:12:14.0094480Z test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0094684Z test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0094880Z test_nn.py::TestNN::test_MultiMarginLoss_weights_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0095009Z test_nn.py::TestNN::test_NLLLoss PASSED [ 55%] 2023-03-31T07:12:14.0095148Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce PASSED [ 55%] 2023-03-31T07:12:14.0095299Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0095445Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce_ignore_index PASSED [ 55%] 2023-03-31T07:12:14.0095663Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce_ignore_index_cuda SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0095832Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce_weights PASSED [ 55%] 2023-03-31T07:12:14.0096012Z test_nn.py::TestNN::test_NLLLoss2d_no_reduce_weights_cuda SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0096150Z test_nn.py::TestNN::test_NLLLossNd_no_reduce PASSED [ 55%] 2023-03-31T07:12:14.0096316Z test_nn.py::TestNN::test_NLLLossNd_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 55%] 2023-03-31T07:12:14.0096457Z test_nn.py::TestNN::test_NLLLossNd_no_reduce_ignore_index PASSED [ 56%] 2023-03-31T07:12:14.0096627Z test_nn.py::TestNN::test_NLLLossNd_no_reduce_ignore_index_cuda SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0096767Z test_nn.py::TestNN::test_NLLLossNd_no_reduce_weights PASSED [ 56%] 2023-03-31T07:12:14.0096945Z test_nn.py::TestNN::test_NLLLossNd_no_reduce_weights_cuda SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0097079Z test_nn.py::TestNN::test_NLLLoss_2d PASSED [ 56%] 2023-03-31T07:12:14.0097244Z test_nn.py::TestNN::test_NLLLoss_2d_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0097409Z test_nn.py::TestNN::test_NLLLoss_2d_cuda_double SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0097605Z test_nn.py::TestNN::test_NLLLoss_2d_cuda_float SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0097766Z test_nn.py::TestNN::test_NLLLoss_2d_cuda_half SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0097891Z test_nn.py::TestNN::test_NLLLoss_2d_ignore_index PASSED [ 56%] 2023-03-31T07:12:14.0098072Z test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0098249Z test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0098431Z test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0098607Z test_nn.py::TestNN::test_NLLLoss_2d_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0098786Z test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction PASSED [ 56%] 2023-03-31T07:12:14.0098969Z test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0099151Z test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0099329Z test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0099494Z test_nn.py::TestNN::test_NLLLoss_2d_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0099629Z test_nn.py::TestNN::test_NLLLoss_2d_weights PASSED [ 56%] 2023-03-31T07:12:14.0099809Z test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 56%] 2023-03-31T07:12:14.0099985Z test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100160Z test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100331Z test_nn.py::TestNN::test_NLLLoss_2d_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100497Z test_nn.py::TestNN::test_NLLLoss_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100658Z test_nn.py::TestNN::test_NLLLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100815Z test_nn.py::TestNN::test_NLLLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0100960Z test_nn.py::TestNN::test_NLLLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0101134Z test_nn.py::TestNN::test_NLLLoss_dim_is_3 PASSED [ 57%] 2023-03-31T07:12:14.0101330Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0101497Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_double SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0101668Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_float SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0101834Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_cuda_half SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0101979Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction PASSED [ 57%] 2023-03-31T07:12:14.0102168Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0102341Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0102528Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0102712Z test_nn.py::TestNN::test_NLLLoss_dim_is_3_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0102850Z test_nn.py::TestNN::test_NLLLoss_higher_dim PASSED [ 57%] 2023-03-31T07:12:14.0103028Z test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0103247Z test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_double SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0103535Z test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_float SKIPPED (Excluded from CUDA tests) [ 57%] 2023-03-31T07:12:14.0103707Z test_nn.py::TestNN::test_NLLLoss_higher_dim_cuda_half SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0103857Z test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction PASSED [ 58%] 2023-03-31T07:12:14.0104036Z test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0104230Z test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0104421Z test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0104653Z test_nn.py::TestNN::test_NLLLoss_higher_dim_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0104792Z test_nn.py::TestNN::test_NLLLoss_ignore_index PASSED [ 58%] 2023-03-31T07:12:14.0104970Z test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0105148Z test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0105322Z test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0105498Z test_nn.py::TestNN::test_NLLLoss_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0105628Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean PASSED [ 58%] 2023-03-31T07:12:14.0105810Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0105987Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0106166Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0106311Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none PASSED [ 58%] 2023-03-31T07:12:14.0106492Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0106669Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0106867Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0107037Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum PASSED [ 58%] 2023-03-31T07:12:14.0107202Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 58%] 2023-03-31T07:12:14.0107381Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0107559Z test_nn.py::TestNN::test_NLLLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0107695Z test_nn.py::TestNN::test_NLLLoss_no_reduce PASSED [ 59%] 2023-03-31T07:12:14.0107859Z test_nn.py::TestNN::test_NLLLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0108002Z test_nn.py::TestNN::test_NLLLoss_no_reduce_ignore_index PASSED [ 59%] 2023-03-31T07:12:14.0108180Z test_nn.py::TestNN::test_NLLLoss_no_reduce_ignore_index_cuda SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0108321Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights PASSED [ 59%] 2023-03-31T07:12:14.0108483Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_cuda SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0108636Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index PASSED [ 59%] 2023-03-31T07:12:14.0108852Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_cuda SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0109009Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_neg PASSED [ 59%] 2023-03-31T07:12:14.0109202Z test_nn.py::TestNN::test_NLLLoss_no_reduce_weights_ignore_index_neg_cuda SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0109340Z test_nn.py::TestNN::test_NLLLoss_sum_reduction PASSED [ 59%] 2023-03-31T07:12:14.0109518Z test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0109701Z test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0109878Z test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0110067Z test_nn.py::TestNN::test_NLLLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0110206Z test_nn.py::TestNN::test_NLLLoss_weights PASSED [ 59%] 2023-03-31T07:12:14.0110379Z test_nn.py::TestNN::test_NLLLoss_weights_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0110550Z test_nn.py::TestNN::test_NLLLoss_weights_cuda_double SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0110719Z test_nn.py::TestNN::test_NLLLoss_weights_cuda_float SKIPPED (Excluded from CUDA tests) [ 59%] 2023-03-31T07:12:14.0110887Z test_nn.py::TestNN::test_NLLLoss_weights_cuda_half SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0111030Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index PASSED [ 60%] 2023-03-31T07:12:14.0111220Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0111394Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_double SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0111584Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_float SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0111765Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_cuda_half SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0111913Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg PASSED [ 60%] 2023-03-31T07:12:14.0112104Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_bfloat16 SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0112295Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_double SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0112513Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_float SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0112720Z test_nn.py::TestNN::test_NLLLoss_weights_ignore_index_neg_cuda_half SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0112873Z test_nn.py::TestNN::test_PReLU_backward_requires_grad_false PASSED [ 60%] 2023-03-31T07:12:14.0112999Z test_nn.py::TestNN::test_PReLU_no_batch_dim PASSED [ 60%] 2023-03-31T07:12:14.0113162Z test_nn.py::TestNN::test_PReLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0113304Z test_nn.py::TestNN::test_Padding122112_3dcircular PASSED [ 60%] 2023-03-31T07:12:14.0113478Z test_nn.py::TestNN::test_Padding122112_3dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0113620Z test_nn.py::TestNN::test_Padding1221_2dcircular PASSED [ 60%] 2023-03-31T07:12:14.0113788Z test_nn.py::TestNN::test_Padding1221_2dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0113929Z test_nn.py::TestNN::test_Padding12_1dcircular PASSED [ 60%] 2023-03-31T07:12:14.0114097Z test_nn.py::TestNN::test_Padding12_1dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 60%] 2023-03-31T07:12:14.0114239Z test_nn.py::TestNN::test_Padding2322_2dcircular PASSED [ 60%] 2023-03-31T07:12:14.0114424Z test_nn.py::TestNN::test_Padding2322_2dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0114565Z test_nn.py::TestNN::test_Padding31_1dcircular PASSED [ 61%] 2023-03-31T07:12:14.0114732Z test_nn.py::TestNN::test_Padding31_1dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0114873Z test_nn.py::TestNN::test_Padding322112_3dcircular PASSED [ 61%] 2023-03-31T07:12:14.0115046Z test_nn.py::TestNN::test_Padding322112_3dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0115185Z test_nn.py::TestNN::test_Padding332122_3dcircular PASSED [ 61%] 2023-03-31T07:12:14.0115358Z test_nn.py::TestNN::test_Padding332122_3dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0115499Z test_nn.py::TestNN::test_Padding3331_2dcircular PASSED [ 61%] 2023-03-31T07:12:14.0115681Z test_nn.py::TestNN::test_Padding3331_2dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0115822Z test_nn.py::TestNN::test_Padding33_1dcircular PASSED [ 61%] 2023-03-31T07:12:14.0115991Z test_nn.py::TestNN::test_Padding33_1dcircular_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0116131Z test_nn.py::TestNN::test_PairwiseDistance PASSED [ 61%] 2023-03-31T07:12:14.0116281Z test_nn.py::TestNN::test_PairwiseDistance_broadcast_lhs PASSED [ 61%] 2023-03-31T07:12:14.0116468Z test_nn.py::TestNN::test_PairwiseDistance_broadcast_lhs_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0116619Z test_nn.py::TestNN::test_PairwiseDistance_broadcast_rhs PASSED [ 61%] 2023-03-31T07:12:14.0116804Z test_nn.py::TestNN::test_PairwiseDistance_broadcast_rhs_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0116958Z test_nn.py::TestNN::test_PairwiseDistance_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0117107Z test_nn.py::TestNN::test_PairwiseDistance_no_batch_dim PASSED [ 61%] 2023-03-31T07:12:14.0117292Z test_nn.py::TestNN::test_PairwiseDistance_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0117447Z test_nn.py::TestNN::test_PairwiseDistance_with_non_default_args PASSED [ 61%] 2023-03-31T07:12:14.0117641Z test_nn.py::TestNN::test_PairwiseDistance_with_non_default_args_cuda SKIPPED (Excluded from CUDA tests) [ 61%] 2023-03-31T07:12:14.0117774Z test_nn.py::TestNN::test_ParameterDict PASSED [ 62%] 2023-03-31T07:12:14.0117919Z test_nn.py::TestNN::test_ParameterDict_replication PASSED [ 62%] 2023-03-31T07:12:14.0118088Z test_nn.py::TestNN::test_ParameterList PASSED [ 62%] 2023-03-31T07:12:14.0118245Z test_nn.py::TestNN::test_ParameterList_meta PASSED [ 62%] 2023-03-31T07:12:14.0118375Z test_nn.py::TestNN::test_ParameterList_replication PASSED [ 62%] 2023-03-31T07:12:14.0118512Z test_nn.py::TestNN::test_PixelShuffle PASSED [ 62%] 2023-03-31T07:12:14.0118675Z test_nn.py::TestNN::test_PixelShuffle_cuda SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0118812Z test_nn.py::TestNN::test_PixelUnshuffle PASSED [ 62%] 2023-03-31T07:12:14.0118975Z test_nn.py::TestNN::test_PixelUnshuffle_cuda SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0119116Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss PASSED [ 62%] 2023-03-31T07:12:14.0119303Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_double SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0119487Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_float SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0119657Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_cuda_half SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0119809Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input PASSED [ 62%] 2023-03-31T07:12:14.0120036Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_double SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0120233Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_float SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0120428Z test_nn.py::TestNN::test_PoissonNLLLoss_full_loss_no_log_input_cuda_half SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0120577Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean PASSED [ 62%] 2023-03-31T07:12:14.0120768Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0120960Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0121150Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 62%] 2023-03-31T07:12:14.0121313Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none PASSED [ 62%] 2023-03-31T07:12:14.0121509Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0121699Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0121885Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0122031Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum PASSED [ 63%] 2023-03-31T07:12:14.0122218Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0122407Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0122598Z test_nn.py::TestNN::test_PoissonNLLLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0122744Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss PASSED [ 63%] 2023-03-31T07:12:14.0122917Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_double SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0123243Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_float SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0123427Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_cuda_half SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0123580Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input PASSED [ 63%] 2023-03-31T07:12:14.0123836Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_double SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0124064Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_float SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0124261Z test_nn.py::TestNN::test_PoissonNLLLoss_no_full_loss_no_log_input_cuda_half SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0124407Z test_nn.py::TestNN::test_PoissonNLLLoss_no_reduce PASSED [ 63%] 2023-03-31T07:12:14.0124583Z test_nn.py::TestNN::test_PoissonNLLLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 63%] 2023-03-31T07:12:14.0124700Z test_nn.py::TestNN::test_RNN_cell PASSED [ 63%] 2023-03-31T07:12:14.0124845Z test_nn.py::TestNN::test_RNN_cell_forward_hidden_size PASSED [ 63%] 2023-03-31T07:12:14.0124988Z test_nn.py::TestNN::test_RNN_cell_forward_input_size PASSED [ 63%] 2023-03-31T07:12:14.0125132Z test_nn.py::TestNN::test_RNN_cell_forward_zero_hidden_size PASSED [ 63%] 2023-03-31T07:12:14.0125277Z test_nn.py::TestNN::test_RNN_cell_no_broadcasting PASSED [ 64%] 2023-03-31T07:12:14.0125423Z test_nn.py::TestNN::test_RNN_change_dropout SKIPPED (needs cudnn >= 5.1) [ 64%] 2023-03-31T07:12:14.0125579Z test_nn.py::TestNN::test_RNN_cpu_vs_cudnn_no_dropout SKIPPED (needs cudnn) [ 64%] 2023-03-31T07:12:14.0125785Z test_nn.py::TestNN::test_RNN_cpu_vs_cudnn_with_dropout SKIPPED (needs cudnn >= 5.1) [ 64%] 2023-03-31T07:12:14.0125920Z test_nn.py::TestNN::test_RNN_cudnn_weight_norm SKIPPED (needs cudnn) [ 64%] 2023-03-31T07:12:14.0126061Z test_nn.py::TestNN::test_RNN_dropout SKIPPED (needs cudnn >= 5.1) [ 64%] 2023-03-31T07:12:14.0126205Z test_nn.py::TestNN::test_RNN_dropout_state SKIPPED (needs cudnn >= 5.1) [ 64%] 2023-03-31T07:12:14.0126339Z test_nn.py::TestNN::test_RNN_input_size_zero PASSED [ 64%] 2023-03-31T07:12:14.0126473Z test_nn.py::TestNN::test_RNN_nonlinearity PASSED [ 64%] 2023-03-31T07:12:14.0126601Z test_nn.py::TestNN::test_RReLU PASSED [ 64%] 2023-03-31T07:12:14.0126748Z test_nn.py::TestNN::test_RReLU_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0126878Z test_nn.py::TestNN::test_RReLU_no_batch_dim PASSED [ 64%] 2023-03-31T07:12:14.0127078Z test_nn.py::TestNN::test_RReLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0127215Z test_nn.py::TestNN::test_RReLU_with_up_down PASSED [ 64%] 2023-03-31T07:12:14.0127376Z test_nn.py::TestNN::test_RReLU_with_up_down_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0127514Z test_nn.py::TestNN::test_RReLU_with_up_down_scalar PASSED [ 64%] 2023-03-31T07:12:14.0127686Z test_nn.py::TestNN::test_RReLU_with_up_down_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0127819Z test_nn.py::TestNN::test_ReLU6_no_batch_dim PASSED [ 64%] 2023-03-31T07:12:14.0127983Z test_nn.py::TestNN::test_ReLU6_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0128119Z test_nn.py::TestNN::test_ReLU_no_batch_dim PASSED [ 64%] 2023-03-31T07:12:14.0128283Z test_nn.py::TestNN::test_ReLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 64%] 2023-03-31T07:12:14.0128410Z test_nn.py::TestNN::test_ReplicationPad3d PASSED [ 65%] 2023-03-31T07:12:14.0128558Z test_nn.py::TestNN::test_ReplicationPad3d_complex PASSED [ 65%] 2023-03-31T07:12:14.0128737Z test_nn.py::TestNN::test_ReplicationPad3d_complex_cuda SKIPPED (Excluded from CUDA tests) [ 65%] 2023-03-31T07:12:14.0128904Z test_nn.py::TestNN::test_ReplicationPad3d_cuda SKIPPED (Excluded from CUDA tests) [ 65%] 2023-03-31T07:12:14.0129050Z test_nn.py::TestNN::test_ReplicationPad3d_no_batch_dim PASSED [ 65%] 2023-03-31T07:12:14.0129233Z test_nn.py::TestNN::test_ReplicationPad3d_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 65%] 2023-03-31T07:12:14.0129399Z test_nn.py::TestNN::test_SELU_no_batch_dim PASSED [ 65%] 2023-03-31T07:12:14.0129581Z test_nn.py::TestNN::test_SELU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 65%] 2023-03-31T07:12:14.0129704Z test_nn.py::TestNN::test_Sequential_add PASSED [ 65%] 2023-03-31T07:12:14.0129845Z test_nn.py::TestNN::test_Sequential_append PASSED [ 65%] 2023-03-31T07:12:14.0129980Z test_nn.py::TestNN::test_Sequential_delitem PASSED [ 65%] 2023-03-31T07:12:14.0130113Z test_nn.py::TestNN::test_Sequential_extend PASSED [ 65%] 2023-03-31T07:12:14.0130251Z test_nn.py::TestNN::test_Sequential_getitem PASSED [ 65%] 2023-03-31T07:12:14.0130383Z test_nn.py::TestNN::test_Sequential_iadd PASSED [ 65%] 2023-03-31T07:12:14.0130516Z test_nn.py::TestNN::test_Sequential_imul PASSED [ 65%] 2023-03-31T07:12:14.0130652Z test_nn.py::TestNN::test_Sequential_insert PASSED [ 65%] 2023-03-31T07:12:14.0130781Z test_nn.py::TestNN::test_Sequential_insert_fail_case PASSED [ 65%] 2023-03-31T07:12:14.0130911Z test_nn.py::TestNN::test_Sequential_mul PASSED [ 65%] 2023-03-31T07:12:14.0131044Z test_nn.py::TestNN::test_Sequential_pop PASSED [ 65%] 2023-03-31T07:12:14.0131218Z test_nn.py::TestNN::test_Sequential_rmul PASSED [ 65%] 2023-03-31T07:12:14.0131355Z test_nn.py::TestNN::test_Sequential_setitem PASSED [ 65%] 2023-03-31T07:12:14.0131490Z test_nn.py::TestNN::test_Sequential_setitem_named PASSED [ 66%] 2023-03-31T07:12:14.0131619Z test_nn.py::TestNN::test_SiLU_no_batch_dim PASSED [ 66%] 2023-03-31T07:12:14.0131782Z test_nn.py::TestNN::test_SiLU_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0131903Z test_nn.py::TestNN::test_Sigmoid_no_batch_dim PASSED [ 66%] 2023-03-31T07:12:14.0132076Z test_nn.py::TestNN::test_Sigmoid_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0132208Z test_nn.py::TestNN::test_SmoothL1Loss PASSED [ 66%] 2023-03-31T07:12:14.0132375Z test_nn.py::TestNN::test_SmoothL1Loss_beta PASSED [ 66%] 2023-03-31T07:12:14.0132539Z test_nn.py::TestNN::test_SmoothL1Loss_beta_cuda SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0132706Z test_nn.py::TestNN::test_SmoothL1Loss_cuda_double SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0132873Z test_nn.py::TestNN::test_SmoothL1Loss_cuda_float SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0133035Z test_nn.py::TestNN::test_SmoothL1Loss_cuda_half SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0133179Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean PASSED [ 66%] 2023-03-31T07:12:14.0133355Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0133544Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0133728Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0133874Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none PASSED [ 66%] 2023-03-31T07:12:14.0134057Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0134239Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0134419Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0134560Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum PASSED [ 66%] 2023-03-31T07:12:14.0134765Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 66%] 2023-03-31T07:12:14.0134971Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0135153Z test_nn.py::TestNN::test_SmoothL1Loss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0135295Z test_nn.py::TestNN::test_SmoothL1Loss_no_reduce PASSED [ 67%] 2023-03-31T07:12:14.0135466Z test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0135611Z test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_scalar PASSED [ 67%] 2023-03-31T07:12:14.0135789Z test_nn.py::TestNN::test_SmoothL1Loss_no_reduce_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0135928Z test_nn.py::TestNN::test_SmoothL1Loss_scalar PASSED [ 67%] 2023-03-31T07:12:14.0136109Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_double SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0136273Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_float SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0136445Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_cuda_half SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0136593Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction PASSED [ 67%] 2023-03-31T07:12:14.0136824Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0137017Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0137208Z test_nn.py::TestNN::test_SmoothL1Loss_scalar_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0137350Z test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction PASSED [ 67%] 2023-03-31T07:12:14.0137535Z test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0137719Z test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0137921Z test_nn.py::TestNN::test_SmoothL1Loss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0138065Z test_nn.py::TestNN::test_SmoothL1Loss_zero_beta PASSED [ 67%] 2023-03-31T07:12:14.0138232Z test_nn.py::TestNN::test_SmoothL1Loss_zero_beta_cuda SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:12:14.0138369Z test_nn.py::TestNN::test_SoftMarginLoss PASSED [ 67%] 2023-03-31T07:12:14.0138541Z test_nn.py::TestNN::test_SoftMarginLoss_cuda_double SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0138710Z test_nn.py::TestNN::test_SoftMarginLoss_cuda_float SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0138880Z test_nn.py::TestNN::test_SoftMarginLoss_cuda_half SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0139031Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean PASSED [ 68%] 2023-03-31T07:12:14.0139212Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0139404Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0139595Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0139743Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none PASSED [ 68%] 2023-03-31T07:12:14.0139931Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0140117Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0140332Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0140507Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum PASSED [ 68%] 2023-03-31T07:12:14.0140692Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0140869Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0141060Z test_nn.py::TestNN::test_SoftMarginLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0141203Z test_nn.py::TestNN::test_SoftMarginLoss_no_reduce PASSED [ 68%] 2023-03-31T07:12:14.0141377Z test_nn.py::TestNN::test_SoftMarginLoss_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0141523Z test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction PASSED [ 68%] 2023-03-31T07:12:14.0141711Z test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_double SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0141899Z test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_float SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0142085Z test_nn.py::TestNN::test_SoftMarginLoss_sum_reduction_cuda_half SKIPPED (Excluded from CUDA tests) [ 68%] 2023-03-31T07:12:14.0142227Z test_nn.py::TestNN::test_Softplus_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0142417Z test_nn.py::TestNN::test_Softplus_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0142558Z test_nn.py::TestNN::test_Softshrink_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0142728Z test_nn.py::TestNN::test_Softshrink_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0142867Z test_nn.py::TestNN::test_Softsign_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0143033Z test_nn.py::TestNN::test_Softsign_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0143169Z test_nn.py::TestNN::test_Tanh_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0143433Z test_nn.py::TestNN::test_Tanh_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0143622Z test_nn.py::TestNN::test_Tanhshrink_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0143782Z test_nn.py::TestNN::test_Tanhshrink_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0143922Z test_nn.py::TestNN::test_Threshold_no_batch_dim PASSED [ 69%] 2023-03-31T07:12:14.0144090Z test_nn.py::TestNN::test_Threshold_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0144255Z test_nn.py::TestNN::test_TransformerDecoderLayer_gelu_activation PASSED [ 69%] 2023-03-31T07:12:14.0144457Z test_nn.py::TestNN::test_TransformerDecoderLayer_gelu_activation_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0144618Z test_nn.py::TestNN::test_TransformerDecoderLayer_relu_activation PASSED [ 69%] 2023-03-31T07:12:14.0144821Z test_nn.py::TestNN::test_TransformerDecoderLayer_relu_activation_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0144985Z test_nn.py::TestNN::test_TransformerEncoderLayer_gelu_activation PASSED [ 69%] 2023-03-31T07:12:14.0145184Z test_nn.py::TestNN::test_TransformerEncoderLayer_gelu_activation_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0145335Z test_nn.py::TestNN::test_TransformerEncoderLayer_relu_activation PASSED [ 69%] 2023-03-31T07:12:14.0145533Z test_nn.py::TestNN::test_TransformerEncoderLayer_relu_activation_cuda SKIPPED (Excluded from CUDA tests) [ 69%] 2023-03-31T07:12:14.0145669Z test_nn.py::TestNN::test_Transformer_cell PASSED [ 69%] 2023-03-31T07:12:14.0145815Z test_nn.py::TestNN::test_Transformer_multilayer_coder PASSED [ 70%] 2023-03-31T07:12:14.0145998Z test_nn.py::TestNN::test_Transformer_multilayer_coder_cuda SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0146206Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean PASSED [ 70%] 2023-03-31T07:12:14.0146474Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_double SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0146674Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_float SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0146873Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_mean_cuda_half SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0147015Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none PASSED [ 70%] 2023-03-31T07:12:14.0147210Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_double SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0147402Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_float SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0147595Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_none_cuda_half SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0147749Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum PASSED [ 70%] 2023-03-31T07:12:14.0147944Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_double SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0148135Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_float SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0148364Z test_nn.py::TestNN::test_TripletMarginLoss_no_batch_dim_sum_cuda_half SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0148508Z test_nn.py::TestNN::test_Unflatten_no_batch_dim PASSED [ 70%] 2023-03-31T07:12:14.0148664Z test_nn.py::TestNN::test_Unflatten_no_batch_dim_cuda SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0148797Z test_nn.py::TestNN::test_Unfold PASSED [ 70%] 2023-03-31T07:12:14.0148945Z test_nn.py::TestNN::test_Unfold_cuda SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0149085Z test_nn.py::TestNN::test_Unfold_int_input PASSED [ 70%] 2023-03-31T07:12:14.0149254Z test_nn.py::TestNN::test_Unfold_int_input_cuda SKIPPED (Excluded from CUDA tests) [ 70%] 2023-03-31T07:12:14.0149426Z test_nn.py::TestNN::test_adaptive_log_softmax PASSED [ 70%] 2023-03-31T07:12:14.0149561Z test_nn.py::TestNN::test_add_module PASSED [ 71%] 2023-03-31T07:12:14.0149712Z test_nn.py::TestNN::test_add_module_raises_error_if_attr_exists PASSED [ 71%] 2023-03-31T07:12:14.0149827Z test_nn.py::TestNN::test_affine_grid PASSED [ 71%] 2023-03-31T07:12:14.0149957Z test_nn.py::TestNN::test_affine_grid_3d PASSED [ 71%] 2023-03-31T07:12:14.0150097Z test_nn.py::TestNN::test_affine_grid_error_checking PASSED [ 71%] 2023-03-31T07:12:14.0150228Z test_nn.py::TestNN::test_assignment PASSED [ 71%] 2023-03-31T07:12:14.0150396Z test_nn.py::TestNN::test_batchnorm_buffer_update_when_stats_are_not_tracked PASSED [ 71%] 2023-03-31T07:12:14.0150547Z test_nn.py::TestNN::test_batchnorm_cudnn_half SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:12:14.0150695Z test_nn.py::TestNN::test_batchnorm_cudnn_nhwc SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:12:14.0150835Z test_nn.py::TestNN::test_batchnorm_nhwc_cpu PASSED [ 71%] 2023-03-31T07:12:14.0150977Z test_nn.py::TestNN::test_batchnorm_nhwc_cuda SKIPPED (CUDA not available) [ 71%] 2023-03-31T07:12:14.0151125Z test_nn.py::TestNN::test_batchnorm_non_contig_cpu_BatchNorm2d PASSED [ 71%] 2023-03-31T07:12:14.0151275Z test_nn.py::TestNN::test_batchnorm_non_contig_cpu_SyncBatchNorm PASSED [ 71%] 2023-03-31T07:12:14.0151449Z test_nn.py::TestNN::test_batchnorm_nonaffine_cuda_half_input SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:12:14.0151619Z test_nn.py::TestNN::test_batchnorm_raises_error_if_bias_is_not_same_size_as_input PASSED [ 71%] 2023-03-31T07:12:14.0151827Z test_nn.py::TestNN::test_batchnorm_raises_error_if_less_than_one_value_per_channel PASSED [ 71%] 2023-03-31T07:12:14.0152032Z test_nn.py::TestNN::test_batchnorm_raises_error_if_running_mean_is_not_same_size_as_input PASSED [ 71%] 2023-03-31T07:12:14.0152214Z test_nn.py::TestNN::test_batchnorm_raises_error_if_running_var_is_not_same_size_as_input PASSED [ 71%] 2023-03-31T07:12:14.0152393Z test_nn.py::TestNN::test_batchnorm_raises_error_if_running_var_or_running_mean_have_forward_grad PASSED [ 71%] 2023-03-31T07:12:14.0152568Z test_nn.py::TestNN::test_batchnorm_raises_error_if_weight_is_not_same_size_as_input PASSED [ 71%] 2023-03-31T07:12:14.0152710Z test_nn.py::TestNN::test_bce_loss_always_nonnegative PASSED [ 71%] 2023-03-31T07:12:14.0152853Z test_nn.py::TestNN::test_bce_loss_broadcasts_weights PASSED [ 72%] 2023-03-31T07:12:14.0152988Z test_nn.py::TestNN::test_bce_loss_input_range PASSED [ 72%] 2023-03-31T07:12:14.0153125Z test_nn.py::TestNN::test_bce_loss_size_mismatch PASSED [ 72%] 2023-03-31T07:12:14.0153276Z test_nn.py::TestNN::test_bce_with_logits_broadcasts_pos_weights PASSED [ 72%] 2023-03-31T07:12:14.0153423Z test_nn.py::TestNN::test_bce_with_logits_broadcasts_weights PASSED [ 72%] 2023-03-31T07:12:14.0153597Z test_nn.py::TestNN::test_bce_with_logits_gives_same_result_as_sigmoid_and_bce_loss PASSED [ 72%] 2023-03-31T07:12:14.0153814Z test_nn.py::TestNN::test_bce_with_logits_gives_same_result_as_sigmoid_and_bce_loss_large_tensors_with_grad PASSED [ 72%] 2023-03-31T07:12:14.0153966Z test_nn.py::TestNN::test_bce_with_logits_has_correct_forward_grad PASSED [ 72%] 2023-03-31T07:12:14.0154116Z test_nn.py::TestNN::test_bce_with_logits_has_correct_grad_at_zero PASSED [ 72%] 2023-03-31T07:12:14.0154284Z test_nn.py::TestNN::test_bce_with_logits_ones_in_pos_weights_are_the_same_as_none PASSED [ 72%] 2023-03-31T07:12:14.0154463Z test_nn.py::TestNN::test_bce_with_logits_raises_if_target_and_input_are_different_size PASSED [ 72%] 2023-03-31T07:12:14.0154607Z test_nn.py::TestNN::test_bce_with_logits_stability PASSED [ 72%] 2023-03-31T07:12:14.0154778Z test_nn.py::TestNN::test_bce_with_logits_with_pos_weight_has_correct_grad_at_zero PASSED [ 72%] 2023-03-31T07:12:14.0154944Z test_nn.py::TestNN::test_bilinear PASSED [ 72%] 2023-03-31T07:12:14.0155071Z test_nn.py::TestNN::test_bilinear_broadcasting PASSED [ 72%] 2023-03-31T07:12:14.0155206Z test_nn.py::TestNN::test_bilinear_no_bias PASSED [ 72%] 2023-03-31T07:12:14.0155344Z test_nn.py::TestNN::test_bilinear_non_contiguous PASSED [ 72%] 2023-03-31T07:12:14.0155733Z test_nn.py::TestNN::test_broadcast_double_backwards_gpu SKIPPED (multi-GPU not supported) [ 72%] 2023-03-31T07:12:14.0155969Z test_nn.py::TestNN::test_broadcast_no_grad SKIPPED (multi-GPU not supported) [ 72%] 2023-03-31T07:12:14.0156220Z test_nn.py::TestNN::test_broadcast_not_requiring_grad SKIPPED (multi-GPU not supported) [ 72%] 2023-03-31T07:12:14.0156363Z test_nn.py::TestNN::test_buffer_not_persistent PASSED [ 72%] 2023-03-31T07:12:14.0156510Z test_nn.py::TestNN::test_buffer_not_persistent_assign PASSED [ 73%] 2023-03-31T07:12:14.0156639Z test_nn.py::TestNN::test_buffer_not_persistent_del PASSED [ 73%] 2023-03-31T07:12:14.0156783Z test_nn.py::TestNN::test_buffer_not_persistent_load PASSED [ 73%] 2023-03-31T07:12:14.0156928Z test_nn.py::TestNN::test_buffer_not_persistent_overwrite PASSED [ 73%] 2023-03-31T07:12:14.0157071Z test_nn.py::TestNN::test_buffers_and_named_buffers PASSED [ 73%] 2023-03-31T07:12:14.0157217Z test_nn.py::TestNN::test_call_supports_python_dict_output PASSED [ 73%] 2023-03-31T07:12:14.0157348Z test_nn.py::TestNN::test_channel_shuffle PASSED [ 73%] 2023-03-31T07:12:14.0157488Z test_nn.py::TestNN::test_channel_shuffle_return_self PASSED [ 73%] 2023-03-31T07:12:14.0157652Z test_nn.py::TestNN::test_children PASSED [ 73%] 2023-03-31T07:12:14.0157810Z test_nn.py::TestNN::test_container_copy PASSED [ 73%] 2023-03-31T07:12:14.0157954Z test_nn.py::TestNN::test_convert_sync_batchnorm SKIPPED (CUDA not available) [ 73%] 2023-03-31T07:12:14.0158101Z test_nn.py::TestNN::test_cosine_embedding_loss_invalid_shape PASSED [ 73%] 2023-03-31T07:12:14.0158250Z test_nn.py::TestNN::test_cosine_embedding_loss_margin_no_reduce PASSED [ 73%] 2023-03-31T07:12:14.0158392Z test_nn.py::TestNN::test_cosine_embedding_loss_no_reduce PASSED [ 73%] 2023-03-31T07:12:14.0158537Z test_nn.py::TestNN::test_cosine_embedding_loss_with_diff_type PASSED [ 73%] 2023-03-31T07:12:14.0158669Z test_nn.py::TestNN::test_cosine_similarity PASSED [ 73%] 2023-03-31T07:12:14.0158802Z test_nn.py::TestNN::test_cross_entropy_loss PASSED [ 73%] 2023-03-31T07:12:14.0158943Z test_nn.py::TestNN::test_cross_entropy_loss_precision PASSED [ 73%] 2023-03-31T07:12:14.0159068Z test_nn.py::TestNN::test_cross_entropy_loss_zero_div PASSED [ 73%] 2023-03-31T07:12:14.0159347Z test_nn.py::TestNN::test_cudnn_rnn_dropout_states_device SKIPPED (CUDNN or multi-gpu not available) [ 73%] 2023-03-31T07:12:14.0159508Z test_nn.py::TestNN::test_cudnn_weight_format SKIPPED (CUDNN not available) [ 73%] 2023-03-31T07:12:14.0159702Z test_nn.py::TestNN::test_cudnn_weight_tying SKIPPED (CUDNN not available) [ 74%] 2023-03-31T07:12:14.0159829Z test_nn.py::TestNN::test_dir PASSED [ 74%] 2023-03-31T07:12:14.0159957Z test_nn.py::TestNN::test_dir_digit PASSED [ 74%] 2023-03-31T07:12:14.0160094Z test_nn.py::TestNN::test_elu_inplace_gradgrad PASSED [ 74%] 2023-03-31T07:12:14.0160230Z test_nn.py::TestNN::test_elu_inplace_on_view PASSED [ 74%] 2023-03-31T07:12:14.0160353Z test_nn.py::TestNN::test_error_RNN_seq_len_zero PASSED [ 74%] 2023-03-31T07:12:14.0160484Z test_nn.py::TestNN::test_extra_state PASSED [ 74%] 2023-03-31T07:12:14.0160630Z test_nn.py::TestNN::test_extra_state_missing_get_extra_state PASSED [ 74%] 2023-03-31T07:12:14.0160801Z test_nn.py::TestNN::test_extra_state_missing_set_extra_state PASSED [ 74%] 2023-03-31T07:12:14.0160933Z test_nn.py::TestNN::test_extra_state_non_dict PASSED [ 74%] 2023-03-31T07:12:14.0161065Z test_nn.py::TestNN::test_fb_fc_packed PASSED [ 74%] 2023-03-31T07:12:14.0161191Z test_nn.py::TestNN::test_flatten PASSED [ 74%] 2023-03-31T07:12:14.0161322Z test_nn.py::TestNN::test_fold_invalid_arg PASSED [ 74%] 2023-03-31T07:12:14.0161446Z test_nn.py::TestNN::test_gaussian_nll_loss_args PASSED [ 74%] 2023-03-31T07:12:14.0161589Z test_nn.py::TestNN::test_gaussian_nll_loss_broadcasting PASSED [ 74%] 2023-03-31T07:12:14.0161716Z test_nn.py::TestNN::test_get_buffer PASSED [ 74%] 2023-03-31T07:12:14.0161857Z test_nn.py::TestNN::test_get_buffer_from_submodules PASSED [ 74%] 2023-03-31T07:12:14.0161993Z test_nn.py::TestNN::test_getattr_with_property XFAIL [ 74%] 2023-03-31T07:12:14.0162124Z test_nn.py::TestNN::test_grid_sample PASSED [ 74%] 2023-03-31T07:12:14.0162255Z test_nn.py::TestNN::test_grid_sample_3d PASSED [ 74%] 2023-03-31T07:12:14.0162393Z test_nn.py::TestNN::test_grid_sample_error_checking PASSED [ 74%] 2023-03-31T07:12:14.0162514Z test_nn.py::TestNN::test_hardtanh_backward PASSED [ 75%] 2023-03-31T07:12:14.0162653Z test_nn.py::TestNN::test_hardtanh_inplace_gradgrad PASSED [ 75%] 2023-03-31T07:12:14.0162792Z test_nn.py::TestNN::test_huber_loss_invalid_delta PASSED [ 75%] 2023-03-31T07:12:14.0162920Z test_nn.py::TestNN::test_inplace_thnn PASSED [ 75%] 2023-03-31T07:12:14.0163277Z test_nn.py::TestNN::test_interpolate PASSED [ 75%] 2023-03-31T07:12:14.0163447Z test_nn.py::TestNN::test_interpolate_bicubic_2d PASSED [ 75%] 2023-03-31T07:12:14.0163622Z test_nn.py::TestNN::test_interpolate_bicubic_2d_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0163767Z test_nn.py::TestNN::test_interpolate_bicubic_2d_zero_dim PASSED [ 75%] 2023-03-31T07:12:14.0163937Z test_nn.py::TestNN::test_interpolate_bicubic_2d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0164078Z test_nn.py::TestNN::test_interpolate_bicubic_scale_2d PASSED [ 75%] 2023-03-31T07:12:14.0164259Z test_nn.py::TestNN::test_interpolate_bicubic_scale_2d_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0164414Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_shared_2d PASSED [ 75%] 2023-03-31T07:12:14.0164613Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_shared_2d_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0164772Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d PASSED [ 75%] 2023-03-31T07:12:14.0164943Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_align_corners PASSED [ 75%] 2023-03-31T07:12:14.0165200Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0165396Z test_nn.py::TestNN::test_interpolate_bicubic_scale_tuple_skewed_2d_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0165527Z test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d PASSED [ 75%] 2023-03-31T07:12:14.0165686Z test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_align_corners PASSED [ 75%] 2023-03-31T07:12:14.0165882Z test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0166064Z test_nn.py::TestNN::test_interpolate_bicubic_tuple_2d_cuda SKIPPED (Excluded from CUDA tests) [ 75%] 2023-03-31T07:12:14.0166210Z test_nn.py::TestNN::test_interpolate_bilinear_2d PASSED [ 75%] 2023-03-31T07:12:14.0166381Z test_nn.py::TestNN::test_interpolate_bilinear_2d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0166579Z test_nn.py::TestNN::test_interpolate_bilinear_2d_zero_dim PASSED [ 76%] 2023-03-31T07:12:14.0166762Z test_nn.py::TestNN::test_interpolate_bilinear_2d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0166907Z test_nn.py::TestNN::test_interpolate_bilinear_scale_2d PASSED [ 76%] 2023-03-31T07:12:14.0167071Z test_nn.py::TestNN::test_interpolate_bilinear_scale_2d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0167228Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_shared_2d PASSED [ 76%] 2023-03-31T07:12:14.0167423Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_shared_2d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0167581Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d PASSED [ 76%] 2023-03-31T07:12:14.0167758Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_align_corners PASSED [ 76%] 2023-03-31T07:12:14.0167976Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0168172Z test_nn.py::TestNN::test_interpolate_bilinear_scale_tuple_skewed_2d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0168324Z test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d PASSED [ 76%] 2023-03-31T07:12:14.0168470Z test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_align_corners PASSED [ 76%] 2023-03-31T07:12:14.0168668Z test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0168847Z test_nn.py::TestNN::test_interpolate_bilinear_tuple_2d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0169026Z test_nn.py::TestNN::test_interpolate_buffer_overflow PASSED [ 76%] 2023-03-31T07:12:14.0169231Z test_nn.py::TestNN::test_interpolate_illegal_memory_access SKIPPED (CUDA unavailable) [ 76%] 2023-03-31T07:12:14.0169372Z test_nn.py::TestNN::test_interpolate_linear_1d PASSED [ 76%] 2023-03-31T07:12:14.0169523Z test_nn.py::TestNN::test_interpolate_linear_1d_align_corners PASSED [ 76%] 2023-03-31T07:12:14.0169708Z test_nn.py::TestNN::test_interpolate_linear_1d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0169877Z test_nn.py::TestNN::test_interpolate_linear_1d_cuda SKIPPED (Excluded from CUDA tests) [ 76%] 2023-03-31T07:12:14.0170009Z test_nn.py::TestNN::test_interpolate_linear_1d_zero_dim PASSED [ 77%] 2023-03-31T07:12:14.0170187Z test_nn.py::TestNN::test_interpolate_linear_1d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0170331Z test_nn.py::TestNN::test_interpolate_linear_scale_1d PASSED [ 77%] 2023-03-31T07:12:14.0170490Z test_nn.py::TestNN::test_interpolate_linear_scale_1d_align_corners PASSED [ 77%] 2023-03-31T07:12:14.0170683Z test_nn.py::TestNN::test_interpolate_linear_scale_1d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0170898Z test_nn.py::TestNN::test_interpolate_linear_scale_1d_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0171043Z test_nn.py::TestNN::test_interpolate_linear_tuple_1d PASSED [ 77%] 2023-03-31T07:12:14.0171219Z test_nn.py::TestNN::test_interpolate_linear_tuple_1d_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0171359Z test_nn.py::TestNN::test_interpolate_nearest_1d PASSED [ 77%] 2023-03-31T07:12:14.0171515Z test_nn.py::TestNN::test_interpolate_nearest_1d_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0171660Z test_nn.py::TestNN::test_interpolate_nearest_1d_zero_dim PASSED [ 77%] 2023-03-31T07:12:14.0171841Z test_nn.py::TestNN::test_interpolate_nearest_1d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0171983Z test_nn.py::TestNN::test_interpolate_nearest_2d PASSED [ 77%] 2023-03-31T07:12:14.0172192Z test_nn.py::TestNN::test_interpolate_nearest_2d_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0172345Z test_nn.py::TestNN::test_interpolate_nearest_2d_launch_configs PASSED [ 77%] 2023-03-31T07:12:14.0172532Z test_nn.py::TestNN::test_interpolate_nearest_2d_launch_configs_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0172675Z test_nn.py::TestNN::test_interpolate_nearest_2d_zero_dim PASSED [ 77%] 2023-03-31T07:12:14.0172843Z test_nn.py::TestNN::test_interpolate_nearest_2d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0172981Z test_nn.py::TestNN::test_interpolate_nearest_3d PASSED [ 77%] 2023-03-31T07:12:14.0173154Z test_nn.py::TestNN::test_interpolate_nearest_3d_cuda SKIPPED (Excluded from CUDA tests) [ 77%] 2023-03-31T07:12:14.0173299Z test_nn.py::TestNN::test_interpolate_nearest_3d_zero_dim PASSED [ 77%] 2023-03-31T07:12:14.0173476Z test_nn.py::TestNN::test_interpolate_nearest_3d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0173621Z test_nn.py::TestNN::test_interpolate_nearest_scale_1d PASSED [ 78%] 2023-03-31T07:12:14.0173802Z test_nn.py::TestNN::test_interpolate_nearest_scale_1d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0173943Z test_nn.py::TestNN::test_interpolate_nearest_scale_2d PASSED [ 78%] 2023-03-31T07:12:14.0174107Z test_nn.py::TestNN::test_interpolate_nearest_scale_2d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0174250Z test_nn.py::TestNN::test_interpolate_nearest_scale_3d PASSED [ 78%] 2023-03-31T07:12:14.0174426Z test_nn.py::TestNN::test_interpolate_nearest_scale_3d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0174603Z test_nn.py::TestNN::test_interpolate_nearest_tuple_1d PASSED [ 78%] 2023-03-31T07:12:14.0174815Z test_nn.py::TestNN::test_interpolate_nearest_tuple_1d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0174960Z test_nn.py::TestNN::test_interpolate_nearest_tuple_2d PASSED [ 78%] 2023-03-31T07:12:14.0175139Z test_nn.py::TestNN::test_interpolate_nearest_tuple_2d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0175282Z test_nn.py::TestNN::test_interpolate_nearest_tuple_3d PASSED [ 78%] 2023-03-31T07:12:14.0175456Z test_nn.py::TestNN::test_interpolate_nearest_tuple_3d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0175585Z test_nn.py::TestNN::test_interpolate_trilinear_3d PASSED [ 78%] 2023-03-31T07:12:14.0175758Z test_nn.py::TestNN::test_interpolate_trilinear_3d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0175902Z test_nn.py::TestNN::test_interpolate_trilinear_3d_zero_dim PASSED [ 78%] 2023-03-31T07:12:14.0176088Z test_nn.py::TestNN::test_interpolate_trilinear_3d_zero_dim_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0176233Z test_nn.py::TestNN::test_interpolate_trilinear_scale_3d PASSED [ 78%] 2023-03-31T07:12:14.0176397Z test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_align_corners PASSED [ 78%] 2023-03-31T07:12:14.0176629Z test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0176814Z test_nn.py::TestNN::test_interpolate_trilinear_scale_3d_cuda SKIPPED (Excluded from CUDA tests) [ 78%] 2023-03-31T07:12:14.0176947Z test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d PASSED [ 79%] 2023-03-31T07:12:14.0177108Z test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_align_corners PASSED [ 79%] 2023-03-31T07:12:14.0177305Z test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_align_corners_cuda SKIPPED (Excluded from CUDA tests) [ 79%] 2023-03-31T07:12:14.0177487Z test_nn.py::TestNN::test_interpolate_trilinear_tuple_3d_cuda SKIPPED (Excluded from CUDA tests) [ 79%] 2023-03-31T07:12:14.0177644Z test_nn.py::TestNN::test_interpolate_undefined_behavior_casting PASSED [ 79%] 2023-03-31T07:12:14.0177819Z test_nn.py::TestNN::test_kl_div_log_softmax_target PASSED [ 79%] 2023-03-31T07:12:14.0177958Z test_nn.py::TestNN::test_kl_div_with_diff_type PASSED [ 79%] 2023-03-31T07:12:14.0178101Z test_nn.py::TestNN::test_kl_div_with_diff_type_log_target PASSED [ 79%] 2023-03-31T07:12:14.0178234Z test_nn.py::TestNN::test_l1_loss_correct PASSED [ 79%] 2023-03-31T07:12:14.0178370Z test_nn.py::TestNN::test_layer_norm_grads_with_create_graph_flag PASSED [ 79%] 2023-03-31T07:12:14.0178524Z test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCOO PASSED [ 79%] 2023-03-31T07:12:14.0178681Z test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCSC PASSED [ 79%] 2023-03-31T07:12:14.0178836Z test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightCSR PASSED [ 79%] 2023-03-31T07:12:14.0179001Z test_nn.py::TestNN::test_linear_autograd_device_cpu_bias_weightStrided PASSED [ 79%] 2023-03-31T07:12:14.0179160Z test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCOO PASSED [ 79%] 2023-03-31T07:12:14.0179315Z test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCSC PASSED [ 79%] 2023-03-31T07:12:14.0179471Z test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightCSR PASSED [ 79%] 2023-03-31T07:12:14.0179624Z test_nn.py::TestNN::test_linear_autograd_device_cpu_nobias_weightStrided PASSED [ 79%] 2023-03-31T07:12:14.0179763Z test_nn.py::TestNN::test_linear_broadcasting PASSED [ 79%] 2023-03-31T07:12:14.0179897Z test_nn.py::TestNN::test_load_state_dict PASSED [ 79%] 2023-03-31T07:12:14.0180033Z test_nn.py::TestNN::test_load_state_dict_BC PASSED [ 79%] 2023-03-31T07:12:14.0180200Z test_nn.py::TestNN::test_load_state_dict_child PASSED [ 80%] 2023-03-31T07:12:14.0180361Z test_nn.py::TestNN::test_load_state_dict_custom PASSED [ 80%] 2023-03-31T07:12:14.0180499Z test_nn.py::TestNN::test_load_state_dict_invalid PASSED [ 80%] 2023-03-31T07:12:14.0180639Z test_nn.py::TestNN::test_load_state_dict_ref_cycle PASSED [ 80%] 2023-03-31T07:12:14.0180763Z test_nn.py::TestNN::test_load_state_dict_type PASSED [ 80%] 2023-03-31T07:12:14.0180895Z test_nn.py::TestNN::test_log_softmax_cpu PASSED [ 80%] 2023-03-31T07:12:14.0181026Z test_nn.py::TestNN::test_log_softmax_dim0 PASSED [ 80%] 2023-03-31T07:12:14.0181188Z test_nn.py::TestNN::test_log_softmax_dim0_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0181320Z test_nn.py::TestNN::test_log_softmax_dim3 PASSED [ 80%] 2023-03-31T07:12:14.0181482Z test_nn.py::TestNN::test_log_softmax_dim3_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0181617Z test_nn.py::TestNN::test_log_softmax_lastdim PASSED [ 80%] 2023-03-31T07:12:14.0181783Z test_nn.py::TestNN::test_log_softmax_lastdim_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0181905Z test_nn.py::TestNN::test_log_softmax_scalar PASSED [ 80%] 2023-03-31T07:12:14.0182100Z test_nn.py::TestNN::test_log_softmax_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0182239Z test_nn.py::TestNN::test_log_softmax_spatial PASSED [ 80%] 2023-03-31T07:12:14.0182403Z test_nn.py::TestNN::test_log_softmax_spatial_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0182545Z test_nn.py::TestNN::test_log_softmax_spatial_special PASSED [ 80%] 2023-03-31T07:12:14.0182721Z test_nn.py::TestNN::test_log_softmax_spatial_special_cuda SKIPPED (Excluded from CUDA tests) [ 80%] 2023-03-31T07:12:14.0182865Z test_nn.py::TestNN::test_loss_equal_input_target_shape PASSED [ 80%] 2023-03-31T07:12:14.0183016Z test_nn.py::TestNN::test_margin_ranking_loss_margin_no_reduce PASSED [ 80%] 2023-03-31T07:12:14.0183158Z test_nn.py::TestNN::test_margin_ranking_loss_no_reduce PASSED [ 80%] 2023-03-31T07:12:14.0183422Z test_nn.py::TestNN::test_module_apply_inplace_op PASSED [ 81%] 2023-03-31T07:12:14.0183572Z test_nn.py::TestNN::test_module_backcompat PASSED [ 81%] 2023-03-31T07:12:14.0183706Z test_nn.py::TestNN::test_module_super_init PASSED [ 81%] 2023-03-31T07:12:14.0183841Z test_nn.py::TestNN::test_module_to_argparse PASSED [ 81%] 2023-03-31T07:12:14.0183968Z test_nn.py::TestNN::test_modules PASSED [ 81%] 2023-03-31T07:12:14.0184104Z test_nn.py::TestNN::test_mse_loss_size_warning PASSED [ 81%] 2023-03-31T07:12:14.0184267Z test_nn.py::TestNN::test_multimarginloss_1d_input_0d_target_no_reduce PASSED [ 81%] 2023-03-31T07:12:14.0184470Z test_nn.py::TestNN::test_multimarginloss_1d_input_0d_target_no_reduce_cuda SKIPPED (Excluded from CUDA tests) [ 81%] 2023-03-31T07:12:14.0184593Z test_nn.py::TestNN::test_named_children PASSED [ 81%] 2023-03-31T07:12:14.0184727Z test_nn.py::TestNN::test_named_modules PASSED [ 81%] 2023-03-31T07:12:14.0184874Z test_nn.py::TestNN::test_named_parameters_remove_duplicate PASSED [ 81%] 2023-03-31T07:12:14.0185012Z test_nn.py::TestNN::test_nested_tensor_from_mask PASSED [ 81%] 2023-03-31T07:12:14.0185152Z test_nn.py::TestNN::test_nested_tensor_from_mask_error PASSED [ 81%] 2023-03-31T07:12:14.0185277Z test_nn.py::TestNN::test_no_grad PASSED [ 81%] 2023-03-31T07:12:14.0185412Z test_nn.py::TestNN::test_non_leaf_parameters PASSED [ 81%] 2023-03-31T07:12:14.0185540Z test_nn.py::TestNN::test_normalize PASSED [ 81%] 2023-03-31T07:12:14.0185711Z test_nn.py::TestNN::test_overwrite_module_params_on_conversion PASSED [ 81%] 2023-03-31T07:12:14.0185916Z test_nn.py::TestNN::test_pack_sequence_batch_sizes_throw SKIPPED (CUDA not available) [ 81%] 2023-03-31T07:12:14.0186052Z test_nn.py::TestNN::test_pad_scalar_error PASSED [ 81%] 2023-03-31T07:12:14.0186184Z test_nn.py::TestNN::test_padding_list PASSED [ 81%] 2023-03-31T07:12:14.0186316Z test_nn.py::TestNN::test_pairwise_distance PASSED [ 81%] 2023-03-31T07:12:14.0186452Z test_nn.py::TestNN::test_parameter_assignment PASSED [ 82%] 2023-03-31T07:12:14.0186594Z test_nn.py::TestNN::test_parameterlistdict_pickle PASSED [ 82%] 2023-03-31T07:12:14.0186748Z test_nn.py::TestNN::test_parameterlistdict_setting_attributes PASSED [ 82%] 2023-03-31T07:12:14.0186881Z test_nn.py::TestNN::test_parameters_and_named_parameters PASSED [ 82%] 2023-03-31T07:12:14.0187018Z test_nn.py::TestNN::test_parameters_to_vector PASSED [ 82%] 2023-03-31T07:12:14.0187147Z test_nn.py::TestNN::test_parse_to PASSED [ 82%] 2023-03-31T07:12:14.0187303Z test_nn.py::TestNN::test_partial_flat_weights SKIPPED (CUDA not available) [ 82%] 2023-03-31T07:12:14.0187466Z test_nn.py::TestNN::test_pdist PASSED [ 82%] 2023-03-31T07:12:14.0187614Z test_nn.py::TestNN::test_pdist_cpu_gradgrad_unimplemented XFAIL [ 82%] 2023-03-31T07:12:14.0187761Z test_nn.py::TestNN::test_pdist_cuda_gradgrad_unimplemented XFAIL [ 82%] 2023-03-31T07:12:14.0187894Z test_nn.py::TestNN::test_pdist_empty_col PASSED [ 82%] 2023-03-31T07:12:14.0188012Z test_nn.py::TestNN::test_pdist_empty_row PASSED [ 82%] 2023-03-31T07:12:14.0188140Z test_nn.py::TestNN::test_pdist_large PASSED [ 82%] 2023-03-31T07:12:14.0188269Z test_nn.py::TestNN::test_pdist_zeros PASSED [ 82%] 2023-03-31T07:12:14.0188407Z test_nn.py::TestNN::test_pixel_shuffle_nhwc_cpu PASSED [ 82%] 2023-03-31T07:12:14.0188543Z test_nn.py::TestNN::test_pixel_shuffle_unshuffle PASSED [ 82%] 2023-03-31T07:12:14.0188717Z test_nn.py::TestNN::test_pointwise_loss_broadcast PASSED [ 82%] 2023-03-31T07:12:14.0188875Z test_nn.py::TestNN::test_pointwise_loss_target_grad_none_reduction PASSED [ 82%] 2023-03-31T07:12:14.0189017Z test_nn.py::TestNN::test_projections_errors_on_gru_and_rnn PASSED [ 82%] 2023-03-31T07:12:14.0189144Z test_nn.py::TestNN::test_projections_lstm_args_check PASSED [ 82%] 2023-03-31T07:12:14.0189448Z test_nn.py::TestNN::test_projections_lstm_check_device SKIPPED (multi-GPU not supported) [ 82%] 2023-03-31T07:12:14.0189595Z test_nn.py::TestNN::test_projections_lstm_initial_hidden_state PASSED [ 83%] 2023-03-31T07:12:14.0189763Z test_nn.py::TestNN::test_register_buffer_allows_overwriting_with_same_name PASSED [ 83%] 2023-03-31T07:12:14.0189921Z test_nn.py::TestNN::test_register_buffer_raises_error_if_attr_exists PASSED [ 83%] 2023-03-31T07:12:14.0190083Z test_nn.py::TestNN::test_register_buffer_raises_error_if_name_is_not_string PASSED [ 83%] 2023-03-31T07:12:14.0190240Z test_nn.py::TestNN::test_register_buffer_raises_error_if_not_tensor PASSED [ 83%] 2023-03-31T07:12:14.0190411Z test_nn.py::TestNN::test_register_parameter_allows_overwriting_with_same_name PASSED [ 83%] 2023-03-31T07:12:14.0190572Z test_nn.py::TestNN::test_register_parameter_raises_error_if_attr_exists PASSED [ 83%] 2023-03-31T07:12:14.0190728Z test_nn.py::TestNN::test_register_parameter_raises_error_if_name_is_not_string PASSED [ 83%] 2023-03-31T07:12:14.0190870Z test_nn.py::TestNN::test_register_state_dict_pre_hook PASSED [ 83%] 2023-03-31T07:12:14.0191029Z test_nn.py::TestNN::test_register_state_dict_pre_hook_backward_compat PASSED [ 83%] 2023-03-31T07:12:14.0191243Z test_nn.py::TestNN::test_relu_inplace_on_view PASSED [ 83%] 2023-03-31T07:12:14.0191398Z test_nn.py::TestNN::test_repr PASSED [ 83%] 2023-03-31T07:12:14.0191531Z test_nn.py::TestNN::test_requires_grad_ PASSED [ 83%] 2023-03-31T07:12:14.0191665Z test_nn.py::TestNN::test_rnn_args_check PASSED [ 83%] 2023-03-31T07:12:14.0191904Z test_nn.py::TestNN::test_rnn_check_device SKIPPED (multi-GPU not supported) [ 83%] 2023-03-31T07:12:14.0192032Z test_nn.py::TestNN::test_rnn_initial_hidden_state PASSED [ 83%] 2023-03-31T07:12:14.0192166Z test_nn.py::TestNN::test_rnn_weight_norm PASSED [ 83%] 2023-03-31T07:12:14.0192296Z test_nn.py::TestNN::test_share_memory PASSED [ 83%] 2023-03-31T07:12:14.0192440Z test_nn.py::TestNN::test_smoothl1loss_intergral_target PASSED [ 83%] 2023-03-31T07:12:14.0192591Z test_nn.py::TestNN::test_smoothl1loss_negative_beta_not_supported PASSED [ 83%] 2023-03-31T07:12:14.0192725Z test_nn.py::TestNN::test_softmax_cpu PASSED [ 83%] 2023-03-31T07:12:14.0192864Z test_nn.py::TestNN::test_softmax_functional_dim0 PASSED [ 84%] 2023-03-31T07:12:14.0193041Z test_nn.py::TestNN::test_softmax_functional_dim0_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0193200Z test_nn.py::TestNN::test_softmax_functional_dim3 PASSED [ 84%] 2023-03-31T07:12:14.0193372Z test_nn.py::TestNN::test_softmax_functional_dim3_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0193514Z test_nn.py::TestNN::test_softmax_functional_scalar PASSED [ 84%] 2023-03-31T07:12:14.0193688Z test_nn.py::TestNN::test_softmax_functional_scalar_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0193822Z test_nn.py::TestNN::test_softmax_lastdim PASSED [ 84%] 2023-03-31T07:12:14.0193983Z test_nn.py::TestNN::test_softmax_lastdim_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0194122Z test_nn.py::TestNN::test_softmax_lastdim_dtype PASSED [ 84%] 2023-03-31T07:12:14.0194294Z test_nn.py::TestNN::test_softmax_lastdim_dtype_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0194450Z test_nn.py::TestNN::test_softmax_spatial PASSED [ 84%] 2023-03-31T07:12:14.0194614Z test_nn.py::TestNN::test_softmax_spatial_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0194749Z test_nn.py::TestNN::test_softmax_spatial_dtype PASSED [ 84%] 2023-03-31T07:12:14.0194916Z test_nn.py::TestNN::test_softmax_spatial_dtype_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0195054Z test_nn.py::TestNN::test_softmax_spatial_special PASSED [ 84%] 2023-03-31T07:12:14.0195223Z test_nn.py::TestNN::test_softmax_spatial_special_cuda SKIPPED (Excluded from CUDA tests) [ 84%] 2023-03-31T07:12:14.0195354Z test_nn.py::TestNN::test_softmin PASSED [ 84%] 2023-03-31T07:12:14.0195488Z test_nn.py::TestNN::test_spectral_norm PASSED [ 84%] 2023-03-31T07:12:14.0195621Z test_nn.py::TestNN::test_spectral_norm_dim PASSED [ 84%] 2023-03-31T07:12:14.0195748Z test_nn.py::TestNN::test_spectral_norm_forward PASSED [ 84%] 2023-03-31T07:12:14.0195894Z test_nn.py::TestNN::test_spectral_norm_load_state_dict PASSED [ 84%] 2023-03-31T07:12:14.0196029Z test_nn.py::TestNN::test_spectral_norm_pickle PASSED [ 85%] 2023-03-31T07:12:14.0196156Z test_nn.py::TestNN::test_state_dict PASSED [ 85%] 2023-03-31T07:12:14.0196320Z test_nn.py::TestNN::test_sync_batchnorm_accuracy_cuda SKIPPED (CUDA not available) [ 85%] 2023-03-31T07:12:14.0196486Z test_nn.py::TestNN::test_sync_batchnorm_backward_elemt SKIPPED (CUDA not available) [ 85%] 2023-03-31T07:12:14.0196622Z test_nn.py::TestNN::test_threshold_bfloat16 PASSED [ 85%] 2023-03-31T07:12:14.0196779Z test_nn.py::TestNN::test_threshold_int PASSED [ 85%] 2023-03-31T07:12:14.0196916Z test_nn.py::TestNN::test_to PASSED [ 85%] 2023-03-31T07:12:14.0197056Z test_nn.py::TestNN::test_train_errors_for_invalid_mode PASSED [ 85%] 2023-03-31T07:12:14.0197196Z test_nn.py::TestNN::test_transformer_args_check PASSED [ 85%] 2023-03-31T07:12:14.0197336Z test_nn.py::TestNN::test_transformer_layer_args_check PASSED [ 85%] 2023-03-31T07:12:14.0197476Z test_nn.py::TestNN::test_transformerdecoder PASSED [ 85%] 2023-03-31T07:12:14.0197625Z test_nn.py::TestNN::test_transformerdecoderlayer PASSED [ 85%] 2023-03-31T07:12:14.0197777Z test_nn.py::TestNN::test_transformerdecoderlayer_gelu PASSED [ 85%] 2023-03-31T07:12:14.0197915Z test_nn.py::TestNN::test_triplet_margin_loss PASSED [ 85%] 2023-03-31T07:12:14.0198043Z test_nn.py::TestNN::test_triplet_margin_loss_no_reduce PASSED [ 85%] 2023-03-31T07:12:14.0198182Z test_nn.py::TestNN::test_triplet_margin_loss_swap PASSED [ 85%] 2023-03-31T07:12:14.0198325Z test_nn.py::TestNN::test_triplet_margin_loss_swap_no_reduce PASSED [ 85%] 2023-03-31T07:12:14.0198451Z test_nn.py::TestNN::test_type PASSED [ 85%] 2023-03-31T07:12:14.0198611Z test_nn.py::TestNN::test_unflatten PASSED [ 85%] 2023-03-31T07:12:14.0198746Z test_nn.py::TestNN::test_unflatten_invalid_arg PASSED [ 85%] 2023-03-31T07:12:14.0198878Z test_nn.py::TestNN::test_unfold_invalid_arg PASSED [ 85%] 2023-03-31T07:12:14.0199034Z test_nn.py::TestNN::test_upsamplingBilinear2d_spatial_invariance PASSED [ 86%] 2023-03-31T07:12:14.0199162Z test_nn.py::TestNN::test_upsamplingLinear1d PASSED [ 86%] 2023-03-31T07:12:14.0199319Z test_nn.py::TestNN::test_upsamplingLinear1d_spatial_invariance PASSED [ 86%] 2023-03-31T07:12:14.0199481Z test_nn.py::TestNN::test_upsamplingTrilinear3d_spatial_invariance PASSED [ 86%] 2023-03-31T07:12:14.0199617Z test_nn.py::TestNN::test_upsampling_bfloat16 PASSED [ 86%] 2023-03-31T07:12:14.0199795Z test_nn.py::TestNN::test_upsampling_not_recompute_scale_factor PASSED [ 86%] 2023-03-31T07:12:14.0199935Z test_nn.py::TestNN::test_upsampling_small_scale PASSED [ 86%] 2023-03-31T07:12:14.0200069Z test_nn.py::TestNN::test_vector_to_parameters PASSED [ 86%] 2023-03-31T07:12:14.0200197Z test_nn.py::TestNN::test_weight_norm PASSED [ 86%] 2023-03-31T07:12:14.0200319Z test_nn.py::TestNN::test_weight_norm_pickle PASSED [ 86%] 2023-03-31T07:12:14.0200443Z test_nn.py::TestNN::test_zero_grad PASSED [ 86%] 2023-03-31T07:12:14.0200587Z test_nn.py::TestFusionEval::test_fuse_module_eval_numerics PASSED [ 86%] 2023-03-31T07:12:14.0200735Z test_nn.py::TestConstantPadNd::test_constant_pad_nd PASSED [ 86%] 2023-03-31T07:12:14.0200892Z test_nn.py::TestConstantPadNd::test_preserves_memory_format PASSED [ 86%] 2023-03-31T07:12:14.0201025Z test_nn.py::TestAddRelu::test_add_relu PASSED [ 86%] 2023-03-31T07:12:14.0201167Z test_nn.py::TestAddRelu::test_add_relu_broadcasting PASSED [ 86%] 2023-03-31T07:12:14.0201318Z test_nn.py::TestFunctionalPickle::test_pickle_softsign PASSED [ 86%] 2023-03-31T07:12:14.0201451Z test_nn.py::TestFusionUtils::test_fuse_conv_bn_requires_grad PASSED [ 86%] 2023-03-31T07:12:14.0201599Z test_nn.py::TestFusionUtils::test_fuse_linear_bn_requires_grad PASSED [ 86%] 2023-03-31T07:12:14.0201753Z test_nn.py::TestNNDeviceTypeCPU::test_BatchNorm_empty_cpu PASSED [ 86%] 2023-03-31T07:12:14.0201905Z test_nn.py::TestNNDeviceTypeCPU::test_Bilinear_empty_cpu PASSED [ 86%] 2023-03-31T07:12:14.0202110Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_cudnn_cpu SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T07:12:14.0202291Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_empty_target_cpu PASSED [ 87%] 2023-03-31T07:12:14.0202491Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_mean_use_module_form_False_cpu PASSED [ 87%] 2023-03-31T07:12:14.0202700Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_mean_use_module_form_True_cpu PASSED [ 87%] 2023-03-31T07:12:14.0202903Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_none_use_module_form_False_cpu PASSED [ 87%] 2023-03-31T07:12:14.0203230Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_none_use_module_form_True_cpu PASSED [ 87%] 2023-03-31T07:12:14.0203430Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_sum_use_module_form_False_cpu PASSED [ 87%] 2023-03-31T07:12:14.0203628Z test_nn.py::TestNNDeviceTypeCPU::test_CTCLoss_no_batch_dim_reduction_sum_use_module_form_True_cpu PASSED [ 87%] 2023-03-31T07:12:14.0203801Z test_nn.py::TestNNDeviceTypeCPU::test_GRU_grad_and_gradgrad_cpu_float64 PASSED [ 87%] 2023-03-31T07:12:14.0203961Z test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_empty_cpu PASSED [ 87%] 2023-03-31T07:12:14.0204120Z test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_general_cpu PASSED [ 87%] 2023-03-31T07:12:14.0204354Z test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_memory_format_cpu PASSED [ 87%] 2023-03-31T07:12:14.0204515Z test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_numeric_cpu PASSED [ 87%] 2023-03-31T07:12:14.0204688Z test_nn.py::TestNNDeviceTypeCPU::test_GroupNorm_raises_error_if_one_value_per_group_cpu PASSED [ 87%] 2023-03-31T07:12:14.0204849Z test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm1d_general_cpu PASSED [ 87%] 2023-03-31T07:12:14.0205011Z test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm2d_general_cpu PASSED [ 87%] 2023-03-31T07:12:14.0205172Z test_nn.py::TestNNDeviceTypeCPU::test_InstanceNorm3d_general_cpu PASSED [ 87%] 2023-03-31T07:12:14.0205349Z test_nn.py::TestNNDeviceTypeCPU::test_LSTM_grad_and_gradgrad_cpu_float64 PASSED [ 87%] 2023-03-31T07:12:14.0205508Z test_nn.py::TestNNDeviceTypeCPU::test_LayerNorm_general_cpu PASSED [ 87%] 2023-03-31T07:12:14.0205727Z test_nn.py::TestNNDeviceTypeCPU::test_LayerNorm_numeric_cpu PASSED [ 87%] 2023-03-31T07:12:14.0205894Z test_nn.py::TestNNDeviceTypeCPU::test_LocalResponseNorm_empty_cpu PASSED [ 87%] 2023-03-31T07:12:14.0206058Z test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_empty_cpu_float32 PASSED [ 87%] 2023-03-31T07:12:14.0206208Z test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_empty_cpu_float64 PASSED [ 88%] 2023-03-31T07:12:14.0206390Z test_nn.py::TestNNDeviceTypeCPU::test_MarginLoss_warnings_cpu SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T07:12:14.0206580Z test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad2d_large_cpu SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T07:12:14.0206760Z test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad3d_large_cpu SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T07:12:14.0206937Z test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad_empty_cpu_complex64 PASSED [ 88%] 2023-03-31T07:12:14.0207108Z test_nn.py::TestNNDeviceTypeCPU::test_ReflectionPad_empty_cpu_float32 PASSED [ 88%] 2023-03-31T07:12:14.0207275Z test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad1d_large_cpu PASSED [ 88%] 2023-03-31T07:12:14.0207440Z test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad2d_large_cpu PASSED [ 88%] 2023-03-31T07:12:14.0207589Z test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad3d_large_cpu PASSED [ 88%] 2023-03-31T07:12:14.0207765Z test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad_empty_cpu_complex128 PASSED [ 88%] 2023-03-31T07:12:14.0207938Z test_nn.py::TestNNDeviceTypeCPU::test_ReplicationPad_empty_cpu_float64 PASSED [ 88%] 2023-03-31T07:12:14.0208118Z test_nn.py::TestNNDeviceTypeCPU::test_TransformerDecoderLayer_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0208287Z test_nn.py::TestNNDeviceTypeCPU::test_TransformerDecoder_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0208510Z test_nn.py::TestNNDeviceTypeCPU::test_TransformerEncoderLayer_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0208711Z test_nn.py::TestNNDeviceTypeCPU::test_TransformerEncoder_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0208873Z test_nn.py::TestNNDeviceTypeCPU::test_Transformer_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0209016Z test_nn.py::TestNNDeviceTypeCPU::test_Unfold_empty_cpu PASSED [ 88%] 2023-03-31T07:12:14.0209199Z test_nn.py::TestNNDeviceTypeCPU::test_activations_bfloat16_cpu SKIPPED (Only runs on cuda) [ 88%] 2023-03-31T07:12:14.0209363Z test_nn.py::TestNNDeviceTypeCPU::test_activations_bfloat16_cpu_cpu PASSED [ 88%] 2023-03-31T07:12:14.0209515Z test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate0_cpu PASSED [ 88%] 2023-03-31T07:12:14.0209668Z test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate45_cpu PASSED [ 88%] 2023-03-31T07:12:14.0209822Z test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotate90_cpu PASSED [ 89%] 2023-03-31T07:12:14.0209982Z test_nn.py::TestNNDeviceTypeCPU::test_affine_2d_rotateRandom_cpu PASSED [ 89%] 2023-03-31T07:12:14.0210141Z test_nn.py::TestNNDeviceTypeCPU::test_affine_3d_rotateRandom_cpu PASSED [ 89%] 2023-03-31T07:12:14.0210291Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_cpu_float32 PASSED [ 89%] 2023-03-31T07:12:14.0210523Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_mixed_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0210720Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_affine_mixed_cpu_float16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0210877Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_cpu_float32 PASSED [ 89%] 2023-03-31T07:12:14.0211071Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_mixed_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0211263Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_eval_mixed_cpu_float16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0211422Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_grad_cpu PASSED [ 89%] 2023-03-31T07:12:14.0211617Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_large_batch_cpu_float16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0211805Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_large_batch_cpu_float32 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0212007Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_cpu_float32 PASSED [ 89%] 2023-03-31T07:12:14.0212213Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_mixed_cpu_bfloat16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0212416Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_simple_average_mixed_cpu_float16 SKIPPED (Only runs on cuda) [ 89%] 2023-03-31T07:12:14.0212576Z test_nn.py::TestNNDeviceTypeCPU::test_batchnorm_update_stats_cpu PASSED [ 89%] 2023-03-31T07:12:14.0212751Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_error_if_nonfinite_cpu PASSED [ 89%] 2023-03-31T07:12:14.0212936Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_0_5_cpu PASSED [ 89%] 2023-03-31T07:12:14.0213120Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_1_5_cpu PASSED [ 89%] 2023-03-31T07:12:14.0213304Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_2_cpu PASSED [ 89%] 2023-03-31T07:12:14.0213487Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_4_cpu PASSED [ 89%] 2023-03-31T07:12:14.0213662Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_False_norm_type_inf_cpu PASSED [ 90%] 2023-03-31T07:12:14.0213842Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_0_5_cpu PASSED [ 90%] 2023-03-31T07:12:14.0214020Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_1_5_cpu PASSED [ 90%] 2023-03-31T07:12:14.0214198Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_2_cpu PASSED [ 90%] 2023-03-31T07:12:14.0214412Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_4_cpu PASSED [ 90%] 2023-03-31T07:12:14.0214619Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_foreach_True_norm_type_inf_cpu PASSED [ 90%] 2023-03-31T07:12:14.0214825Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_multi_device_foreach_False_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:12:14.0215032Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_norm_multi_device_foreach_True_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:12:14.0215203Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_value_foreach_False_cpu PASSED [ 90%] 2023-03-31T07:12:14.0215360Z test_nn.py::TestNNDeviceTypeCPU::test_clip_grad_value_foreach_True_cpu PASSED [ 90%] 2023-03-31T07:12:14.0215525Z test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_bfloat16 PASSED [ 90%] 2023-03-31T07:12:14.0215692Z test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_complex128 PASSED [ 90%] 2023-03-31T07:12:14.0215852Z test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_float32 PASSED [ 90%] 2023-03-31T07:12:14.0216014Z test_nn.py::TestNNDeviceTypeCPU::test_conv_empty_input_cpu_float64 PASSED [ 90%] 2023-03-31T07:12:14.0216226Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_consistent_index_target_and_probs_cpu PASSED [ 90%] 2023-03-31T07:12:14.0216405Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_errors_cpu PASSED [ 90%] 2023-03-31T07:12:14.0216635Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_weight_ignore_indices_cpu PASSED [ 90%] 2023-03-31T07:12:14.0216805Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_label_smoothing_with_probs_cpu PASSED [ 90%] 2023-03-31T07:12:14.0217009Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_mean_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:12:14.0217211Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_none_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:12:14.0217416Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_large_tensor_reduction_sum_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:12:14.0217606Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_index_target_unit_weights_cpu PASSED [ 91%] 2023-03-31T07:12:14.0217819Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_one_hot_target_cpu PASSED [ 91%] 2023-03-31T07:12:14.0218008Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_all_reductions_cpu PASSED [ 91%] 2023-03-31T07:12:14.0218229Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_mean_weighted_False_cpu PASSED [ 91%] 2023-03-31T07:12:14.0218446Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_mean_weighted_True_cpu PASSED [ 91%] 2023-03-31T07:12:14.0218665Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_none_weighted_False_cpu PASSED [ 91%] 2023-03-31T07:12:14.0218870Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_none_weighted_True_cpu PASSED [ 91%] 2023-03-31T07:12:14.0219088Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_sum_weighted_False_cpu PASSED [ 91%] 2023-03-31T07:12:14.0219309Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_no_batch_dim_reduction_sum_weighted_True_cpu PASSED [ 91%] 2023-03-31T07:12:14.0219495Z test_nn.py::TestNNDeviceTypeCPU::test_cross_entropy_loss_prob_target_unit_weights_cpu PASSED [ 91%] 2023-03-31T07:12:14.0219644Z test_nn.py::TestNNDeviceTypeCPU::test_ctc_loss_cpu PASSED [ 91%] 2023-03-31T07:12:14.0219816Z test_nn.py::TestNNDeviceTypeCPU::test_ctc_loss_cudnn_cpu SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T07:12:14.0219986Z test_nn.py::TestNNDeviceTypeCPU::test_device_mask_cpu SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T07:12:14.0220149Z test_nn.py::TestNNDeviceTypeCPU::test_elu_inplace_overlap_cpu PASSED [ 91%] 2023-03-31T07:12:14.0220348Z test_nn.py::TestNNDeviceTypeCPU::test_elu_inplace_with_neg_alpha_cpu PASSED [ 91%] 2023-03-31T07:12:14.0220507Z test_nn.py::TestNNDeviceTypeCPU::test_fold_cpu PASSED [ 91%] 2023-03-31T07:12:14.0220658Z test_nn.py::TestNNDeviceTypeCPU::test_glu_bfloat16_cpu PASSED [ 91%] 2023-03-31T07:12:14.0220849Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_half_precision_cpu SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T07:12:14.0221028Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_cpu SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T07:12:14.0221200Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_2d_cpu_float32 PASSED [ 91%] 2023-03-31T07:12:14.0221404Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_2d_cpu_float64 SKIPPED (Insufficient cpu memory) [ 91%] 2023-03-31T07:12:14.0221574Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_3d_cpu_float32 PASSED [ 92%] 2023-03-31T07:12:14.0221784Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_large_index_3d_cpu_float64 SKIPPED (Insufficient cpu memory) [ 92%] 2023-03-31T07:12:14.0221938Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_nan_inf_cpu_float32 PASSED [ 92%] 2023-03-31T07:12:14.0222102Z test_nn.py::TestNNDeviceTypeCPU::test_grid_sample_nan_inf_cpu_float64 PASSED [ 92%] 2023-03-31T07:12:14.0222264Z test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_bfloat16 PASSED [ 92%] 2023-03-31T07:12:14.0222457Z test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_float32 PASSED [ 92%] 2023-03-31T07:12:14.0222619Z test_nn.py::TestNNDeviceTypeCPU::test_groupnorm_nhwc_cpu_float64 PASSED [ 92%] 2023-03-31T07:12:14.0222775Z test_nn.py::TestNNDeviceTypeCPU::test_gumbel_softmax_cpu_float32 PASSED [ 92%] 2023-03-31T07:12:14.0222931Z test_nn.py::TestNNDeviceTypeCPU::test_gumbel_softmax_cpu_float64 PASSED [ 92%] 2023-03-31T07:12:14.0223085Z test_nn.py::TestNNDeviceTypeCPU::test_hardsigmoid_grad_cpu PASSED [ 92%] 2023-03-31T07:12:14.0223225Z test_nn.py::TestNNDeviceTypeCPU::test_hardswish_grad_cpu PASSED [ 92%] 2023-03-31T07:12:14.0223485Z test_nn.py::TestNNDeviceTypeCPU::test_hardswish_inplace_overlap_cpu PASSED [ 92%] 2023-03-31T07:12:14.0223711Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_for_single_spatial_element_during_training_cpu PASSED [ 92%] 2023-03-31T07:12:14.0224018Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_False_affine_False_cpu PASSED [ 92%] 2023-03-31T07:12:14.0224277Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_False_affine_True_cpu PASSED [ 92%] 2023-03-31T07:12:14.0224529Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_True_affine_False_cpu PASSED [ 92%] 2023-03-31T07:12:14.0224786Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm1d_no_batch_dim_True_affine_True_cpu PASSED [ 92%] 2023-03-31T07:12:14.0225049Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_False_affine_False_cpu PASSED [ 92%] 2023-03-31T07:12:14.0225305Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_False_affine_True_cpu PASSED [ 92%] 2023-03-31T07:12:14.0225553Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_True_affine_False_cpu PASSED [ 92%] 2023-03-31T07:12:14.0225809Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm2d_no_batch_dim_True_affine_True_cpu PASSED [ 92%] 2023-03-31T07:12:14.0226065Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_False_affine_False_cpu PASSED [ 93%] 2023-03-31T07:12:14.0226333Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_False_affine_True_cpu PASSED [ 93%] 2023-03-31T07:12:14.0226624Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_True_affine_False_cpu PASSED [ 93%] 2023-03-31T07:12:14.0226881Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_input_channels_is_not_num_features_InstanceNorm3d_no_batch_dim_True_affine_True_cpu PASSED [ 93%] 2023-03-31T07:12:14.0227083Z test_nn.py::TestNNDeviceTypeCPU::test_instancenorm_raises_error_if_less_than_one_value_per_channel_cpu PASSED [ 93%] 2023-03-31T07:12:14.0227246Z test_nn.py::TestNNDeviceTypeCPU::test_invalid_reduction_strings_cpu PASSED [ 93%] 2023-03-31T07:12:14.0227434Z test_nn.py::TestNNDeviceTypeCPU::test_layernorm_half_precision_cpu SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T07:12:14.0227619Z test_nn.py::TestNNDeviceTypeCPU::test_layernorm_weight_bias_cpu SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T07:12:14.0227791Z test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_overlap_cpu PASSED [ 93%] 2023-03-31T07:12:14.0227961Z test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_with_neg_slope_cpu PASSED [ 93%] 2023-03-31T07:12:14.0228124Z test_nn.py::TestNNDeviceTypeCPU::test_leaky_relu_inplace_with_zero_slope_cpu PASSED [ 93%] 2023-03-31T07:12:14.0228328Z test_nn.py::TestNNDeviceTypeCPU::test_linear_empty_cpu PASSED [ 93%] 2023-03-31T07:12:14.0228489Z test_nn.py::TestNNDeviceTypeCPU::test_log_softmax_big_cpu_float32 PASSED [ 93%] 2023-03-31T07:12:14.0228644Z test_nn.py::TestNNDeviceTypeCPU::test_logsigmoid_out_cpu PASSED [ 93%] 2023-03-31T07:12:14.0228860Z test_nn.py::TestNNDeviceTypeCPU::test_lstmcell_backward_only_one_output_grad_cpu_float64 SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T07:12:14.0229051Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_TxT_layout_cpu SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T07:12:14.0229207Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_cpu PASSED [ 93%] 2023-03-31T07:12:14.0229401Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_devices_parity_cpu SKIPPED (Only runs on cuda) [ 93%] 2023-03-31T07:12:14.0229614Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_forward_with_nans_cpu PASSED [ 93%] 2023-03-31T07:12:14.0229760Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_grad_cpu PASSED [ 93%] 2023-03-31T07:12:14.0229923Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_mask_types_cpu PASSED [ 93%] 2023-03-31T07:12:14.0230119Z test_nn.py::TestNNDeviceTypeCPU::test_masked_softmax_transformer_layout_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:12:14.0230279Z test_nn.py::TestNNDeviceTypeCPU::test_mish_inplace_overlap_cpu PASSED [ 94%] 2023-03-31T07:12:14.0230437Z test_nn.py::TestNNDeviceTypeCPU::test_module_to_empty_cpu_float32 PASSED [ 94%] 2023-03-31T07:12:14.0230593Z test_nn.py::TestNNDeviceTypeCPU::test_module_to_empty_cpu_float64 PASSED [ 94%] 2023-03-31T07:12:14.0230764Z test_nn.py::TestNNDeviceTypeCPU::test_multi_margin_loss_errors_cpu PASSED [ 94%] 2023-03-31T07:12:14.0230923Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_all_ignored_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231083Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_byte_target_matches_long_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231262Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_mean_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231437Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_none_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231610Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_empty_tensor_reduction_sum_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231776Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_invalid_target_dim_cpu PASSED [ 94%] 2023-03-31T07:12:14.0231943Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_invalid_weights_cpu PASSED [ 94%] 2023-03-31T07:12:14.0232174Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_mean_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:12:14.0232403Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_none_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:12:14.0232602Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_large_tensor_reduction_sum_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:12:14.0232754Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_mismatched_batch_cpu PASSED [ 94%] 2023-03-31T07:12:14.0232926Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_out_of_bounds_ignore_index_cpu PASSED [ 94%] 2023-03-31T07:12:14.0233091Z test_nn.py::TestNNDeviceTypeCPU::test_nll_loss_total_weight_is_zero_cpu PASSED [ 94%] 2023-03-31T07:12:14.0233242Z test_nn.py::TestNNDeviceTypeCPU::test_nn_empty_cpu PASSED [ 94%] 2023-03-31T07:12:14.0233394Z test_nn.py::TestNNDeviceTypeCPU::test_nn_scalars_cpu PASSED [ 94%] 2023-03-31T07:12:14.0233552Z test_nn.py::TestNNDeviceTypeCPU::test_nn_scalars_reductions_cpu PASSED [ 94%] 2023-03-31T07:12:14.0233724Z test_nn.py::TestNNDeviceTypeCPU::test_nonlinearity_propagate_nan_cpu PASSED [ 95%] 2023-03-31T07:12:14.0233874Z test_nn.py::TestNNDeviceTypeCPU::test_one_hot_cpu PASSED [ 95%] 2023-03-31T07:12:14.0234079Z test_nn.py::TestNNDeviceTypeCPU::test_overwrite_module_params_on_conversion_cpu_device_cpu SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0234270Z test_nn.py::TestNNDeviceTypeCPU::test_pad_cpu_complex128 PASSED [ 95%] 2023-03-31T07:12:14.0234421Z test_nn.py::TestNNDeviceTypeCPU::test_pad_cpu_float64 PASSED [ 95%] 2023-03-31T07:12:14.0234613Z test_nn.py::TestNNDeviceTypeCPU::test_prelu_backward_32bit_indexing_cpu SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0234792Z test_nn.py::TestNNDeviceTypeCPU::test_rnn_fused_cpu_float32 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0234967Z test_nn.py::TestNNDeviceTypeCPU::test_rnn_fused_cpu_float64 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0235138Z test_nn.py::TestNNDeviceTypeCPU::test_rnn_retain_variables_cpu_float64 PASSED [ 95%] 2023-03-31T07:12:14.0235303Z test_nn.py::TestNNDeviceTypeCPU::test_save_lstm_compatibility_cpu PASSED [ 95%] 2023-03-31T07:12:14.0235461Z test_nn.py::TestNNDeviceTypeCPU::test_silu_inplace_overlap_cpu PASSED [ 95%] 2023-03-31T07:12:14.0235666Z test_nn.py::TestNNDeviceTypeCPU::test_skip_init_cpu PASSED [ 95%] 2023-03-31T07:12:14.0235825Z test_nn.py::TestNNDeviceTypeCPU::test_smooth_l1_loss_bfloat16_cpu PASSED [ 95%] 2023-03-31T07:12:14.0235992Z test_nn.py::TestNNDeviceTypeCPU::test_smooth_l1_loss_vs_huber_loss_cpu PASSED [ 95%] 2023-03-31T07:12:14.0236163Z test_nn.py::TestNNDeviceTypeCPU::test_smoothl1loss_backward_zero_beta_cpu PASSED [ 95%] 2023-03-31T07:12:14.0236354Z test_nn.py::TestNNDeviceTypeCPU::test_softmax_64bit_indexing_cpu_float16 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0236510Z test_nn.py::TestNNDeviceTypeCPU::test_softmax_bfloat16_cpu PASSED [ 95%] 2023-03-31T07:12:14.0236685Z test_nn.py::TestNNDeviceTypeCPU::test_softmax_cpu_float16 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0236860Z test_nn.py::TestNNDeviceTypeCPU::test_softmax_cpu_float32 SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:12:14.0237009Z test_nn.py::TestNNDeviceTypeCPU::test_softmax_results_cpu_float32 PASSED [ 95%] 2023-03-31T07:12:14.0237174Z test_nn.py::TestNNDeviceTypeCPU::test_softplus_inplace_overlap_cpu PASSED [ 95%] 2023-03-31T07:12:14.0237335Z test_nn.py::TestNNDeviceTypeCPU::test_softplus_low_threshold_cpu PASSED [ 96%] 2023-03-31T07:12:14.0237483Z test_nn.py::TestNNDeviceTypeCPU::test_softshrink_cpu PASSED [ 96%] 2023-03-31T07:12:14.0237653Z test_nn.py::TestNNDeviceTypeCPU::test_softshrink_inplace_overlap_cpu PASSED [ 96%] 2023-03-31T07:12:14.0237811Z test_nn.py::TestNNDeviceTypeCPU::test_softshrink_negative_cpu PASSED [ 96%] 2023-03-31T07:12:14.0237977Z test_nn.py::TestNNDeviceTypeCPU::test_threshold_inplace_overlap_cpu PASSED [ 96%] 2023-03-31T07:12:14.0238166Z test_nn.py::TestNNDeviceTypeCPU::test_to_complex_cpu PASSED [ 96%] 2023-03-31T07:12:14.0238362Z test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_cpu_float32 PASSED [ 96%] 2023-03-31T07:12:14.0238556Z test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_fast_path_cpu_float64 PASSED [ 96%] 2023-03-31T07:12:14.0238746Z test_nn.py::TestNNDeviceTypeCPU::test_transformerencoderlayer_gelu_cpu_float32 PASSED [ 96%] 2023-03-31T07:12:14.0238922Z test_nn.py::TestNNDeviceTypeCPU::test_triplet_margin_with_distance_loss_cpu PASSED [ 96%] 2023-03-31T07:12:14.0239120Z test_nn.py::TestNNDeviceTypeCPU::test_triplet_margin_with_distance_loss_default_parity_cpu PASSED [ 96%] 2023-03-31T07:12:14.0239413Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_3_output_size_32_cpu PASSED [ 96%] 2023-03-31T07:12:14.0239708Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_3_output_size_600_cpu PASSED [ 96%] 2023-03-31T07:12:14.0240001Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_5_output_size_32_cpu PASSED [ 96%] 2023-03-31T07:12:14.0240320Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_False_num_channels_5_output_size_600_cpu PASSED [ 96%] 2023-03-31T07:12:14.0240611Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_3_output_size_32_cpu PASSED [ 96%] 2023-03-31T07:12:14.0240897Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_3_output_size_600_cpu PASSED [ 96%] 2023-03-31T07:12:14.0241168Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_5_output_size_32_cpu PASSED [ 96%] 2023-03-31T07:12:14.0241460Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_False_align_corners_True_num_channels_5_output_size_600_cpu PASSED [ 96%] 2023-03-31T07:12:14.0241779Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_3_output_size_32_cpu PASSED [ 96%] 2023-03-31T07:12:14.0242064Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_3_output_size_600_cpu PASSED [ 96%] 2023-03-31T07:12:14.0242349Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0242637Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_False_num_channels_5_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0242923Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_3_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0243379Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_3_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0243664Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0243996Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_channels_last_antialias_True_align_corners_True_num_channels_5_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0244320Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_3_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0244618Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_3_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0244910Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0245198Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_False_num_channels_5_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0245487Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_3_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0245761Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_3_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0246096Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0246384Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_False_align_corners_True_num_channels_5_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0246672Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_3_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0246962Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_3_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0247305Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0247591Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_False_num_channels_5_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0247877Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_3_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0248163Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_3_output_size_600_cpu PASSED [ 97%] 2023-03-31T07:12:14.0248452Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_5_output_size_32_cpu PASSED [ 97%] 2023-03-31T07:12:14.0248738Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiLinear2d_consistency_memory_format_torch_contiguous_format_antialias_True_align_corners_True_num_channels_5_output_size_600_cpu PASSED [ 98%] 2023-03-31T07:12:14.0248994Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bicubic_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0249258Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bicubic_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0249548Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bilinear_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0249826Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_False_mode_bilinear_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0250083Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bicubic_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0250343Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bicubic_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0250595Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bilinear_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0250853Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_False_align_corners_True_mode_bilinear_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0251106Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bicubic_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0251365Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bicubic_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0251656Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bilinear_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0251913Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_False_mode_bilinear_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0252166Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bicubic_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0252427Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bicubic_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0252665Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bilinear_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0252956Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBiMode2d_antialias_True_align_corners_True_mode_bilinear_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0253183Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_aa_correctness_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0253409Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_aa_correctness_memory_format_torch_contiguous_format_cpu PASSED [ 98%] 2023-03-31T07:12:14.0253589Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBicubic2d_correctness_cpu PASSED [ 98%] 2023-03-31T07:12:14.0253818Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBilinear2d_aa_correctness_memory_format_torch_channels_last_cpu PASSED [ 98%] 2023-03-31T07:12:14.0254048Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingBilinear2d_aa_correctness_memory_format_torch_contiguous_format_cpu PASSED [ 99%] 2023-03-31T07:12:14.0254235Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0254402Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_cpu PASSED [ 99%] 2023-03-31T07:12:14.0254609Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest1d_launch_config_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0254776Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0254943Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_cpu PASSED [ 99%] 2023-03-31T07:12:14.0255146Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_config_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0255378Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_fail_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0255607Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest2d_launch_rocm_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0255786Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0255953Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_cpu PASSED [ 99%] 2023-03-31T07:12:14.0256158Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearest3d_launch_config_cpu SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0256351Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact1d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0256523Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact1d_rescale_cpu PASSED [ 99%] 2023-03-31T07:12:14.0256710Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact2d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0256896Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingNearestExact3d_correctness_cpu PASSED [ 99%] 2023-03-31T07:12:14.0257091Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingTrilinear3d_align_corners_False_cpu PASSED [ 99%] 2023-03-31T07:12:14.0257286Z test_nn.py::TestNNDeviceTypeCPU::test_upsamplingTrilinear3d_align_corners_True_cpu PASSED [ 99%] 2023-03-31T07:12:14.0257497Z test_nn.py::TestNNDeviceTypeCPU::test_upsampling_64bit_indexing_channels_last_cpu_float16 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0257697Z test_nn.py::TestNNDeviceTypeCPU::test_variable_sequence_cpu_float32 PASSED [ 99%] 2023-03-31T07:12:14.0257898Z test_nn.py::TestNNDeviceTypeCPU::test_warp_softmax_64bit_indexing_cpu_float16 SKIPPED (Only runs on cuda) [ 99%] 2023-03-31T07:12:14.0258078Z test_nn.py::TestNNDeviceTypeCPU::test_warp_softmax_64bit_indexing_cpu_float32 SKIPPED (Only runs on cuda) [100%] 2023-03-31T07:12:14.0258103Z 2023-03-31T07:12:14.0258476Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nn/test_nn-2e6c9c8f712a39c2.xml - 2023-03-31T07:12:14.0258611Z ========== 1065 passed, 1040 skipped, 3 xfailed in 115.88s (0:01:55) =========== 2023-03-31T07:12:14.0258794Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:12:14.0258838Z 2023-03-31T07:12:14.0259208Z ##[endgroup] 2023-03-31T07:12:14.0259489Z FINISHED PRINTING LOG FILE of test_nn (/var/lib/jenkins/workspace/test/test-reports/test_nn_7anrhld6.log) 2023-03-31T07:12:14.0259498Z 2023-03-31T07:12:14.0259688Z Running test_cpp_extensions_jit ... [2023-03-31 07:12:13.939094] 2023-03-31T07:12:14.0260186Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cpp_extensions_jit.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:12:13.939462] 2023-03-31T07:13:01.5106147Z 2023-03-31T07:13:01.5106630Z Expand the folded group to see the log file of test_cpp_extensions_jit 2023-03-31T07:13:01.5107425Z ##[group]PRINTING LOG FILE of test_cpp_extensions_jit (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_jit_dy4jhjqx.log) 2023-03-31T07:13:01.5110420Z Test results will be stored in test-reports/python-pytest/test_cpp_extensions_jit/test_cpp_extensions_jit-936fb117f18bad7f.xml 2023-03-31T07:13:01.5111099Z ============================= test session starts ============================== 2023-03-31T07:13:01.5111743Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:01.5112095Z cachedir: .pytest_cache 2023-03-31T07:13:01.5112625Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:01.5113011Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:01.5113544Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:01.5113860Z collecting ... collected 23 items 2023-03-31T07:13:01.5117132Z Running 23 items in this shard: test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_autograd_from_cpp, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_compilation_error_formatting, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_has_same_output_as_python, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_has_up_to_date_attributes, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_python_inter_op, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_python_inter_op_with_cuda, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_custom_compound_op_autograd, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_half_support, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_custom_op_cuda, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_cuda, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_multiple_sources_and_no_functions, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_throws_when_functions_is_bad, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_with_functions_as_dict, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_with_functions_as_list, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_compile_extension, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cuda_archflags, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cuda_extension, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cudnn_extension, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_lenient_flag_handling_in_jit_extensions, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_reload_jit_extension, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_returns_shared_library_path_when_is_python_module_is_true, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_set_default_type_also_changes_aten_default_type, test/test_cpp_extensions_jit.py::TestCppExtensionJIT::test_warning 2023-03-31T07:13:01.5120093Z 2023-03-31T07:13:01.5120312Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_autograd_from_cpp PASSED [ 4%] 2023-03-31T07:13:01.5120812Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_compilation_error_formatting PASSED [ 8%] 2023-03-31T07:13:01.5122811Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_has_same_output_as_python [1/2] c++ -MMD -MF cpp_frontend_extension.o.d -DTORCH_EXTENSION_NAME=cpp_frontend_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/workspace/test/cpp_extensions/cpp_frontend_extension.cpp -o cpp_frontend_extension.o 2023-03-31T07:13:01.5124539Z [2/2] c++ cpp_frontend_extension.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o cpp_frontend_extension.so 2023-03-31T07:13:01.5125180Z PASSED [ 13%] 2023-03-31T07:13:01.5125777Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_has_up_to_date_attributes PASSED [ 17%] 2023-03-31T07:13:01.5126198Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_python_inter_op PASSED [ 21%] 2023-03-31T07:13:01.5126810Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_cpp_frontend_module_python_inter_op_with_cuda SKIPPED (CUDA not found) [ 26%] 2023-03-31T07:13:01.5129294Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_custom_compound_op_autograd [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=is_python_module -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/is_python_module/main.cpp -o main.o 2023-03-31T07:13:01.5130720Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o is_python_module.so 2023-03-31T07:13:01.5131112Z PASSED [ 30%] 2023-03-31T07:13:01.5131423Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_half_support SKIPPED (Temporarily disabled) [ 34%] 2023-03-31T07:13:01.5131958Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_custom_op_cuda SKIPPED (Temporarily disabled) [ 39%] 2023-03-31T07:13:01.5132414Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_cuda SKIPPED (Temporarily disabled) [ 43%] 2023-03-31T07:13:01.5134429Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_multiple_sources_and_no_functions [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=inline_jit_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/inline_jit_extension/main.cpp -o main.o 2023-03-31T07:13:01.5135935Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o inline_jit_extension.so 2023-03-31T07:13:01.5136269Z PASSED [ 47%] 2023-03-31T07:13:01.5136666Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_throws_when_functions_is_bad PASSED [ 52%] 2023-03-31T07:13:01.5138593Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_with_functions_as_dict [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=inline_jit_extension_with_functions_dict -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/inline_jit_extension_with_functions_dict/main.cpp -o main.o 2023-03-31T07:13:01.5140032Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o inline_jit_extension_with_functions_dict.so 2023-03-31T07:13:01.5140451Z PASSED [ 56%] 2023-03-31T07:13:01.5142331Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_inline_jit_compile_extension_with_functions_as_list [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=inline_jit_extension_with_functions_list -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/inline_jit_extension_with_functions_list/main.cpp -o main.o 2023-03-31T07:13:01.5143916Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o inline_jit_extension_with_functions_list.so 2023-03-31T07:13:01.5144352Z PASSED [ 60%] 2023-03-31T07:13:01.5146057Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_compile_extension [1/3] c++ -MMD -MF jit_extension.o.d -DTORCH_EXTENSION_NAME=jit_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -I/var/lib/jenkins/workspace/test/cpp_extensions -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -g -c /var/lib/jenkins/workspace/test/cpp_extensions/jit_extension.cpp -o jit_extension.o 2023-03-31T07:13:01.5148506Z [2/3] c++ -MMD -MF jit_extension2.o.d -DTORCH_EXTENSION_NAME=jit_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -I/var/lib/jenkins/workspace/test/cpp_extensions -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -g -c /var/lib/jenkins/workspace/test/cpp_extensions/jit_extension2.cpp -o jit_extension2.o 2023-03-31T07:13:01.5149828Z [3/3] c++ jit_extension.o jit_extension2.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o jit_extension.so 2023-03-31T07:13:01.5150260Z PASSED [ 65%] 2023-03-31T07:13:01.5150572Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cuda_archflags SKIPPED (CUDA not found) [ 69%] 2023-03-31T07:13:01.5150965Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cuda_extension SKIPPED (CUDA not found) [ 73%] 2023-03-31T07:13:01.5151380Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_jit_cudnn_extension SKIPPED (CuDNN not found) [ 78%] 2023-03-31T07:13:01.5153247Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_lenient_flag_handling_in_jit_extensions [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=lenient_flag_handling_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -I/var/lib/jenkins/workspace/test/cpp_extensions -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -g -O0 -Wall -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/lenient_flag_handling_extension/main.cpp -o main.o 2023-03-31T07:13:01.5154639Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o lenient_flag_handling_extension.so 2023-03-31T07:13:01.5154981Z PASSED [ 82%] 2023-03-31T07:13:01.5156539Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_reload_jit_extension [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=reloaded_jit_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/reloaded_jit_extension/main.cpp -o main.o 2023-03-31T07:13:01.5157855Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o reloaded_jit_extension.so 2023-03-31T07:13:01.5159465Z [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=reloaded_jit_extension_v1 -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/reloaded_jit_extension/main.cpp -o main.o 2023-03-31T07:13:01.5160784Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o reloaded_jit_extension_v1.so 2023-03-31T07:13:01.5162743Z [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=reloaded_jit_extension_v2 -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/reloaded_jit_extension/main.cpp -o main.o 2023-03-31T07:13:01.5164140Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o reloaded_jit_extension_v2.so 2023-03-31T07:13:01.5164478Z PASSED [ 86%] 2023-03-31T07:13:01.5166670Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_returns_shared_library_path_when_is_python_module_is_true [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=is_python_module_v1 -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/is_python_module/main.cpp -o main.o 2023-03-31T07:13:01.5168085Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o is_python_module_v1.so 2023-03-31T07:13:01.5168457Z PASSED [ 91%] 2023-03-31T07:13:01.5170049Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_set_default_type_also_changes_aten_default_type [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=test_set_default_type -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py38_cpu/test_set_default_type/main.cpp -o main.o 2023-03-31T07:13:01.5171438Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o test_set_default_type.so 2023-03-31T07:13:01.5171753Z PASSED [ 95%] 2023-03-31T07:13:01.5172135Z test_cpp_extensions_jit.py::TestCppExtensionJIT::test_warning [W main.cpp:12] Warning: Error with CPUDoubleType (function foo) 2023-03-31T07:13:01.5172513Z [W main.cpp:12] Warning: Error with CPUDoubleType (function foo) 2023-03-31T07:13:01.5172798Z [W main.cpp:12] Warning: Error with CPUDoubleType (function foo) 2023-03-31T07:13:01.5173115Z [W main.cpp:12] Warning: Error with CPUDoubleType (function foo) 2023-03-31T07:13:01.5173497Z PASSED [100%]No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:13:01.5173674Z 2023-03-31T07:13:01.5173679Z 2023-03-31T07:13:01.5174047Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cpp_extensions_jit/test_cpp_extensions_jit-936fb117f18bad7f.xml - 2023-03-31T07:13:01.5174424Z ======================== 16 passed, 7 skipped in 43.38s ======================== 2023-03-31T07:13:01.5174751Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:01.5175002Z 2023-03-31T07:13:01.5175336Z ##[endgroup] 2023-03-31T07:13:01.5175765Z FINISHED PRINTING LOG FILE of test_cpp_extensions_jit (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_jit_dy4jhjqx.log) 2023-03-31T07:13:01.5176016Z 2023-03-31T07:13:01.5176187Z Running nn/test_pooling ... [2023-03-31 07:13:01.511057] 2023-03-31T07:13:01.5176805Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'nn/test_pooling.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:01.511426] 2023-03-31T07:13:27.8781853Z 2023-03-31T07:13:27.8782503Z Expand the folded group to see the log file of nn/test_pooling 2023-03-31T07:13:27.8783223Z ##[group]PRINTING LOG FILE of nn/test_pooling (/var/lib/jenkins/workspace/test/test-reports/nn-test_pooling_mzdpmx1r.log) 2023-03-31T07:13:27.8783969Z Test results will be stored in test-reports/python-pytest/nn.test_pooling/nn.test_pooling-29fb58834ee6cdf0.xml 2023-03-31T07:13:27.8784560Z ============================= test session starts ============================== 2023-03-31T07:13:27.8785285Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:27.8785794Z cachedir: .pytest_cache 2023-03-31T07:13:27.8786261Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:27.8786632Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:27.8787161Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:27.8798966Z collecting ... collected 90 items 2023-03-31T07:13:27.8809361Z Running 90 items in this shard: test/nn/test_pooling.py::TestAvgPool::test_avg_pool1d_ceil_mode, test/nn/test_pooling.py::TestAvgPool::test_avg_pool2d_ceil_mode, test/nn/test_pooling.py::TestAvgPool::test_avg_pool3d_ceil_mode, test/nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool2d, test/nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool2d_with_divisor, test/nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool3d, test/nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool3d_with_divisor, test/nn/test_pooling.py::TestPoolingNN::test_MaxUnpool2d_output_size, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_launch_config_backward, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_launch_config_forward, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_non_contiguous, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_bfloat16, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_size_none, test/nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_size_overflow, test/nn/test_pooling.py::TestPoolingNN::test_max_unpool, test/nn/test_pooling.py::TestPoolingNN::test_max_unpool2d_nhwc_cpu, test/nn/test_pooling.py::TestPoolingNN::test_max_unpool3d_input_check, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool1d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool2d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool3d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool_zero_batch_dim_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AvgPool2d_empty_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AvgPool3d_backward_after_cat_dim1_device_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_batch_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_out_size_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_samples_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_batch_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_out_size_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_samples_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool1d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool2d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool3d_indices_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool_zero_batch_dim_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case10_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case1_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case2_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case3_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case4_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case5_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case6_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case7_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case8_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case9_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_zero_batch_dim_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_avg_pool2d_output_size_one_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_avg_pool3d_output_size_one_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pool_odd_size_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_empty_output_size_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_empty_output_size_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_max_nhwc_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_max_nhwc_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int16, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int8, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_uint8, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_zero_batch_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_zero_batch_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_bfloat16_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_nhwc_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_nhwc_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool2d_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool3d_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool_nan_inf_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_corner_cases_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_corner_cases_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_bfloat16_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_indices_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_nhwc_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_nhwc_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float16, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool_nan_inf_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_maxpool3d_non_square_backward_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_maxpool_indices_no_batch_dim_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool3d_large_size_int64_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool3d_size_one_feature_dim_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool_invalid_size_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool_large_size_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_bfloat16_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_large_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_max_nhwc_cpu_float32, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_max_nhwc_cpu_float64, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_shape_cpu, test/nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_zero_stride_cpu 2023-03-31T07:13:27.8819073Z 2023-03-31T07:13:27.8819226Z nn/test_pooling.py::TestAvgPool::test_avg_pool1d_ceil_mode PASSED [ 1%] 2023-03-31T07:13:27.8819602Z nn/test_pooling.py::TestAvgPool::test_avg_pool2d_ceil_mode PASSED [ 2%] 2023-03-31T07:13:27.8819963Z nn/test_pooling.py::TestAvgPool::test_avg_pool3d_ceil_mode PASSED [ 3%] 2023-03-31T07:13:27.8820294Z nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool2d PASSED [ 4%] 2023-03-31T07:13:27.8820638Z nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool2d_with_divisor PASSED [ 5%] 2023-03-31T07:13:27.8820984Z nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool3d PASSED [ 6%] 2023-03-31T07:13:27.8821332Z nn/test_pooling.py::TestAvgPool::test_doubletensor_avg_pool3d_with_divisor PASSED [ 7%] 2023-03-31T07:13:27.8821681Z nn/test_pooling.py::TestPoolingNN::test_MaxUnpool2d_output_size PASSED [ 8%] 2023-03-31T07:13:27.8822006Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc PASSED [ 10%] 2023-03-31T07:13:27.8822398Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_launch_config_backward SKIPPED (CUDA unavailable) [ 11%] 2023-03-31T07:13:27.8822838Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_launch_config_forward SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:27.8823227Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_avg_nhwc_non_contiguous PASSED [ 13%] 2023-03-31T07:13:27.8823588Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_bfloat16 PASSED [ 14%] 2023-03-31T07:13:27.8824053Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_size_none PASSED [ 15%] 2023-03-31T07:13:27.8824408Z nn/test_pooling.py::TestPoolingNN::test_adaptive_pooling_size_overflow PASSED [ 16%] 2023-03-31T07:13:27.8824734Z nn/test_pooling.py::TestPoolingNN::test_max_unpool PASSED [ 17%] 2023-03-31T07:13:27.8825067Z nn/test_pooling.py::TestPoolingNN::test_max_unpool2d_nhwc_cpu PASSED [ 18%] 2023-03-31T07:13:27.8825403Z nn/test_pooling.py::TestPoolingNN::test_max_unpool3d_input_check PASSED [ 20%] 2023-03-31T07:13:27.8825779Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool1d_indices_cpu_float32 PASSED [ 21%] 2023-03-31T07:13:27.8826216Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool2d_indices_cpu_float32 PASSED [ 22%] 2023-03-31T07:13:27.8826645Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool3d_indices_cpu_float32 PASSED [ 23%] 2023-03-31T07:13:27.8827105Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AdaptiveMaxPool_zero_batch_dim_cpu PASSED [ 24%] 2023-03-31T07:13:27.8827493Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AvgPool2d_empty_cpu PASSED [ 25%] 2023-03-31T07:13:27.8827926Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_AvgPool3d_backward_after_cat_dim1_device_cpu SKIPPED (Only runs on cuda) [ 26%] 2023-03-31T07:13:27.8828376Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_batch_cpu PASSED [ 27%] 2023-03-31T07:13:27.8828806Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_out_size_cpu PASSED [ 28%] 2023-03-31T07:13:27.8829226Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool2d_zero_samples_cpu PASSED [ 30%] 2023-03-31T07:13:27.8829653Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_batch_cpu PASSED [ 31%] 2023-03-31T07:13:27.8830085Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_out_size_cpu PASSED [ 32%] 2023-03-31T07:13:27.8830500Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_FractionalMaxPool3d_zero_samples_cpu PASSED [ 33%] 2023-03-31T07:13:27.8830917Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool1d_indices_cpu_float32 PASSED [ 34%] 2023-03-31T07:13:27.8831317Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool2d_indices_cpu_float32 PASSED [ 35%] 2023-03-31T07:13:27.8831718Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool3d_indices_cpu_float32 PASSED [ 36%] 2023-03-31T07:13:27.8832099Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxPool_zero_batch_dim_cpu PASSED [ 37%] 2023-03-31T07:13:27.8832601Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case10_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 38%] 2023-03-31T07:13:27.8833183Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case1_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 40%] 2023-03-31T07:13:27.8833717Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case2_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 41%] 2023-03-31T07:13:27.8834231Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case3_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 42%] 2023-03-31T07:13:27.8834749Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case4_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 43%] 2023-03-31T07:13:27.8835280Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case5_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 44%] 2023-03-31T07:13:27.8835810Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case6_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 45%] 2023-03-31T07:13:27.8836365Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case7_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 46%] 2023-03-31T07:13:27.8836878Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case8_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 47%] 2023-03-31T07:13:27.8837401Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_index_errors_case9_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 48%] 2023-03-31T07:13:27.8837863Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_MaxUnpool_zero_batch_dim_cpu PASSED [ 50%] 2023-03-31T07:13:27.8838277Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_avg_pool2d_output_size_one_cpu PASSED [ 51%] 2023-03-31T07:13:27.8838693Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_avg_pool3d_output_size_one_cpu PASSED [ 52%] 2023-03-31T07:13:27.8839212Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pool_odd_size_cpu SKIPPED (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 53%] 2023-03-31T07:13:27.8839694Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_empty_output_size_cpu_float32 PASSED [ 54%] 2023-03-31T07:13:27.8840138Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_empty_output_size_cpu_float64 PASSED [ 55%] 2023-03-31T07:13:27.8840551Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_max_nhwc_cpu_float32 PASSED [ 56%] 2023-03-31T07:13:27.8840967Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_max_nhwc_cpu_float64 PASSED [ 57%] 2023-03-31T07:13:27.8841395Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int16 PASSED [ 58%] 2023-03-31T07:13:27.8841817Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int32 PASSED [ 60%] 2023-03-31T07:13:27.8842249Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int64 PASSED [ 61%] 2023-03-31T07:13:27.8842682Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_int8 PASSED [ 62%] 2023-03-31T07:13:27.8843273Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_no_suppot_input_cpu_uint8 PASSED [ 63%] 2023-03-31T07:13:27.8843689Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_zero_batch_cpu_float32 PASSED [ 64%] 2023-03-31T07:13:27.8844116Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_adaptive_pooling_zero_batch_cpu_float64 PASSED [ 65%] 2023-03-31T07:13:27.8844517Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_bfloat16_cpu PASSED [ 66%] 2023-03-31T07:13:27.8844987Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_nhwc_cpu_float32 PASSED [ 67%] 2023-03-31T07:13:27.8845406Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_avg_pool2d_nhwc_cpu_float64 PASSED [ 68%] 2023-03-31T07:13:27.8845801Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool2d_cpu PASSED [ 70%] 2023-03-31T07:13:27.8846198Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool3d_cpu PASSED [ 71%] 2023-03-31T07:13:27.8846606Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_fractional_max_pool_nan_inf_cpu_float32 PASSED [ 72%] 2023-03-31T07:13:27.8847007Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_corner_cases_cpu_float32 PASSED [ 73%] 2023-03-31T07:13:27.8847422Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_corner_cases_cpu_float64 PASSED [ 74%] 2023-03-31T07:13:27.8847817Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_cpu_float32 PASSED [ 75%] 2023-03-31T07:13:27.8848182Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool1d_cpu_float64 PASSED [ 76%] 2023-03-31T07:13:27.8848562Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_bfloat16_cpu PASSED [ 77%] 2023-03-31T07:13:27.8848960Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_cpu SKIPPED (Only runs on cuda) [ 78%] 2023-03-31T07:13:27.8849472Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_indices_cpu SKIPPED (Only runs on cuda) [ 80%] 2023-03-31T07:13:27.8849871Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_nhwc_cpu_float32 PASSED [ 81%] 2023-03-31T07:13:27.8850265Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool2d_nhwc_cpu_float64 PASSED [ 82%] 2023-03-31T07:13:27.8850683Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float16 SKIPPED (Only runs on cuda) [ 83%] 2023-03-31T07:13:27.8851117Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float32 SKIPPED (Only runs on cuda) [ 84%] 2023-03-31T07:13:27.8851537Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool3d_ndhwc_cpu_float64 SKIPPED (Only runs on cuda) [ 85%] 2023-03-31T07:13:27.8851949Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_max_pool_nan_inf_cpu_float32 PASSED [ 86%] 2023-03-31T07:13:27.8852406Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_maxpool3d_non_square_backward_cpu PASSED [ 87%] 2023-03-31T07:13:27.8852824Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_maxpool_indices_no_batch_dim_cpu_float32 PASSED [ 88%] 2023-03-31T07:13:27.8853240Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool3d_large_size_int64_cpu SKIPPED (Only runs on cuda) [ 90%] 2023-03-31T07:13:27.8853681Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool3d_size_one_feature_dim_cpu SKIPPED (Only runs on cuda) [ 91%] 2023-03-31T07:13:27.8854097Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool_invalid_size_cpu_float32 PASSED [ 92%] 2023-03-31T07:13:27.8854481Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pool_large_size_cpu_float32 PASSED [ 93%] 2023-03-31T07:13:27.8854888Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_bfloat16_cpu SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:13:27.8855308Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_large_cpu SKIPPED (Only runs on cuda) [ 95%] 2023-03-31T07:13:27.8855710Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_max_nhwc_cpu_float32 PASSED [ 96%] 2023-03-31T07:13:27.8856092Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_max_nhwc_cpu_float64 PASSED [ 97%] 2023-03-31T07:13:27.8856484Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_shape_cpu PASSED [ 98%] 2023-03-31T07:13:27.8856867Z nn/test_pooling.py::TestPoolingNNDeviceTypeCPU::test_pooling_zero_stride_cpu PASSED [100%] 2023-03-31T07:13:27.8857078Z 2023-03-31T07:13:27.8857543Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_pooling/nn.test_pooling-29fb58834ee6cdf0.xml - 2023-03-31T07:13:27.8857963Z ======================= 67 passed, 23 skipped in 24.20s ======================== 2023-03-31T07:13:27.8858328Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:27.8858535Z 2023-03-31T07:13:27.8858882Z ##[endgroup] 2023-03-31T07:13:27.8859290Z FINISHED PRINTING LOG FILE of nn/test_pooling (/var/lib/jenkins/workspace/test/test-reports/nn-test_pooling_mzdpmx1r.log) 2023-03-31T07:13:27.8859526Z 2023-03-31T07:13:27.8859709Z Running test_cpp_api_parity ... [2023-03-31 07:13:27.878640] 2023-03-31T07:13:27.8860348Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cpp_api_parity.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:27.878959] 2023-03-31T07:13:41.6318573Z 2023-03-31T07:13:41.6319241Z Expand the folded group to see the log file of test_cpp_api_parity 2023-03-31T07:13:41.6320392Z ##[group]PRINTING LOG FILE of test_cpp_api_parity (/var/lib/jenkins/workspace/test/test-reports/test_cpp_api_parity_c6t8ir31.log) 2023-03-31T07:13:41.6321262Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:13:41.6322131Z Test results will be stored in test-reports/python-pytest/test_cpp_api_parity/test_cpp_api_parity-e0456c88272dc2e6.xml 2023-03-31T07:13:41.6322963Z ============================= test session starts ============================== 2023-03-31T07:13:41.6323879Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:41.6324313Z cachedir: .pytest_cache 2023-03-31T07:13:41.6325013Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:41.6325581Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:41.6326257Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:41.6326868Z collecting ... collected 862 items 2023-03-31T07:13:41.6475331Z Running 862 items in this shard: test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride1_pad0_gpu_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride1_pad0_gpu_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_general_output, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_general_output_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride1_pad0_gpu_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride1_pad0_gpu_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_fixedkw_output, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_fixedkw_output_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_general_output, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_general_output_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_input_nooverlap, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_input_nooverlap_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_scalar_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_scalar_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_scalar_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_scalar_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_not_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_not_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_simple_average, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_simple_average_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_2d_simple_average, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_2d_simple_average_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_momentum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_momentum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_3d_simple_average, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_3d_simple_average_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_momentum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_momentum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_int_target_lengths_tensors, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_int_target_lengths_tensors_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_lengths_tensors, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_lengths_tensors_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_lengths_tensors, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_lengths_tensors_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_circular_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_circular_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_groups, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_groups_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1size1, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1size1_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2size1, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2size1_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_valid, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_valid_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_reflect_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_reflect_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_replicate_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_replicate_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zeros_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zeros_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_circular_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_circular_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_padded, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_padded_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_strided, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_strided_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_with_multiplier, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_with_multiplier_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_thnn, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_thnn_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_valid, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_valid_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_padding, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_padding_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_reflect_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_reflect_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_replicate_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_replicate_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_strided, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_strided_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zeros_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zeros_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_1x1x1_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_1x1x1_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_circular_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_circular_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_strided, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_strided_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_groups, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_groups_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_valid, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_valid_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_replicate_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_replicate_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_padding, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_padding_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zero_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zero_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zeros_stride2_pad2, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zeros_stride2_pad2_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_groups, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_groups_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_groups, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_groups_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_dilated, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_dilated_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_margin, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_margin_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_dim_is_3, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_dim_is_3_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_higher_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_higher_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossMapLRN2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossMapLRN2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_discontiguous, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_discontiguous_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_padding_idx, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_padding_idx_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_padding_idx, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_padding_idx_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sparse, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sparse_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_padding_idx, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_padding_idx_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_discontiguous, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_discontiguous_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_sparse, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_sparse_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_int_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_int_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_int_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_int_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_asymsize, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_asymsize_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_GN, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_GN_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_large_batch, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_large_batch_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_IN, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_IN_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_LN, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_LN_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_large_feature, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_large_feature_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_IN, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_IN_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_LN, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_LN_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_large_feature, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_large_feature_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_margin, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_margin_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_scalar_margin, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_scalar_margin_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HuberLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HuberLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_log_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_log_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_log_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_log_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_norm, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_norm_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_norm, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_norm_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_elementwise_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_elementwise_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_empty_elementwise_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_empty_elementwise_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_no_elementwise_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_no_elementwise_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_elementwise_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_elementwise_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_affine_large_feature, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_affine_large_feature_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_elementwise_affine, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_elementwise_affine_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_bias, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_bias_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_2d_uneven_pad, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_2d_uneven_pad_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_3d_custom_params, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_3d_custom_params_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_prec, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_prec_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_margin, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_margin_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_3d_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_3d_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_4d_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_4d_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_padding, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_padding_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_margin, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_margin_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_p, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_p_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_dim_is_3, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_dim_is_3_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_higher_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_higher_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_neg, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_neg_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_lhs, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_lhs_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_rhs, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_rhs_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_with_non_default_args, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_with_non_default_args_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelShuffle, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelShuffle_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelUnshuffle, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelUnshuffle_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_no_log_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_no_log_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_no_log_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_no_log_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_complex, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_complex_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_has_parity, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_has_parity_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_no_parity, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_no_parity_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_gelu_activation, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_gelu_activation_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_relu_activation, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_relu_activation_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_gelu_activation, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_gelu_activation_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_relu_activation, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_relu_activation_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Transformer_multilayer_coder, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Transformer_multilayer_coder_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_mean, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_mean_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_none, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_none_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_sum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_sum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unflatten_no_batch_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unflatten_no_batch_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_int_input, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_int_input_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_legacy_enum, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_legacy_enum_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_margin_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_margin_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HuberLoss_delta, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HuberLoss_delta_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_log_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_log_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_log_target, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_log_target_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_log_target_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_log_target_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_target_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_target_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_complex, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_complex_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_0d_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_0d_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_1d_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_1d_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_index_neg, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_index_neg_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_weights_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_weights_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_1d_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_1d_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_margin_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_margin_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_p_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_p_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_weights_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_weights_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_neg, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_neg_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding122112_3dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding122112_3dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding1221_2dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding1221_2dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding12_1dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding12_1dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding2322_2dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding2322_2dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding31_1dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding31_1dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding322112_3dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding322112_3dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding332122_3dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding332122_3dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding3331_2dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding3331_2dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding33_1dcircular, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding33_1dcircular_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_PoissonNLLLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_PoissonNLLLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_beta, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_beta_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_zero_beta, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_zero_beta_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SoftMarginLoss_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SoftMarginLoss_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_shared_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_shared_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_shared_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_shared_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_tuple_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_tuple_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_launch_configs, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_launch_configs_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_1d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_1d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_2d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_2d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_zero_dim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_zero_dim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_align_corners, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_align_corners_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim0, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim0_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim3, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim3_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_lastdim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_lastdim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_special, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_special_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_multimarginloss_1d_input_0d_target_no_reduce, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_multimarginloss_1d_input_0d_target_no_reduce_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_has_parity, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_has_parity_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_no_parity, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_no_parity_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim0, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim0_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim3, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim3_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_scalar, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_scalar_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_dtype, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_dtype_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_dtype, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_dtype_cuda, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_special, test/test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_special_cuda 2023-03-31T07:13:41.6571365Z 2023-03-31T07:13:41.6572467Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 0%] 2023-03-31T07:13:41.6573067Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 0%] 2023-03-31T07:13:41.6573677Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 0%] 2023-03-31T07:13:41.6574305Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 0%] 2023-03-31T07:13:41.6574898Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 0%] 2023-03-31T07:13:41.6575491Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 0%] 2023-03-31T07:13:41.6576127Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 0%] 2023-03-31T07:13:41.6576742Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool1d_stride_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 0%] 2023-03-31T07:13:41.6577316Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 1%] 2023-03-31T07:13:41.6577889Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 1%] 2023-03-31T07:13:41.6578472Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor <- test/cpp_api_parity/module_impl_check.py PASSED [ 1%] 2023-03-31T07:13:41.6579072Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 1%] 2023-03-31T07:13:41.6579672Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 1%] 2023-03-31T07:13:41.6580280Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 1%] 2023-03-31T07:13:41.6581139Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 1%] 2023-03-31T07:13:41.6581862Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_divisor_stride_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 1%] 2023-03-31T07:13:41.6582468Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 1%] 2023-03-31T07:13:41.6583160Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 2%] 2023-03-31T07:13:41.6583743Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 2%] 2023-03-31T07:13:41.6584344Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 2%] 2023-03-31T07:13:41.6584938Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 2%] 2023-03-31T07:13:41.6585552Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool2d_stride_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 2%] 2023-03-31T07:13:41.6586175Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 2%] 2023-03-31T07:13:41.6586755Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 2%] 2023-03-31T07:13:41.6587405Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor <- test/cpp_api_parity/module_impl_check.py PASSED [ 2%] 2023-03-31T07:13:41.6588009Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 3%] 2023-03-31T07:13:41.6588588Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 3%] 2023-03-31T07:13:41.6589200Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride1_pad0_gpu_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 3%] 2023-03-31T07:13:41.6589861Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride1_pad0_gpu_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 3%] 2023-03-31T07:13:41.6590516Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 3%] 2023-03-31T07:13:41.6591115Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 3%] 2023-03-31T07:13:41.6591747Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 3%] 2023-03-31T07:13:41.6592391Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output <- test/cpp_api_parity/module_impl_check.py PASSED [ 3%] 2023-03-31T07:13:41.6593073Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_fixedkw_output_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 3%] 2023-03-31T07:13:41.6593743Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_general_output <- test/cpp_api_parity/module_impl_check.py PASSED [ 4%] 2023-03-31T07:13:41.6594408Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_general_output_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 4%] 2023-03-31T07:13:41.6595107Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap <- test/cpp_api_parity/module_impl_check.py PASSED [ 4%] 2023-03-31T07:13:41.6595821Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_divisor_stride_pad_gpu_input_nooverlap_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 4%] 2023-03-31T07:13:41.6596449Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 4%] 2023-03-31T07:13:41.6597043Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 4%] 2023-03-31T07:13:41.6597633Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 4%] 2023-03-31T07:13:41.6598217Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride1_pad0_gpu_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 4%] 2023-03-31T07:13:41.6598863Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride1_pad0_gpu_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 4%] 2023-03-31T07:13:41.6599500Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 5%] 2023-03-31T07:13:41.6600107Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 5%] 2023-03-31T07:13:41.6600713Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 5%] 2023-03-31T07:13:41.6601383Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_fixedkw_output <- test/cpp_api_parity/module_impl_check.py PASSED [ 5%] 2023-03-31T07:13:41.6602051Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_fixedkw_output_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 5%] 2023-03-31T07:13:41.6602685Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_general_output <- test/cpp_api_parity/module_impl_check.py PASSED [ 5%] 2023-03-31T07:13:41.6604988Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_general_output_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 5%] 2023-03-31T07:13:41.6605637Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_input_nooverlap <- test/cpp_api_parity/module_impl_check.py PASSED [ 5%] 2023-03-31T07:13:41.6606299Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_AvgPool3d_stride_pad_gpu_input_nooverlap_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 6%] 2023-03-31T07:13:41.6606875Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 6%] 2023-03-31T07:13:41.6607455Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 6%] 2023-03-31T07:13:41.6608039Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py XFAIL [ 6%] 2023-03-31T07:13:41.6608654Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 6%] 2023-03-31T07:13:41.6609245Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py XFAIL [ 6%] 2023-03-31T07:13:41.6609845Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 6%] 2023-03-31T07:13:41.6610441Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py XFAIL [ 6%] 2023-03-31T07:13:41.6611147Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 6%] 2023-03-31T07:13:41.6611791Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_scalar_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 7%] 2023-03-31T07:13:41.6612390Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_scalar_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 7%] 2023-03-31T07:13:41.6612970Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 7%] 2023-03-31T07:13:41.6613567Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCELoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 7%] 2023-03-31T07:13:41.6614158Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 7%] 2023-03-31T07:13:41.6614758Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 7%] 2023-03-31T07:13:41.6615381Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py XFAIL [ 7%] 2023-03-31T07:13:41.6616082Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 7%] 2023-03-31T07:13:41.6616757Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py XFAIL [ 8%] 2023-03-31T07:13:41.6617389Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 8%] 2023-03-31T07:13:41.6618019Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py XFAIL [ 8%] 2023-03-31T07:13:41.6618664Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 8%] 2023-03-31T07:13:41.6619298Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_scalar_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 8%] 2023-03-31T07:13:41.6619950Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_scalar_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 8%] 2023-03-31T07:13:41.6620558Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 8%] 2023-03-31T07:13:41.6621199Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BCEWithLogitsLoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 8%] 2023-03-31T07:13:41.6621803Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 8%] 2023-03-31T07:13:41.6622415Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 9%] 2023-03-31T07:13:41.6623082Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_not_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 9%] 2023-03-31T07:13:41.6623718Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_3d_input_not_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 9%] 2023-03-31T07:13:41.6624318Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 9%] 2023-03-31T07:13:41.6624955Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 9%] 2023-03-31T07:13:41.6625580Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_simple_average <- test/cpp_api_parity/module_impl_check.py PASSED [ 9%] 2023-03-31T07:13:41.6626226Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_affine_simple_average_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 9%] 2023-03-31T07:13:41.6626829Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 9%] 2023-03-31T07:13:41.6627449Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 9%] 2023-03-31T07:13:41.6628057Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 10%] 2023-03-31T07:13:41.6628973Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_not_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 10%] 2023-03-31T07:13:41.6629599Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 10%] 2023-03-31T07:13:41.6630266Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm1d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 10%] 2023-03-31T07:13:41.6630896Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 10%] 2023-03-31T07:13:41.6631457Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_2d_simple_average <- test/cpp_api_parity/module_impl_check.py PASSED [ 10%] 2023-03-31T07:13:41.6632136Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_2d_simple_average_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 10%] 2023-03-31T07:13:41.6632926Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 10%] 2023-03-31T07:13:41.6633510Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_momentum <- test/cpp_api_parity/module_impl_check.py PASSED [ 11%] 2023-03-31T07:13:41.6634114Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_momentum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 11%] 2023-03-31T07:13:41.6634709Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 11%] 2023-03-31T07:13:41.6635320Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 11%] 2023-03-31T07:13:41.6635934Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 11%] 2023-03-31T07:13:41.6636558Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_not_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 11%] 2023-03-31T07:13:41.6637165Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 11%] 2023-03-31T07:13:41.6638158Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm2d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 11%] 2023-03-31T07:13:41.6639145Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 11%] 2023-03-31T07:13:41.6640195Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_3d_simple_average <- test/cpp_api_parity/module_impl_check.py PASSED [ 12%] 2023-03-31T07:13:41.6641451Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_3d_simple_average_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:41.6642635Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:41.6643868Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_momentum <- test/cpp_api_parity/module_impl_check.py PASSED [ 12%] 2023-03-31T07:13:41.6644970Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_momentum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:41.6646064Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 12%] 2023-03-31T07:13:41.6647175Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:41.6648306Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 12%] 2023-03-31T07:13:41.6649470Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_not_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 12%] 2023-03-31T07:13:41.6650677Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 13%] 2023-03-31T07:13:41.6651885Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_BatchNorm3d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 13%] 2023-03-31T07:13:41.6653015Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_int_target_lengths_tensors <- test/cpp_api_parity/module_impl_check.py PASSED [ 13%] 2023-03-31T07:13:41.6654215Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_int_target_lengths_tensors_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 13%] 2023-03-31T07:13:41.6655319Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_lengths_tensors <- test/cpp_api_parity/module_impl_check.py PASSED [ 13%] 2023-03-31T07:13:41.6656466Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_2d_lengths_tensors_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 13%] 2023-03-31T07:13:41.6657535Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_lengths_tensors <- test/cpp_api_parity/module_impl_check.py PASSED [ 13%] 2023-03-31T07:13:41.6658676Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CTCLoss_lengths_tensors_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 13%] 2023-03-31T07:13:41.6659714Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 14%] 2023-03-31T07:13:41.6660728Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_circular_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 14%] 2023-03-31T07:13:41.6661889Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_circular_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 14%] 2023-03-31T07:13:41.6663101Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 14%] 2023-03-31T07:13:41.6664142Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 14%] 2023-03-31T07:13:41.6665187Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 14%] 2023-03-31T07:13:41.6666229Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_groups <- test/cpp_api_parity/module_impl_check.py PASSED [ 14%] 2023-03-31T07:13:41.6667457Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_groups_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 14%] 2023-03-31T07:13:41.6668556Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1 <- test/cpp_api_parity/module_impl_check.py PASSED [ 14%] 2023-03-31T07:13:41.6669604Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 15%] 2023-03-31T07:13:41.6670636Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1size1 <- test/cpp_api_parity/module_impl_check.py PASSED [ 15%] 2023-03-31T07:13:41.6671654Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad1size1_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 15%] 2023-03-31T07:13:41.6672667Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 15%] 2023-03-31T07:13:41.6673679Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 15%] 2023-03-31T07:13:41.6674701Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2size1 <- test/cpp_api_parity/module_impl_check.py PASSED [ 15%] 2023-03-31T07:13:41.6675821Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad2size1_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 15%] 2023-03-31T07:13:41.6676843Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same <- test/cpp_api_parity/module_impl_check.py PASSED [ 15%] 2023-03-31T07:13:41.6677898Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 16%] 2023-03-31T07:13:41.6678986Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 16%] 2023-03-31T07:13:41.6680081Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 16%] 2023-03-31T07:13:41.6681191Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 16%] 2023-03-31T07:13:41.6682305Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_same_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 16%] 2023-03-31T07:13:41.6683557Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_valid <- test/cpp_api_parity/module_impl_check.py PASSED [ 16%] 2023-03-31T07:13:41.6684652Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_pad_valid_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 16%] 2023-03-31T07:13:41.6685702Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_reflect_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 16%] 2023-03-31T07:13:41.6686861Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_reflect_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 16%] 2023-03-31T07:13:41.6687988Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_replicate_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 17%] 2023-03-31T07:13:41.6689160Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_replicate_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 17%] 2023-03-31T07:13:41.6690209Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 17%] 2023-03-31T07:13:41.6691279Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 17%] 2023-03-31T07:13:41.6692492Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 17%] 2023-03-31T07:13:41.6693677Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 17%] 2023-03-31T07:13:41.6694794Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zeros_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 17%] 2023-03-31T07:13:41.6695957Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv1d_zeros_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 17%] 2023-03-31T07:13:41.6697024Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 17%] 2023-03-31T07:13:41.6698089Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_circular_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 18%] 2023-03-31T07:13:41.6699240Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_circular_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 18%] 2023-03-31T07:13:41.6700387Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 18%] 2023-03-31T07:13:41.6701556Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise <- test/cpp_api_parity/module_impl_check.py PASSED [ 18%] 2023-03-31T07:13:41.6702585Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 18%] 2023-03-31T07:13:41.6703717Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 18%] 2023-03-31T07:13:41.6704848Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 18%] 2023-03-31T07:13:41.6705856Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_padded <- test/cpp_api_parity/module_impl_check.py PASSED [ 18%] 2023-03-31T07:13:41.6706887Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_padded_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 19%] 2023-03-31T07:13:41.6707970Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_strided <- test/cpp_api_parity/module_impl_check.py PASSED [ 19%] 2023-03-31T07:13:41.6709092Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_strided_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 19%] 2023-03-31T07:13:41.6710207Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_with_multiplier <- test/cpp_api_parity/module_impl_check.py PASSED [ 19%] 2023-03-31T07:13:41.6711343Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_depthwise_with_multiplier_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 19%] 2023-03-31T07:13:41.6712385Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 19%] 2023-03-31T07:13:41.6713463Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 19%] 2023-03-31T07:13:41.6714516Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups <- test/cpp_api_parity/module_impl_check.py PASSED [ 19%] 2023-03-31T07:13:41.6715564Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 19%] 2023-03-31T07:13:41.6716582Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_thnn <- test/cpp_api_parity/module_impl_check.py PASSED [ 20%] 2023-03-31T07:13:41.6717810Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_groups_thnn_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 20%] 2023-03-31T07:13:41.6718913Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 20%] 2023-03-31T07:13:41.6719971Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 20%] 2023-03-31T07:13:41.6721013Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same <- test/cpp_api_parity/module_impl_check.py PASSED [ 20%] 2023-03-31T07:13:41.6722048Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 20%] 2023-03-31T07:13:41.6723286Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 20%] 2023-03-31T07:13:41.6724383Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_same_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 20%] 2023-03-31T07:13:41.6725489Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_valid <- test/cpp_api_parity/module_impl_check.py PASSED [ 20%] 2023-03-31T07:13:41.6726611Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_pad_valid_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 21%] 2023-03-31T07:13:41.6727801Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_padding <- test/cpp_api_parity/module_impl_check.py PASSED [ 21%] 2023-03-31T07:13:41.6728936Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_padding_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 21%] 2023-03-31T07:13:41.6729968Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_reflect_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 21%] 2023-03-31T07:13:41.6731053Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_reflect_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 21%] 2023-03-31T07:13:41.6732162Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_replicate_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 21%] 2023-03-31T07:13:41.6733314Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_replicate_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 21%] 2023-03-31T07:13:41.6734340Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_strided <- test/cpp_api_parity/module_impl_check.py PASSED [ 21%] 2023-03-31T07:13:41.6735401Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_strided_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 22%] 2023-03-31T07:13:41.6736421Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 22%] 2023-03-31T07:13:41.6737492Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 22%] 2023-03-31T07:13:41.6738554Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zeros_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 22%] 2023-03-31T07:13:41.6739697Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv2d_zeros_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 22%] 2023-03-31T07:13:41.6740720Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 22%] 2023-03-31T07:13:41.6741717Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_1x1x1_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 22%] 2023-03-31T07:13:41.6742789Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_1x1x1_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 22%] 2023-03-31T07:13:41.6744171Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_circular_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 22%] 2023-03-31T07:13:41.6745353Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_circular_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 23%] 2023-03-31T07:13:41.6746436Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 23%] 2023-03-31T07:13:41.6747435Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 23%] 2023-03-31T07:13:41.6748512Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 23%] 2023-03-31T07:13:41.6749606Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_strided <- test/cpp_api_parity/module_impl_check.py PASSED [ 23%] 2023-03-31T07:13:41.6750701Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_dilated_strided_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 23%] 2023-03-31T07:13:41.6751766Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_groups <- test/cpp_api_parity/module_impl_check.py PASSED [ 23%] 2023-03-31T07:13:41.6752922Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_groups_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 23%] 2023-03-31T07:13:41.6753965Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 24%] 2023-03-31T07:13:41.6755104Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 24%] 2023-03-31T07:13:41.6756034Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same <- test/cpp_api_parity/module_impl_check.py PASSED [ 24%] 2023-03-31T07:13:41.6757047Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 24%] 2023-03-31T07:13:41.6758147Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 24%] 2023-03-31T07:13:41.6759300Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_same_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 24%] 2023-03-31T07:13:41.6760373Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_valid <- test/cpp_api_parity/module_impl_check.py PASSED [ 24%] 2023-03-31T07:13:41.6761415Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_pad_valid_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 24%] 2023-03-31T07:13:41.6762490Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_replicate_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 24%] 2023-03-31T07:13:41.6763801Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_replicate_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 25%] 2023-03-31T07:13:41.6764792Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 25%] 2023-03-31T07:13:41.6765835Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 25%] 2023-03-31T07:13:41.6766889Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_padding <- test/cpp_api_parity/module_impl_check.py PASSED [ 25%] 2023-03-31T07:13:41.6767984Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_stride_padding_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 25%] 2023-03-31T07:13:41.6769184Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zero_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 25%] 2023-03-31T07:13:41.6770283Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zero_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 25%] 2023-03-31T07:13:41.6771364Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zeros_stride2_pad2 <- test/cpp_api_parity/module_impl_check.py PASSED [ 25%] 2023-03-31T07:13:41.6772432Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Conv3d_zeros_stride2_pad2_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 25%] 2023-03-31T07:13:41.6773435Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 26%] 2023-03-31T07:13:41.6774470Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 26%] 2023-03-31T07:13:41.6775490Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 26%] 2023-03-31T07:13:41.6776628Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 26%] 2023-03-31T07:13:41.6777849Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_groups <- test/cpp_api_parity/module_impl_check.py PASSED [ 26%] 2023-03-31T07:13:41.6778976Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_groups_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 26%] 2023-03-31T07:13:41.6780200Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 26%] 2023-03-31T07:13:41.6781337Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose1d_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 26%] 2023-03-31T07:13:41.6782420Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 27%] 2023-03-31T07:13:41.6783690Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 27%] 2023-03-31T07:13:41.6784775Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 27%] 2023-03-31T07:13:41.6785875Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 27%] 2023-03-31T07:13:41.6786967Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_groups <- test/cpp_api_parity/module_impl_check.py PASSED [ 27%] 2023-03-31T07:13:41.6788023Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_groups_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 27%] 2023-03-31T07:13:41.6789086Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 27%] 2023-03-31T07:13:41.6790191Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose2d_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 27%] 2023-03-31T07:13:41.6791220Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 27%] 2023-03-31T07:13:41.6792308Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 28%] 2023-03-31T07:13:41.6793370Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_dilated <- test/cpp_api_parity/module_impl_check.py PASSED [ 28%] 2023-03-31T07:13:41.6794584Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ConvTranspose3d_dilated_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 28%] 2023-03-31T07:13:41.6795752Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 28%] 2023-03-31T07:13:41.6796883Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 28%] 2023-03-31T07:13:41.6797970Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_margin <- test/cpp_api_parity/module_impl_check.py PASSED [ 28%] 2023-03-31T07:13:41.6799120Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_margin_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 28%] 2023-03-31T07:13:41.6800275Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 28%] 2023-03-31T07:13:41.6801473Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 29%] 2023-03-31T07:13:41.6802623Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py PASSED [ 29%] 2023-03-31T07:13:41.6804095Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 29%] 2023-03-31T07:13:41.6805287Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 29%] 2023-03-31T07:13:41.6806651Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CosineEmbeddingLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 29%] 2023-03-31T07:13:41.6807792Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 29%] 2023-03-31T07:13:41.6808794Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 29%] 2023-03-31T07:13:41.6809895Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 29%] 2023-03-31T07:13:41.6811005Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 29%] 2023-03-31T07:13:41.6812138Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 30%] 2023-03-31T07:13:41.6813274Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing <- test/cpp_api_parity/module_impl_check.py PASSED [ 30%] 2023-03-31T07:13:41.6814393Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 30%] 2023-03-31T07:13:41.6815537Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 30%] 2023-03-31T07:13:41.6816799Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 30%] 2023-03-31T07:13:41.6818091Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction <- test/cpp_api_parity/module_impl_check.py PASSED [ 30%] 2023-03-31T07:13:41.6819392Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_sum_reduction_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 30%] 2023-03-31T07:13:41.6820728Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight <- test/cpp_api_parity/module_impl_check.py PASSED [ 30%] 2023-03-31T07:13:41.6822114Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_indices_target_smoothing_weight_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 30%] 2023-03-31T07:13:41.6823420Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target <- test/cpp_api_parity/module_impl_check.py PASSED [ 31%] 2023-03-31T07:13:41.6824569Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 31%] 2023-03-31T07:13:41.6825697Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing <- test/cpp_api_parity/module_impl_check.py PASSED [ 31%] 2023-03-31T07:13:41.6826965Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 31%] 2023-03-31T07:13:41.6828158Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction <- test/cpp_api_parity/module_impl_check.py PASSED [ 31%] 2023-03-31T07:13:41.6829599Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_sum_reduction_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 31%] 2023-03-31T07:13:41.6830942Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight <- test/cpp_api_parity/module_impl_check.py PASSED [ 31%] 2023-03-31T07:13:41.6832218Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_smoothing_weight_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 31%] 2023-03-31T07:13:41.6833477Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 32%] 2023-03-31T07:13:41.6834691Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_prob_target_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 32%] 2023-03-31T07:13:41.6835883Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 32%] 2023-03-31T07:13:41.6837038Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_2d_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 32%] 2023-03-31T07:13:41.6838206Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing <- test/cpp_api_parity/module_impl_check.py PASSED [ 32%] 2023-03-31T07:13:41.6839450Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 32%] 2023-03-31T07:13:41.6840661Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 32%] 2023-03-31T07:13:41.6841988Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 32%] 2023-03-31T07:13:41.6843405Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction <- test/cpp_api_parity/module_impl_check.py PASSED [ 32%] 2023-03-31T07:13:41.6844669Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 33%] 2023-03-31T07:13:41.6846080Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 33%] 2023-03-31T07:13:41.6847499Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_indices_target_smoothing_sum_reduction_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 33%] 2023-03-31T07:13:41.6848692Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target <- test/cpp_api_parity/module_impl_check.py PASSED [ 33%] 2023-03-31T07:13:41.6849798Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 33%] 2023-03-31T07:13:41.6850875Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing <- test/cpp_api_parity/module_impl_check.py PASSED [ 33%] 2023-03-31T07:13:41.6852061Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 33%] 2023-03-31T07:13:41.6853254Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction <- test/cpp_api_parity/module_impl_check.py PASSED [ 33%] 2023-03-31T07:13:41.6854593Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_smoothing_sum_reduction_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 33%] 2023-03-31T07:13:41.6855807Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 34%] 2023-03-31T07:13:41.6857136Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_3d_prob_target_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 34%] 2023-03-31T07:13:41.6858275Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target <- test/cpp_api_parity/module_impl_check.py PASSED [ 34%] 2023-03-31T07:13:41.6859404Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 34%] 2023-03-31T07:13:41.6860522Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 34%] 2023-03-31T07:13:41.6861745Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_4d_prob_target_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 34%] 2023-03-31T07:13:41.6862900Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 34%] 2023-03-31T07:13:41.6864068Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_dim_is_3 <- test/cpp_api_parity/module_impl_check.py PASSED [ 34%] 2023-03-31T07:13:41.6865200Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_dim_is_3_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 35%] 2023-03-31T07:13:41.6866335Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_higher_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 35%] 2023-03-31T07:13:41.6867472Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_higher_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 35%] 2023-03-31T07:13:41.6868620Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 35%] 2023-03-31T07:13:41.6869807Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossEntropyLoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 35%] 2023-03-31T07:13:41.6870973Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossMapLRN2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 35%] 2023-03-31T07:13:41.6872091Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_CrossMapLRN2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 35%] 2023-03-31T07:13:41.6873137Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding <- test/cpp_api_parity/module_impl_check.py PASSED [ 35%] 2023-03-31T07:13:41.6874189Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_discontiguous <- test/cpp_api_parity/module_impl_check.py PASSED [ 35%] 2023-03-31T07:13:41.6875363Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_discontiguous_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 36%] 2023-03-31T07:13:41.6876493Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max <- test/cpp_api_parity/module_impl_check.py PASSED [ 36%] 2023-03-31T07:13:41.6877624Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 36%] 2023-03-31T07:13:41.6878771Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_padding_idx <- test/cpp_api_parity/module_impl_check.py PASSED [ 36%] 2023-03-31T07:13:41.6880033Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_max_padding_idx_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 36%] 2023-03-31T07:13:41.6881153Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 36%] 2023-03-31T07:13:41.6882345Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 36%] 2023-03-31T07:13:41.6883944Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_padding_idx <- test/cpp_api_parity/module_impl_check.py PASSED [ 36%] 2023-03-31T07:13:41.6884697Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_mean_padding_idx_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 37%] 2023-03-31T07:13:41.6885314Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sparse <- test/cpp_api_parity/module_impl_check.py PASSED [ 37%] 2023-03-31T07:13:41.6885949Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sparse_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 37%] 2023-03-31T07:13:41.6886548Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 37%] 2023-03-31T07:13:41.6887151Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 37%] 2023-03-31T07:13:41.6887764Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_padding_idx <- test/cpp_api_parity/module_impl_check.py PASSED [ 37%] 2023-03-31T07:13:41.6888412Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_EmbeddingBag_sum_padding_idx_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 37%] 2023-03-31T07:13:41.6889053Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 37%] 2023-03-31T07:13:41.6889645Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_discontiguous <- test/cpp_api_parity/module_impl_check.py PASSED [ 37%] 2023-03-31T07:13:41.6890292Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_discontiguous_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 38%] 2023-03-31T07:13:41.6890908Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_sparse <- test/cpp_api_parity/module_impl_check.py PASSED [ 38%] 2023-03-31T07:13:41.6891687Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Embedding_sparse_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 38%] 2023-03-31T07:13:41.6892298Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten <- test/cpp_api_parity/module_impl_check.py PASSED [ 38%] 2023-03-31T07:13:41.6892884Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 38%] 2023-03-31T07:13:41.6893478Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 38%] 2023-03-31T07:13:41.6894103Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Flatten_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 38%] 2023-03-31T07:13:41.6894680Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold <- test/cpp_api_parity/module_impl_check.py PASSED [ 38%] 2023-03-31T07:13:41.6895249Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 38%] 2023-03-31T07:13:41.6895831Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_int_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 39%] 2023-03-31T07:13:41.6896436Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_int_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 39%] 2023-03-31T07:13:41.6897090Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 39%] 2023-03-31T07:13:41.6897709Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 39%] 2023-03-31T07:13:41.6898373Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_int_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 39%] 2023-03-31T07:13:41.6899007Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Fold_no_batch_dim_int_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 39%] 2023-03-31T07:13:41.6899644Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio <- test/cpp_api_parity/module_impl_check.py PASSED [ 39%] 2023-03-31T07:13:41.6900286Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 39%] 2023-03-31T07:13:41.6900948Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 40%] 2023-03-31T07:13:41.6901628Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 40%] 2023-03-31T07:13:41.6902321Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples <- test/cpp_api_parity/module_impl_check.py PASSED [ 40%] 2023-03-31T07:13:41.6903129Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_ratio_no_batch_dim_no_random_samples_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 40%] 2023-03-31T07:13:41.6903812Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size <- test/cpp_api_parity/module_impl_check.py PASSED [ 40%] 2023-03-31T07:13:41.6904476Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 40%] 2023-03-31T07:13:41.6905133Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 40%] 2023-03-31T07:13:41.6905832Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 40%] 2023-03-31T07:13:41.6906566Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples <- test/cpp_api_parity/module_impl_check.py PASSED [ 40%] 2023-03-31T07:13:41.6907351Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool2d_size_no_batch_dim_no_random_samples_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:13:41.6908033Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_asymsize <- test/cpp_api_parity/module_impl_check.py PASSED [ 41%] 2023-03-31T07:13:41.6908718Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_asymsize_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 41%] 2023-03-31T07:13:41.6909350Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio <- test/cpp_api_parity/module_impl_check.py PASSED [ 41%] 2023-03-31T07:13:41.6910009Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 41%] 2023-03-31T07:13:41.6910657Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 41%] 2023-03-31T07:13:41.6911373Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 41%] 2023-03-31T07:13:41.6912065Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples <- test/cpp_api_parity/module_impl_check.py PASSED [ 41%] 2023-03-31T07:13:41.6912842Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_ratio_no_batch_dim_no_random_samples_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 41%] 2023-03-31T07:13:41.6913514Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size <- test/cpp_api_parity/module_impl_check.py PASSED [ 42%] 2023-03-31T07:13:41.6914151Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 42%] 2023-03-31T07:13:41.6914797Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 42%] 2023-03-31T07:13:41.6915447Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 42%] 2023-03-31T07:13:41.6916145Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples <- test/cpp_api_parity/module_impl_check.py PASSED [ 42%] 2023-03-31T07:13:41.6916885Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_FractionalMaxPool3d_size_no_batch_dim_no_random_samples_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 42%] 2023-03-31T07:13:41.6917530Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 42%] 2023-03-31T07:13:41.6918108Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_GN <- test/cpp_api_parity/module_impl_check.py PASSED [ 42%] 2023-03-31T07:13:41.6918721Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_GN_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 43%] 2023-03-31T07:13:41.6919369Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 43%] 2023-03-31T07:13:41.6919982Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_large_batch <- test/cpp_api_parity/module_impl_check.py PASSED [ 43%] 2023-03-31T07:13:41.6920676Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_affine_large_batch_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 43%] 2023-03-31T07:13:41.6921305Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_IN <- test/cpp_api_parity/module_impl_check.py PASSED [ 43%] 2023-03-31T07:13:41.6921923Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_IN_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 43%] 2023-03-31T07:13:41.6922523Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_LN <- test/cpp_api_parity/module_impl_check.py PASSED [ 43%] 2023-03-31T07:13:41.6923449Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_1d_no_affine_LN_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 43%] 2023-03-31T07:13:41.6924039Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 43%] 2023-03-31T07:13:41.6924645Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 44%] 2023-03-31T07:13:41.6925259Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_large_feature <- test/cpp_api_parity/module_impl_check.py PASSED [ 44%] 2023-03-31T07:13:41.6925990Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_affine_large_feature_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 44%] 2023-03-31T07:13:41.6926599Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_IN <- test/cpp_api_parity/module_impl_check.py PASSED [ 44%] 2023-03-31T07:13:41.6927273Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_IN_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 44%] 2023-03-31T07:13:41.6927878Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_LN <- test/cpp_api_parity/module_impl_check.py PASSED [ 44%] 2023-03-31T07:13:41.6928502Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_LN_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 44%] 2023-03-31T07:13:41.6929128Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_large_feature <- test/cpp_api_parity/module_impl_check.py PASSED [ 44%] 2023-03-31T07:13:41.6929766Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_GroupNorm_2d_no_affine_large_feature_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 45%] 2023-03-31T07:13:41.6930382Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 45%] 2023-03-31T07:13:41.6931003Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 45%] 2023-03-31T07:13:41.6931623Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_margin <- test/cpp_api_parity/module_impl_check.py PASSED [ 45%] 2023-03-31T07:13:41.6932250Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_margin_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 45%] 2023-03-31T07:13:41.6932886Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py XFAIL [ 45%] 2023-03-31T07:13:41.6933544Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 45%] 2023-03-31T07:13:41.6934184Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py XFAIL [ 45%] 2023-03-31T07:13:41.6934916Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 45%] 2023-03-31T07:13:41.6935577Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py XFAIL [ 46%] 2023-03-31T07:13:41.6936231Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 46%] 2023-03-31T07:13:41.6936870Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_scalar_margin <- test/cpp_api_parity/module_impl_check.py PASSED [ 46%] 2023-03-31T07:13:41.6937529Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HingeEmbeddingLoss_scalar_margin_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 46%] 2023-03-31T07:13:41.6938107Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HuberLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 46%] 2023-03-31T07:13:41.6938691Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_HuberLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 46%] 2023-03-31T07:13:41.6939268Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 46%] 2023-03-31T07:13:41.6939912Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 46%] 2023-03-31T07:13:41.6940499Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 46%] 2023-03-31T07:13:41.6941162Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 47%] 2023-03-31T07:13:41.6941779Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 47%] 2023-03-31T07:13:41.6942418Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 47%] 2023-03-31T07:13:41.6943132Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 47%] 2023-03-31T07:13:41.6943794Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm1d_tracking_stats_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 47%] 2023-03-31T07:13:41.6944408Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 47%] 2023-03-31T07:13:41.6945002Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 47%] 2023-03-31T07:13:41.6945605Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 47%] 2023-03-31T07:13:41.6946223Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 48%] 2023-03-31T07:13:41.6946840Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 48%] 2023-03-31T07:13:41.6947475Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 48%] 2023-03-31T07:13:41.6948123Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 48%] 2023-03-31T07:13:41.6948780Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm2d_tracking_stats_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 48%] 2023-03-31T07:13:41.6949440Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 48%] 2023-03-31T07:13:41.6950067Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 48%] 2023-03-31T07:13:41.6950674Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 48%] 2023-03-31T07:13:41.6951288Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 48%] 2023-03-31T07:13:41.6951907Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats <- test/cpp_api_parity/module_impl_check.py PASSED [ 49%] 2023-03-31T07:13:41.6952548Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 49%] 2023-03-31T07:13:41.6953189Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 49%] 2023-03-31T07:13:41.6953896Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_InstanceNorm3d_tracking_stats_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 49%] 2023-03-31T07:13:41.6954479Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 49%] 2023-03-31T07:13:41.6955096Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 49%] 2023-03-31T07:13:41.6955675Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_log_target <- test/cpp_api_parity/module_impl_check.py PASSED [ 49%] 2023-03-31T07:13:41.6956286Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_log_target_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 49%] 2023-03-31T07:13:41.6956864Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar <- test/cpp_api_parity/module_impl_check.py PASSED [ 50%] 2023-03-31T07:13:41.6957463Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 50%] 2023-03-31T07:13:41.6958068Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_log_target <- test/cpp_api_parity/module_impl_check.py PASSED [ 50%] 2023-03-31T07:13:41.6958701Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_KLDivLoss_scalar_log_target_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 50%] 2023-03-31T07:13:41.6959276Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss <- test/cpp_api_parity/module_impl_check.py PASSED [ 50%] 2023-03-31T07:13:41.6959835Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 50%] 2023-03-31T07:13:41.6960401Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_scalar <- test/cpp_api_parity/module_impl_check.py PASSED [ 50%] 2023-03-31T07:13:41.6960988Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_L1Loss_scalar_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 50%] 2023-03-31T07:13:41.6961557Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 50%] 2023-03-31T07:13:41.6962127Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 51%] 2023-03-31T07:13:41.6962704Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 51%] 2023-03-31T07:13:41.6963544Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 51%] 2023-03-31T07:13:41.6964180Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_norm <- test/cpp_api_parity/module_impl_check.py PASSED [ 51%] 2023-03-31T07:13:41.6964770Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool1d_norm_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 51%] 2023-03-31T07:13:41.6965320Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 51%] 2023-03-31T07:13:41.6965898Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 51%] 2023-03-31T07:13:41.6966462Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_norm <- test/cpp_api_parity/module_impl_check.py PASSED [ 51%] 2023-03-31T07:13:41.6967049Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LPPool2d_norm_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 51%] 2023-03-31T07:13:41.6967641Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_elementwise_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 52%] 2023-03-31T07:13:41.6968324Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_elementwise_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 52%] 2023-03-31T07:13:41.6968963Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_empty_elementwise_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 52%] 2023-03-31T07:13:41.6969775Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_empty_elementwise_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 52%] 2023-03-31T07:13:41.6970390Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_no_elementwise_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 52%] 2023-03-31T07:13:41.6971038Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_1d_no_elementwise_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 52%] 2023-03-31T07:13:41.6971660Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_elementwise_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 52%] 2023-03-31T07:13:41.6972298Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_elementwise_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 52%] 2023-03-31T07:13:41.6972918Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_affine_large_feature <- test/cpp_api_parity/module_impl_check.py PASSED [ 53%] 2023-03-31T07:13:41.6973569Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_affine_large_feature_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 53%] 2023-03-31T07:13:41.6974203Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_elementwise_affine <- test/cpp_api_parity/module_impl_check.py PASSED [ 53%] 2023-03-31T07:13:41.6974847Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LayerNorm_3d_no_elementwise_affine_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 53%] 2023-03-31T07:13:41.6975430Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear <- test/cpp_api_parity/module_impl_check.py PASSED [ 53%] 2023-03-31T07:13:41.6975989Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 53%] 2023-03-31T07:13:41.6976560Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 53%] 2023-03-31T07:13:41.6977163Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 53%] 2023-03-31T07:13:41.6977783Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_bias <- test/cpp_api_parity/module_impl_check.py PASSED [ 53%] 2023-03-31T07:13:41.6978391Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Linear_no_bias_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 54%] 2023-03-31T07:13:41.6978982Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 54%] 2023-03-31T07:13:41.6979599Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 54%] 2023-03-31T07:13:41.6980212Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_2d_uneven_pad <- test/cpp_api_parity/module_impl_check.py PASSED [ 54%] 2023-03-31T07:13:41.6980841Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_2d_uneven_pad_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 54%] 2023-03-31T07:13:41.6981477Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_3d_custom_params <- test/cpp_api_parity/module_impl_check.py PASSED [ 54%] 2023-03-31T07:13:41.6982131Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_LocalResponseNorm_3d_custom_params_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 54%] 2023-03-31T07:13:41.6982752Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 54%] 2023-03-31T07:13:41.6983418Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 54%] 2023-03-31T07:13:41.6983983Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_prec <- test/cpp_api_parity/module_impl_check.py PASSED [ 55%] 2023-03-31T07:13:41.6984571Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_prec_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 55%] 2023-03-31T07:13:41.6985144Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_scalar <- test/cpp_api_parity/module_impl_check.py PASSED [ 55%] 2023-03-31T07:13:41.6985728Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MSELoss_scalar_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 55%] 2023-03-31T07:13:41.6986297Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 55%] 2023-03-31T07:13:41.6986906Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 55%] 2023-03-31T07:13:41.6987506Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_margin <- test/cpp_api_parity/module_impl_check.py PASSED [ 55%] 2023-03-31T07:13:41.6988138Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_margin_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 55%] 2023-03-31T07:13:41.6988749Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 56%] 2023-03-31T07:13:41.6989406Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 56%] 2023-03-31T07:13:41.6990033Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py PASSED [ 56%] 2023-03-31T07:13:41.6990685Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 56%] 2023-03-31T07:13:41.6991302Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 56%] 2023-03-31T07:13:41.6991991Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MarginRankingLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 56%] 2023-03-31T07:13:41.6992614Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 56%] 2023-03-31T07:13:41.6993195Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 56%] 2023-03-31T07:13:41.6993755Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 56%] 2023-03-31T07:13:41.6994358Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool1d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 57%] 2023-03-31T07:13:41.6994940Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_3d_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 57%] 2023-03-31T07:13:41.6995542Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_3d_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 57%] 2023-03-31T07:13:41.6996119Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_4d_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 57%] 2023-03-31T07:13:41.6996732Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool2d_4d_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 57%] 2023-03-31T07:13:41.6997324Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 57%] 2023-03-31T07:13:41.6999358Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 57%] 2023-03-31T07:13:41.6999969Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride <- test/cpp_api_parity/module_impl_check.py PASSED [ 57%] 2023-03-31T07:13:41.7000574Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 58%] 2023-03-31T07:13:41.7001156Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_padding <- test/cpp_api_parity/module_impl_check.py PASSED [ 58%] 2023-03-31T07:13:41.7001783Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MaxPool3d_stride_padding_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 58%] 2023-03-31T07:13:41.7002386Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 58%] 2023-03-31T07:13:41.7003011Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 58%] 2023-03-31T07:13:41.7003803Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 58%] 2023-03-31T07:13:41.7004453Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 58%] 2023-03-31T07:13:41.7005084Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 58%] 2023-03-31T07:13:41.7005758Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 58%] 2023-03-31T07:13:41.7006399Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py PASSED [ 59%] 2023-03-31T07:13:41.7007063Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 59%] 2023-03-31T07:13:41.7007790Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 59%] 2023-03-31T07:13:41.7008498Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelMarginLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 59%] 2023-03-31T07:13:41.7009111Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 59%] 2023-03-31T07:13:41.7009757Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 59%] 2023-03-31T07:13:41.7010410Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 59%] 2023-03-31T07:13:41.7011095Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 59%] 2023-03-31T07:13:41.7011765Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py PASSED [ 59%] 2023-03-31T07:13:41.7012428Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 60%] 2023-03-31T07:13:41.7013097Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 60%] 2023-03-31T07:13:41.7013907Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 60%] 2023-03-31T07:13:41.7014568Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 60%] 2023-03-31T07:13:41.7015218Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiLabelSoftMarginLoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 60%] 2023-03-31T07:13:41.7015830Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 60%] 2023-03-31T07:13:41.7016393Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_1d <- test/cpp_api_parity/module_impl_check.py PASSED [ 60%] 2023-03-31T07:13:41.7016998Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_1d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 60%] 2023-03-31T07:13:41.7017611Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 61%] 2023-03-31T07:13:41.7018205Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_margin <- test/cpp_api_parity/module_impl_check.py PASSED [ 61%] 2023-03-31T07:13:41.7018828Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_margin_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 61%] 2023-03-31T07:13:41.7019422Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_p <- test/cpp_api_parity/module_impl_check.py PASSED [ 61%] 2023-03-31T07:13:41.7020027Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_p_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 61%] 2023-03-31T07:13:41.7020612Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 61%] 2023-03-31T07:13:41.7021242Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_MultiMarginLoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 61%] 2023-03-31T07:13:41.7021862Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 61%] 2023-03-31T07:13:41.7022431Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d <- test/cpp_api_parity/module_impl_check.py PASSED [ 61%] 2023-03-31T07:13:41.7023080Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 62%] 2023-03-31T07:13:41.7023665Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 62%] 2023-03-31T07:13:41.7024284Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 62%] 2023-03-31T07:13:41.7024878Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 62%] 2023-03-31T07:13:41.7025471Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_2d_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 62%] 2023-03-31T07:13:41.7026079Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 62%] 2023-03-31T07:13:41.7026641Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_dim_is_3 <- test/cpp_api_parity/module_impl_check.py PASSED [ 62%] 2023-03-31T07:13:41.7027234Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_dim_is_3_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 62%] 2023-03-31T07:13:41.7027840Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_higher_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 62%] 2023-03-31T07:13:41.7028480Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_higher_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 63%] 2023-03-31T07:13:41.7029069Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 63%] 2023-03-31T07:13:41.7029678Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 63%] 2023-03-31T07:13:41.7030261Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py XFAIL [ 63%] 2023-03-31T07:13:41.7030879Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 63%] 2023-03-31T07:13:41.7031473Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py XFAIL [ 63%] 2023-03-31T07:13:41.7032089Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 63%] 2023-03-31T07:13:41.7032678Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py XFAIL [ 63%] 2023-03-31T07:13:41.7033269Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 64%] 2023-03-31T07:13:41.7033851Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights <- test/cpp_api_parity/module_impl_check.py PASSED [ 64%] 2023-03-31T07:13:41.7034443Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 64%] 2023-03-31T07:13:41.7035043Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index <- test/cpp_api_parity/module_impl_check.py PASSED [ 64%] 2023-03-31T07:13:41.7035690Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 64%] 2023-03-31T07:13:41.7036338Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_neg <- test/cpp_api_parity/module_impl_check.py PASSED [ 64%] 2023-03-31T07:13:41.7036978Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_NLLLoss_weights_ignore_index_neg_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 64%] 2023-03-31T07:13:41.7037575Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance <- test/cpp_api_parity/module_impl_check.py PASSED [ 64%] 2023-03-31T07:13:41.7038155Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_lhs <- test/cpp_api_parity/module_impl_check.py PASSED [ 64%] 2023-03-31T07:13:41.7038798Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_lhs_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 65%] 2023-03-31T07:13:41.7039428Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_rhs <- test/cpp_api_parity/module_impl_check.py PASSED [ 65%] 2023-03-31T07:13:41.7040072Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_broadcast_rhs_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 65%] 2023-03-31T07:13:41.7040693Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 65%] 2023-03-31T07:13:41.7041300Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 65%] 2023-03-31T07:13:41.7042010Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 65%] 2023-03-31T07:13:41.7042649Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_with_non_default_args <- test/cpp_api_parity/module_impl_check.py PASSED [ 65%] 2023-03-31T07:13:41.7043465Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PairwiseDistance_with_non_default_args_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 65%] 2023-03-31T07:13:41.7044055Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelShuffle <- test/cpp_api_parity/module_impl_check.py PASSED [ 66%] 2023-03-31T07:13:41.7044642Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelShuffle_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 66%] 2023-03-31T07:13:41.7045222Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelUnshuffle <- test/cpp_api_parity/module_impl_check.py PASSED [ 66%] 2023-03-31T07:13:41.7045822Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PixelUnshuffle_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 66%] 2023-03-31T07:13:41.7046405Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss <- test/cpp_api_parity/module_impl_check.py PASSED [ 66%] 2023-03-31T07:13:41.7047024Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 66%] 2023-03-31T07:13:41.7047638Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_no_log_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 66%] 2023-03-31T07:13:41.7048284Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_full_loss_no_log_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 66%] 2023-03-31T07:13:41.7048892Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss <- test/cpp_api_parity/module_impl_check.py PASSED [ 66%] 2023-03-31T07:13:41.7049515Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 67%] 2023-03-31T07:13:41.7050206Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_no_log_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 67%] 2023-03-31T07:13:41.7050904Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_PoissonNLLLoss_no_full_loss_no_log_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 67%] 2023-03-31T07:13:41.7051480Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU <- test/cpp_api_parity/module_impl_check.py PASSED [ 67%] 2023-03-31T07:13:41.7052047Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:13:41.7052627Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down <- test/cpp_api_parity/module_impl_check.py PASSED [ 67%] 2023-03-31T07:13:41.7053238Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:13:41.7053840Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_scalar <- test/cpp_api_parity/module_impl_check.py PASSED [ 67%] 2023-03-31T07:13:41.7054453Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_RReLU_with_up_down_scalar_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (Excluded from CUDA tests) [ 67%] 2023-03-31T07:13:41.7055050Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d <- test/cpp_api_parity/module_impl_check.py PASSED [ 68%] 2023-03-31T07:13:41.7055667Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_complex <- test/cpp_api_parity/module_impl_check.py PASSED [ 68%] 2023-03-31T07:13:41.7056337Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_complex_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 68%] 2023-03-31T07:13:41.7056970Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 68%] 2023-03-31T07:13:41.7057575Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 68%] 2023-03-31T07:13:41.7058207Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_ReplicationPad3d_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 68%] 2023-03-31T07:13:41.7058818Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_has_parity <- test/cpp_api_parity/module_impl_check.py PASSED [ 68%] 2023-03-31T07:13:41.7059427Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_has_parity_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 68%] 2023-03-31T07:13:41.7060026Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_no_parity <- test/cpp_api_parity/module_impl_check.py XFAIL [ 69%] 2023-03-31T07:13:41.7060646Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SampleModule_no_parity_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 69%] 2023-03-31T07:13:41.7061221Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss <- test/cpp_api_parity/module_impl_check.py PASSED [ 69%] 2023-03-31T07:13:41.7061809Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 69%] 2023-03-31T07:13:41.7062373Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_scalar <- test/cpp_api_parity/module_impl_check.py PASSED [ 69%] 2023-03-31T07:13:41.7063054Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SmoothL1Loss_scalar_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 69%] 2023-03-31T07:13:41.7063647Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss <- test/cpp_api_parity/module_impl_check.py PASSED [ 69%] 2023-03-31T07:13:41.7064297Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 69%] 2023-03-31T07:13:41.7064916Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py XFAIL [ 69%] 2023-03-31T07:13:41.7065554Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 70%] 2023-03-31T07:13:41.7066177Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py XFAIL [ 70%] 2023-03-31T07:13:41.7066814Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 70%] 2023-03-31T07:13:41.7067423Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py XFAIL [ 70%] 2023-03-31T07:13:41.7068053Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_SoftMarginLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 70%] 2023-03-31T07:13:41.7068698Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_gelu_activation <- test/cpp_api_parity/module_impl_check.py PASSED [ 70%] 2023-03-31T07:13:41.7069380Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_gelu_activation_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 70%] 2023-03-31T07:13:41.7070073Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_relu_activation <- test/cpp_api_parity/module_impl_check.py PASSED [ 70%] 2023-03-31T07:13:41.7070770Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerDecoderLayer_relu_activation_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 70%] 2023-03-31T07:13:41.7071433Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_gelu_activation <- test/cpp_api_parity/module_impl_check.py PASSED [ 71%] 2023-03-31T07:13:41.7072107Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_gelu_activation_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:13:41.7072770Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_relu_activation <- test/cpp_api_parity/module_impl_check.py PASSED [ 71%] 2023-03-31T07:13:41.7073435Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TransformerEncoderLayer_relu_activation_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:13:41.7074072Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Transformer_multilayer_coder <- test/cpp_api_parity/module_impl_check.py PASSED [ 71%] 2023-03-31T07:13:41.7074705Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Transformer_multilayer_coder_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:13:41.7075334Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_mean <- test/cpp_api_parity/module_impl_check.py PASSED [ 71%] 2023-03-31T07:13:41.7075971Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_mean_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 71%] 2023-03-31T07:13:41.7076600Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_none <- test/cpp_api_parity/module_impl_check.py PASSED [ 72%] 2023-03-31T07:13:41.7077251Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_none_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 72%] 2023-03-31T07:13:41.7077913Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_sum <- test/cpp_api_parity/module_impl_check.py PASSED [ 72%] 2023-03-31T07:13:41.7078589Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_TripletMarginLoss_no_batch_dim_sum_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 72%] 2023-03-31T07:13:41.7079182Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unflatten_no_batch_dim <- test/cpp_api_parity/module_impl_check.py PASSED [ 72%] 2023-03-31T07:13:41.7079793Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unflatten_no_batch_dim_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 72%] 2023-03-31T07:13:41.7080364Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold <- test/cpp_api_parity/module_impl_check.py PASSED [ 72%] 2023-03-31T07:13:41.7080939Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 72%] 2023-03-31T07:13:41.7081490Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_int_input <- test/cpp_api_parity/module_impl_check.py PASSED [ 72%] 2023-03-31T07:13:41.7082086Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_Unfold_int_input_cuda <- test/cpp_api_parity/module_impl_check.py SKIPPED (CUDA unavailable) [ 73%] 2023-03-31T07:13:41.7082690Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 73%] 2023-03-31T07:13:41.7083487Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 73%] 2023-03-31T07:13:41.7084159Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 73%] 2023-03-31T07:13:41.7084853Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 73%] 2023-03-31T07:13:41.7085500Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 73%] 2023-03-31T07:13:41.7086164Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 73%] 2023-03-31T07:13:41.7086817Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 73%] 2023-03-31T07:13:41.7087491Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCELoss_weights_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 74%] 2023-03-31T07:13:41.7088151Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_legacy_enum <- test/cpp_api_parity/functional_impl_check.py PASSED [ 74%] 2023-03-31T07:13:41.7088835Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_legacy_enum_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 74%] 2023-03-31T07:13:41.7089494Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 74%] 2023-03-31T07:13:41.7090155Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 74%] 2023-03-31T07:13:41.7090824Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 74%] 2023-03-31T07:13:41.7091520Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_BCEWithLogitsLoss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 74%] 2023-03-31T07:13:41.7092242Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_margin_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 74%] 2023-03-31T07:13:41.7092983Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_margin_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 74%] 2023-03-31T07:13:41.7093634Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 75%] 2023-03-31T07:13:41.7094316Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HingeEmbeddingLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 75%] 2023-03-31T07:13:41.7094952Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HuberLoss_delta <- test/cpp_api_parity/functional_impl_check.py PASSED [ 75%] 2023-03-31T07:13:41.7095595Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_HuberLoss_delta_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 75%] 2023-03-31T07:13:41.7096214Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 75%] 2023-03-31T07:13:41.7096860Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 75%] 2023-03-31T07:13:41.7097538Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_log_target <- test/cpp_api_parity/functional_impl_check.py PASSED [ 75%] 2023-03-31T07:13:41.7098251Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_log_target_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 75%] 2023-03-31T07:13:41.7098913Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 75%] 2023-03-31T07:13:41.7099566Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 76%] 2023-03-31T07:13:41.7100228Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_log_target <- test/cpp_api_parity/functional_impl_check.py PASSED [ 76%] 2023-03-31T07:13:41.7100922Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_no_reduce_scalar_log_target_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 76%] 2023-03-31T07:13:41.7101599Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_log_target_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 76%] 2023-03-31T07:13:41.7102285Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_log_target_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 76%] 2023-03-31T07:13:41.7102935Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_target_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 76%] 2023-03-31T07:13:41.7103712Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_KLDivLoss_with_target_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 76%] 2023-03-31T07:13:41.7104344Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 76%] 2023-03-31T07:13:41.7104958Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_complex <- test/cpp_api_parity/functional_impl_check.py PASSED [ 77%] 2023-03-31T07:13:41.7105643Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_complex_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 77%] 2023-03-31T07:13:41.7106349Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 77%] 2023-03-31T07:13:41.7106977Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 77%] 2023-03-31T07:13:41.7107628Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_L1Loss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 77%] 2023-03-31T07:13:41.7108243Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 77%] 2023-03-31T07:13:41.7108882Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 77%] 2023-03-31T07:13:41.7109522Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 77%] 2023-03-31T07:13:41.7110176Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MSELoss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 77%] 2023-03-31T07:13:41.7110848Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_0d_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 78%] 2023-03-31T07:13:41.7111572Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_0d_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 78%] 2023-03-31T07:13:41.7112278Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_1d_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 78%] 2023-03-31T07:13:41.7112982Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_1d_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 78%] 2023-03-31T07:13:41.7113659Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_index_neg <- test/cpp_api_parity/functional_impl_check.py PASSED [ 78%] 2023-03-31T07:13:41.7114350Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_index_neg_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 78%] 2023-03-31T07:13:41.7115012Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 78%] 2023-03-31T07:13:41.7115703Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelMarginLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 78%] 2023-03-31T07:13:41.7116388Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 79%] 2023-03-31T07:13:41.7117098Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 79%] 2023-03-31T07:13:41.7117795Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_weights_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 79%] 2023-03-31T07:13:41.7118523Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiLabelSoftMarginLoss_weights_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 79%] 2023-03-31T07:13:41.7119209Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_1d_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 79%] 2023-03-31T07:13:41.7119913Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_1d_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 79%] 2023-03-31T07:13:41.7120607Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_margin_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 79%] 2023-03-31T07:13:41.7121277Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_margin_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 79%] 2023-03-31T07:13:41.7121933Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 79%] 2023-03-31T07:13:41.7122598Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 80%] 2023-03-31T07:13:41.7123393Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_p_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 80%] 2023-03-31T07:13:41.7124053Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_p_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 80%] 2023-03-31T07:13:41.7124717Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_weights_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 80%] 2023-03-31T07:13:41.7125459Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_MultiMarginLoss_weights_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 80%] 2023-03-31T07:13:41.7126148Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 80%] 2023-03-31T07:13:41.7126799Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 80%] 2023-03-31T07:13:41.7127438Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_ignore_index <- test/cpp_api_parity/functional_impl_check.py PASSED [ 80%] 2023-03-31T07:13:41.7128122Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_ignore_index_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 80%] 2023-03-31T07:13:41.7128776Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_weights <- test/cpp_api_parity/functional_impl_check.py PASSED [ 81%] 2023-03-31T07:13:41.7129444Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss2d_no_reduce_weights_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 81%] 2023-03-31T07:13:41.7130073Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 81%] 2023-03-31T07:13:41.7130717Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 81%] 2023-03-31T07:13:41.7131365Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_ignore_index <- test/cpp_api_parity/functional_impl_check.py PASSED [ 81%] 2023-03-31T07:13:41.7132039Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_ignore_index_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 81%] 2023-03-31T07:13:41.7132703Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_weights <- test/cpp_api_parity/functional_impl_check.py PASSED [ 81%] 2023-03-31T07:13:41.7133396Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLossNd_no_reduce_weights_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 81%] 2023-03-31T07:13:41.7134075Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 82%] 2023-03-31T07:13:41.7134715Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 82%] 2023-03-31T07:13:41.7135356Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_ignore_index <- test/cpp_api_parity/functional_impl_check.py PASSED [ 82%] 2023-03-31T07:13:41.7136017Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_ignore_index_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 82%] 2023-03-31T07:13:41.7136666Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights <- test/cpp_api_parity/functional_impl_check.py PASSED [ 82%] 2023-03-31T07:13:41.7153902Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 82%] 2023-03-31T07:13:41.7154617Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index <- test/cpp_api_parity/functional_impl_check.py PASSED [ 82%] 2023-03-31T07:13:41.7155334Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 82%] 2023-03-31T07:13:41.7156156Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_neg <- test/cpp_api_parity/functional_impl_check.py PASSED [ 82%] 2023-03-31T07:13:41.7156927Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_NLLLoss_no_reduce_weights_ignore_index_neg_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 83%] 2023-03-31T07:13:41.7157584Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding122112_3dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 83%] 2023-03-31T07:13:41.7158245Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding122112_3dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 83%] 2023-03-31T07:13:41.7158887Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding1221_2dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 83%] 2023-03-31T07:13:41.7159548Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding1221_2dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 83%] 2023-03-31T07:13:41.7160168Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding12_1dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 83%] 2023-03-31T07:13:41.7160818Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding12_1dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 83%] 2023-03-31T07:13:41.7161461Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding2322_2dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 83%] 2023-03-31T07:13:41.7162117Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding2322_2dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 83%] 2023-03-31T07:13:41.7162752Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding31_1dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 84%] 2023-03-31T07:13:41.7163558Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding31_1dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 84%] 2023-03-31T07:13:41.7164279Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding322112_3dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 84%] 2023-03-31T07:13:41.7164993Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding322112_3dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 84%] 2023-03-31T07:13:41.7165638Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding332122_3dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 84%] 2023-03-31T07:13:41.7166286Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding332122_3dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 84%] 2023-03-31T07:13:41.7166936Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding3331_2dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 84%] 2023-03-31T07:13:41.7167596Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding3331_2dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 84%] 2023-03-31T07:13:41.7168232Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding33_1dcircular <- test/cpp_api_parity/functional_impl_check.py PASSED [ 85%] 2023-03-31T07:13:41.7168879Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_Padding33_1dcircular_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 85%] 2023-03-31T07:13:41.7169502Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_PoissonNLLLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 85%] 2023-03-31T07:13:41.7170218Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_PoissonNLLLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 85%] 2023-03-31T07:13:41.7170897Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_beta <- test/cpp_api_parity/functional_impl_check.py PASSED [ 85%] 2023-03-31T07:13:41.7171546Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_beta_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 85%] 2023-03-31T07:13:41.7172165Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 85%] 2023-03-31T07:13:41.7172823Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 85%] 2023-03-31T07:13:41.7173476Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 85%] 2023-03-31T07:13:41.7174152Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_no_reduce_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 86%] 2023-03-31T07:13:41.7174798Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_zero_beta <- test/cpp_api_parity/functional_impl_check.py PASSED [ 86%] 2023-03-31T07:13:41.7175438Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SmoothL1Loss_zero_beta_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 86%] 2023-03-31T07:13:41.7176080Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SoftMarginLoss_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 86%] 2023-03-31T07:13:41.7176740Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_SoftMarginLoss_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 86%] 2023-03-31T07:13:41.7177388Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 86%] 2023-03-31T07:13:41.7178061Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 86%] 2023-03-31T07:13:41.7178744Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 86%] 2023-03-31T07:13:41.7179426Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_2d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 87%] 2023-03-31T07:13:41.7180079Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 87%] 2023-03-31T07:13:41.7180756Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 87%] 2023-03-31T07:13:41.7181416Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_shared_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 87%] 2023-03-31T07:13:41.7182126Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_shared_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 87%] 2023-03-31T07:13:41.7182810Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 87%] 2023-03-31T07:13:41.7183596Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 87%] 2023-03-31T07:13:41.7184391Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 87%] 2023-03-31T07:13:41.7185139Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_scale_tuple_skewed_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 87%] 2023-03-31T07:13:41.7185807Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 88%] 2023-03-31T07:13:41.7186456Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 88%] 2023-03-31T07:13:41.7187166Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 88%] 2023-03-31T07:13:41.7187858Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bicubic_tuple_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 88%] 2023-03-31T07:13:41.7188498Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 88%] 2023-03-31T07:13:41.7189155Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 88%] 2023-03-31T07:13:41.7189810Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 88%] 2023-03-31T07:13:41.7190493Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_2d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 88%] 2023-03-31T07:13:41.7191141Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 88%] 2023-03-31T07:13:41.7191847Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 89%] 2023-03-31T07:13:41.7192554Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_shared_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 89%] 2023-03-31T07:13:41.7193258Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_shared_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 89%] 2023-03-31T07:13:41.7193933Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 89%] 2023-03-31T07:13:41.7194625Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 89%] 2023-03-31T07:13:41.7195373Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 89%] 2023-03-31T07:13:41.7196117Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_scale_tuple_skewed_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 89%] 2023-03-31T07:13:41.7196775Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 89%] 2023-03-31T07:13:41.7197446Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 90%] 2023-03-31T07:13:41.7198209Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 90%] 2023-03-31T07:13:41.7198916Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_bilinear_tuple_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 90%] 2023-03-31T07:13:41.7199555Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 90%] 2023-03-31T07:13:41.7200191Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 90%] 2023-03-31T07:13:41.7200868Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 90%] 2023-03-31T07:13:41.7201556Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 90%] 2023-03-31T07:13:41.7202209Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 90%] 2023-03-31T07:13:41.7202876Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_1d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 90%] 2023-03-31T07:13:41.7203663Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 91%] 2023-03-31T07:13:41.7204314Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 91%] 2023-03-31T07:13:41.7205025Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 91%] 2023-03-31T07:13:41.7205801Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_scale_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 91%] 2023-03-31T07:13:41.7206490Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_tuple_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 91%] 2023-03-31T07:13:41.7207149Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_linear_tuple_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 91%] 2023-03-31T07:13:41.7207793Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 91%] 2023-03-31T07:13:41.7208449Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 91%] 2023-03-31T07:13:41.7209108Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 91%] 2023-03-31T07:13:41.7209772Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_1d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 92%] 2023-03-31T07:13:41.7210418Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 92%] 2023-03-31T07:13:41.7211113Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 92%] 2023-03-31T07:13:41.7211816Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_launch_configs <- test/cpp_api_parity/functional_impl_check.py PASSED [ 92%] 2023-03-31T07:13:41.7212525Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_launch_configs_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 92%] 2023-03-31T07:13:41.7213178Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 92%] 2023-03-31T07:13:41.7213852Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_2d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 92%] 2023-03-31T07:13:41.7214500Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 92%] 2023-03-31T07:13:41.7215157Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 93%] 2023-03-31T07:13:41.7215798Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 93%] 2023-03-31T07:13:41.7216476Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_3d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 93%] 2023-03-31T07:13:41.7217129Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 93%] 2023-03-31T07:13:41.7217801Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 93%] 2023-03-31T07:13:41.7218456Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 93%] 2023-03-31T07:13:41.7219155Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 93%] 2023-03-31T07:13:41.7219834Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 93%] 2023-03-31T07:13:41.7220494Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_scale_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 93%] 2023-03-31T07:13:41.7221143Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_1d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 94%] 2023-03-31T07:13:41.7221819Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_1d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 94%] 2023-03-31T07:13:41.7222462Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_2d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 94%] 2023-03-31T07:13:41.7223201Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_2d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 94%] 2023-03-31T07:13:41.7223861Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 94%] 2023-03-31T07:13:41.7224527Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_nearest_tuple_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 94%] 2023-03-31T07:13:41.7225201Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 94%] 2023-03-31T07:13:41.7225900Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 94%] 2023-03-31T07:13:41.7226562Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_zero_dim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 95%] 2023-03-31T07:13:41.7227242Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_3d_zero_dim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 95%] 2023-03-31T07:13:41.7227902Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 95%] 2023-03-31T07:13:41.7228554Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 95%] 2023-03-31T07:13:41.7229270Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 95%] 2023-03-31T07:13:41.7229980Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_scale_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 95%] 2023-03-31T07:13:41.7230635Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d <- test/cpp_api_parity/functional_impl_check.py PASSED [ 95%] 2023-03-31T07:13:41.7231276Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_align_corners <- test/cpp_api_parity/functional_impl_check.py PASSED [ 95%] 2023-03-31T07:13:41.7231995Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_align_corners_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 95%] 2023-03-31T07:13:41.7232740Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_interpolate_trilinear_tuple_3d_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 96%] 2023-03-31T07:13:41.7233399Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim0 <- test/cpp_api_parity/functional_impl_check.py PASSED [ 96%] 2023-03-31T07:13:41.7234035Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim0_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 96%] 2023-03-31T07:13:41.7234631Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim3 <- test/cpp_api_parity/functional_impl_check.py PASSED [ 96%] 2023-03-31T07:13:41.7235269Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_dim3_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 96%] 2023-03-31T07:13:41.7235898Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_lastdim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 96%] 2023-03-31T07:13:41.7236542Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_lastdim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 96%] 2023-03-31T07:13:41.7237149Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 96%] 2023-03-31T07:13:41.7237780Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 96%] 2023-03-31T07:13:41.7238441Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial <- test/cpp_api_parity/functional_impl_check.py PASSED [ 97%] 2023-03-31T07:13:41.7239111Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 97%] 2023-03-31T07:13:41.7239760Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_special <- test/cpp_api_parity/functional_impl_check.py PASSED [ 97%] 2023-03-31T07:13:41.7240408Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_log_softmax_spatial_special_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 97%] 2023-03-31T07:13:41.7241084Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_multimarginloss_1d_input_0d_target_no_reduce <- test/cpp_api_parity/functional_impl_check.py PASSED [ 97%] 2023-03-31T07:13:41.7241784Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_multimarginloss_1d_input_0d_target_no_reduce_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 97%] 2023-03-31T07:13:41.7242450Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_has_parity <- test/cpp_api_parity/functional_impl_check.py PASSED [ 97%] 2023-03-31T07:13:41.7243230Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_has_parity_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 97%] 2023-03-31T07:13:41.7243885Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_no_parity <- test/cpp_api_parity/functional_impl_check.py XFAIL [ 98%] 2023-03-31T07:13:41.7244551Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_sample_functional_no_parity_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 98%] 2023-03-31T07:13:41.7245192Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim0 <- test/cpp_api_parity/functional_impl_check.py PASSED [ 98%] 2023-03-31T07:13:41.7245864Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim0_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (Excluded from CUDA tests) [ 98%] 2023-03-31T07:13:41.7246546Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim3 <- test/cpp_api_parity/functional_impl_check.py PASSED [ 98%] 2023-03-31T07:13:41.7247267Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_dim3_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (Excluded from CUDA tests) [ 98%] 2023-03-31T07:13:41.7247917Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_scalar <- test/cpp_api_parity/functional_impl_check.py PASSED [ 98%] 2023-03-31T07:13:41.7248590Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_functional_scalar_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (Excluded from CUDA tests) [ 98%] 2023-03-31T07:13:41.7249208Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim <- test/cpp_api_parity/functional_impl_check.py PASSED [ 98%] 2023-03-31T07:13:41.7249850Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 99%] 2023-03-31T07:13:41.7250478Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_dtype <- test/cpp_api_parity/functional_impl_check.py PASSED [ 99%] 2023-03-31T07:13:41.7251145Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_lastdim_dtype_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (Excluded from CUDA tests) [ 99%] 2023-03-31T07:13:41.7251781Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial <- test/cpp_api_parity/functional_impl_check.py PASSED [ 99%] 2023-03-31T07:13:41.7252443Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [ 99%] 2023-03-31T07:13:41.7253132Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_dtype <- test/cpp_api_parity/functional_impl_check.py PASSED [ 99%] 2023-03-31T07:13:41.7253798Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_dtype_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (Excluded from CUDA tests) [ 99%] 2023-03-31T07:13:41.7254447Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_special <- test/cpp_api_parity/functional_impl_check.py PASSED [ 99%] 2023-03-31T07:13:41.7255087Z test_cpp_api_parity.py::TestCppApiParity::test_torch_nn_functional_softmax_spatial_special_cuda <- test/cpp_api_parity/functional_impl_check.py SKIPPED (CUDA unavailable) [100%] 2023-03-31T07:13:41.7255385Z 2023-03-31T07:13:41.7255740Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cpp_api_parity/test_cpp_api_parity-e0456c88272dc2e6.xml - 2023-03-31T07:13:41.7256135Z ================= 414 passed, 431 skipped, 17 xfailed in 6.48s ================= 2023-03-31T07:13:41.7256478Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:41.7256683Z 2023-03-31T07:13:41.7257036Z ##[endgroup] 2023-03-31T07:13:41.7257467Z FINISHED PRINTING LOG FILE of test_cpp_api_parity (/var/lib/jenkins/workspace/test/test-reports/test_cpp_api_parity_c6t8ir31.log) 2023-03-31T07:13:41.7257710Z 2023-03-31T07:13:41.7257897Z Running test_mobile_optimizer ... [2023-03-31 07:13:41.636077] 2023-03-31T07:13:41.7258535Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_mobile_optimizer.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:41.636579] 2023-03-31T07:13:46.8006889Z 2023-03-31T07:13:46.8007456Z Expand the folded group to see the log file of test_mobile_optimizer 2023-03-31T07:13:46.8008589Z ##[group]PRINTING LOG FILE of test_mobile_optimizer (/var/lib/jenkins/workspace/test/test-reports/test_mobile_optimizer_lr26e4oo.log) 2023-03-31T07:13:46.8009940Z Test results will be stored in test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-741c2318ad0d4c63.xml 2023-03-31T07:13:46.8010439Z ============================= test session starts ============================== 2023-03-31T07:13:46.8010856Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:46.8011189Z cachedir: .pytest_cache 2023-03-31T07:13:46.8011643Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:46.8012087Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:46.8012552Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:46.8012932Z collecting ... collected 7 items 2023-03-31T07:13:46.8013847Z Running 7 items in this shard: test/test_mobile_optimizer.py::TestOptimizer::test_clone_module_with_class, test/test_mobile_optimizer.py::TestOptimizer::test_generate_mobile_module_lints, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_mobilenet_optimize_for_mobile, test/test_mobile_optimizer.py::TestOptimizer::test_optimize_for_mobile, test/test_mobile_optimizer.py::TestOptimizer::test_preserve_bundled_inputs_methods, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures 2023-03-31T07:13:46.8014626Z 2023-03-31T07:13:46.8014795Z test_mobile_optimizer.py::TestOptimizer::test_clone_module_with_class PASSED [ 14%] 2023-03-31T07:13:46.8015237Z test_mobile_optimizer.py::TestOptimizer::test_generate_mobile_module_lints PASSED [ 28%] 2023-03-31T07:13:46.8015676Z test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params PASSED [ 42%] 2023-03-31T07:13:46.8016095Z test_mobile_optimizer.py::TestOptimizer::test_mobilenet_optimize_for_mobile PASSED [ 57%] 2023-03-31T07:13:46.8016510Z test_mobile_optimizer.py::TestOptimizer::test_optimize_for_mobile PASSED [ 71%] 2023-03-31T07:13:46.8016959Z test_mobile_optimizer.py::TestOptimizer::test_preserve_bundled_inputs_methods PASSED [ 85%] 2023-03-31T07:13:46.8017319Z test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures PASSED [100%] 2023-03-31T07:13:46.8017602Z 2023-03-31T07:13:46.8017974Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-741c2318ad0d4c63.xml - 2023-03-31T07:13:46.8018430Z ============================== 7 passed in 2.86s =============================== 2023-03-31T07:13:46.8018749Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:46.8018968Z 2023-03-31T07:13:46.8019256Z ##[endgroup] 2023-03-31T07:13:46.8019690Z FINISHED PRINTING LOG FILE of test_mobile_optimizer (/var/lib/jenkins/workspace/test/test-reports/test_mobile_optimizer_lr26e4oo.log) 2023-03-31T07:13:46.8020022Z 2023-03-31T07:13:46.8020263Z Running test_cpp_extensions_open_device_registration ... [2023-03-31 07:13:46.800953] 2023-03-31T07:13:46.8021068Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cpp_extensions_open_device_registration.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:46.801279] 2023-03-31T07:13:51.3631598Z 2023-03-31T07:13:51.3632253Z Expand the folded group to see the log file of test_cpp_extensions_open_device_registration 2023-03-31T07:13:51.3633563Z ##[group]PRINTING LOG FILE of test_cpp_extensions_open_device_registration (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_open_device_registration_ymgastf1.log) 2023-03-31T07:13:51.3634954Z Test results will be stored in test-reports/python-pytest/test_cpp_extensions_open_device_registration/test_cpp_extensions_open_device_registration-7c212efa3ee59096.xml 2023-03-31T07:13:51.3635578Z ============================= test session starts ============================== 2023-03-31T07:13:51.3636585Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:51.3637104Z cachedir: .pytest_cache 2023-03-31T07:13:51.3637814Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:51.3638421Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:51.3639141Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:51.3639619Z collecting ... collected 1 item 2023-03-31T07:13:51.3640243Z Running 1 items in this shard: test/test_cpp_extensions_open_device_registration.py::TestCppExtensionOpenRgistration::test_open_device_registration 2023-03-31T07:13:51.3640696Z 2023-03-31T07:13:51.3643392Z test_cpp_extensions_open_device_registration.py::TestCppExtensionOpenRgistration::test_open_device_registration [1/2] c++ -MMD -MF open_registration_extension.o.d -DTORCH_EXTENSION_NAME=custom_device_extension -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_clang\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1002\" -I/var/lib/jenkins/workspace/test/cpp_extensions -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -isystem /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -isystem /opt/conda/envs/py_3.8/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=1 -fPIC -std=c++17 -g -c /var/lib/jenkins/workspace/test/cpp_extensions/open_registration_extension.cpp -o open_registration_extension.o 2023-03-31T07:13:51.3645708Z [2/2] c++ open_registration_extension.o -shared -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o custom_device_extension.so 2023-03-31T07:13:51.3646370Z PASSED [100%]No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:13:51.3646610Z 2023-03-31T07:13:51.3646620Z 2023-03-31T07:13:51.3647219Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cpp_extensions_open_device_registration/test_cpp_extensions_open_device_registration-7c212efa3ee59096.xml - 2023-03-31T07:13:51.3647822Z ============================== 1 passed in 2.19s =============================== 2023-03-31T07:13:51.3648226Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:51.3648497Z 2023-03-31T07:13:51.3648860Z ##[endgroup] 2023-03-31T07:13:51.3649559Z FINISHED PRINTING LOG FILE of test_cpp_extensions_open_device_registration (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_open_device_registration_ymgastf1.log) 2023-03-31T07:13:51.3649960Z 2023-03-31T07:13:51.3650208Z Running test_namedtuple_return_api ... [2023-03-31 07:13:51.363514] 2023-03-31T07:13:51.3651094Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_namedtuple_return_api.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:51.363910] 2023-03-31T07:13:54.9850055Z 2023-03-31T07:13:54.9850632Z Expand the folded group to see the log file of test_namedtuple_return_api 2023-03-31T07:13:54.9852874Z ##[group]PRINTING LOG FILE of test_namedtuple_return_api (/var/lib/jenkins/workspace/test/test-reports/test_namedtuple_return_api_8xp0_qem.log) 2023-03-31T07:13:54.9853615Z Test results will be stored in test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-d91e6b677e7304dd.xml 2023-03-31T07:13:54.9853985Z ============================= test session starts ============================== 2023-03-31T07:13:54.9854479Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:54.9854760Z cachedir: .pytest_cache 2023-03-31T07:13:54.9855445Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:54.9855957Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:54.9856418Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:54.9856806Z collecting ... collected 3 items 2023-03-31T07:13:54.9857329Z Running 3 items in this shard: test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_import_return_types, test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_namedtuple_return, test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_native_functions_yaml 2023-03-31T07:13:54.9857831Z 2023-03-31T07:13:54.9858013Z test_namedtuple_return_api.py::TestNamedTupleAPI::test_import_return_types PASSED [ 33%] 2023-03-31T07:13:54.9858480Z test_namedtuple_return_api.py::TestNamedTupleAPI::test_namedtuple_return PASSED [ 66%] 2023-03-31T07:13:54.9858874Z test_namedtuple_return_api.py::TestNamedTupleAPI::test_native_functions_yaml PASSED [100%] 2023-03-31T07:13:54.9859106Z 2023-03-31T07:13:54.9859483Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-d91e6b677e7304dd.xml - 2023-03-31T07:13:54.9859950Z ============================== 3 passed in 1.74s =============================== 2023-03-31T07:13:54.9860269Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:54.9860542Z 2023-03-31T07:13:54.9860771Z ##[endgroup] 2023-03-31T07:13:54.9861366Z FINISHED PRINTING LOG FILE of test_namedtuple_return_api (/var/lib/jenkins/workspace/test/test-reports/test_namedtuple_return_api_8xp0_qem.log) 2023-03-31T07:13:54.9861624Z 2023-03-31T07:13:54.9861808Z Running test_fake_tensor ... [2023-03-31 07:13:54.985326] 2023-03-31T07:13:54.9862657Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_fake_tensor.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:54.985695] 2023-03-31T07:13:57.7511547Z 2023-03-31T07:13:57.7512099Z Expand the folded group to see the log file of test_fake_tensor 2023-03-31T07:13:57.7513174Z ##[group]PRINTING LOG FILE of test_fake_tensor (/var/lib/jenkins/workspace/test/test-reports/test_fake_tensor_tywhji31.log) 2023-03-31T07:13:57.7514280Z Test results will be stored in test-reports/python-pytest/test_fake_tensor/test_fake_tensor-76f6a1f7b24b7983.xml 2023-03-31T07:13:57.7514868Z ============================= test session starts ============================== 2023-03-31T07:13:57.7515516Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:13:57.7515915Z cachedir: .pytest_cache 2023-03-31T07:13:57.7516659Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:13:57.7517290Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:13:57.7517990Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:13:57.7518489Z collecting ... collected 64 items 2023-03-31T07:13:57.7527228Z Running 64 items in this shard: test/test_fake_tensor.py::FakeTensorTest::test_allow_meta, test/test_fake_tensor.py::FakeTensorTest::test_basic, test/test_fake_tensor.py::FakeTensorTest::test_binary_op_type_promotion, test/test_fake_tensor.py::FakeTensorTest::test_constructor, test/test_fake_tensor.py::FakeTensorTest::test_cpu_fallback, test/test_fake_tensor.py::FakeTensorTest::test_cuda_lstm, test/test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_with_fallback, test/test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_without_fallback, test/test_fake_tensor.py::FakeTensorTest::test_data_dependent_operator, test/test_fake_tensor.py::FakeTensorTest::test_deepcopy, test/test_fake_tensor.py::FakeTensorTest::test_fake_dispatch_keys, test/test_fake_tensor.py::FakeTensorTest::test_fake_grad_copy, test/test_fake_tensor.py::FakeTensorTest::test_fake_mode_error, test/test_fake_tensor.py::FakeTensorTest::test_fallback_memory_prop, test/test_fake_tensor.py::FakeTensorTest::test_from_numpy, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu, test/test_fake_tensor.py::FakeTensorTest::test_like_constructor, test/test_fake_tensor.py::FakeTensorTest::test_mode, test/test_fake_tensor.py::FakeTensorTest::test_nan_to_num, test/test_fake_tensor.py::FakeTensorTest::test_new, test/test_fake_tensor.py::FakeTensorTest::test_non_kwarg_device, test/test_fake_tensor.py::FakeTensorTest::test_non_overlapping_stride_zero, test/test_fake_tensor.py::FakeTensorTest::test_non_parameter_grad, test/test_fake_tensor.py::FakeTensorTest::test_normalize_device, test/test_fake_tensor.py::FakeTensorTest::test_parameter_instantiation, test/test_fake_tensor.py::FakeTensorTest::test_print_in_fake_mode, test/test_fake_tensor.py::FakeTensorTest::test_randperm, test/test_fake_tensor.py::FakeTensorTest::test_recursive_invocation, test/test_fake_tensor.py::FakeTensorTest::test_scalar_inputs, test/test_fake_tensor.py::FakeTensorTest::test_setitem, test/test_fake_tensor.py::FakeTensorTest::test_shape_take_not_device, test/test_fake_tensor.py::FakeTensorTest::test_throw, test/test_fake_tensor.py::FakeTensorTest::test_type_as, test/test_fake_tensor.py::FakeTensorTest::test_upsample_bilinear_small_channels, test/test_fake_tensor.py::FakeTensorTest::test_zero_dim, test/test_fake_tensor.py::FakeTensorConstHandling::test_aliased_const_write, test/test_fake_tensor.py::FakeTensorConstHandling::test_constant_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_constant_propagate_through_functions, test/test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_batch_norm_cpu, test/test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_in_intlist_repro, test/test_fake_tensor.py::FakeTensorConstHandling::test_inplace_add, test/test_fake_tensor.py::FakeTensorConstHandling::test_inplace_view_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_shared_storage_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_shared_storages, test/test_fake_tensor.py::FakeTensorConstHandling::test_simple, test/test_fake_tensor.py::FakeTensorConverterTest::test_dead_key, test/test_fake_tensor.py::FakeTensorConverterTest::test_dead_weak_ref, test/test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_from_meta, test/test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_to_meta, test/test_fake_tensor.py::FakeTensorConverterTest::test_no_active_mode, test/test_fake_tensor.py::FakeTensorConverterTest::test_no_ref_cycle, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_mode_error, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_non_view, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_view, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_conv_c1_backward, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_embedding_bag_private, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_like_ops, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_no_dispatch_with_like_function, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_non_kwarg_only_device, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_sparse_new, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_constructors_all_have_kwarg_device, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_new, test/test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module, test/test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module_with_optional_args 2023-03-31T07:13:57.7534593Z 2023-03-31T07:13:57.7534832Z test_fake_tensor.py::FakeTensorTest::test_allow_meta PASSED [ 1%] 2023-03-31T07:13:57.7535246Z test_fake_tensor.py::FakeTensorTest::test_basic PASSED [ 3%] 2023-03-31T07:13:57.7535880Z test_fake_tensor.py::FakeTensorTest::test_binary_op_type_promotion PASSED [ 4%] 2023-03-31T07:13:57.7536372Z test_fake_tensor.py::FakeTensorTest::test_constructor PASSED [ 6%] 2023-03-31T07:13:57.7536826Z test_fake_tensor.py::FakeTensorTest::test_cpu_fallback SKIPPED (requires cuda) [ 7%] 2023-03-31T07:13:57.7537273Z test_fake_tensor.py::FakeTensorTest::test_cuda_lstm SKIPPED (requires cuda) [ 9%] 2023-03-31T07:13:57.7537749Z test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_with_fallback SKIPPED (requires cuda) [ 10%] 2023-03-31T07:13:57.7538257Z test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_without_fallback SKIPPED (requires cuda) [ 12%] 2023-03-31T07:13:57.7538717Z test_fake_tensor.py::FakeTensorTest::test_data_dependent_operator PASSED [ 14%] 2023-03-31T07:13:57.7539153Z test_fake_tensor.py::FakeTensorTest::test_deepcopy PASSED [ 15%] 2023-03-31T07:13:57.7539575Z test_fake_tensor.py::FakeTensorTest::test_fake_dispatch_keys PASSED [ 17%] 2023-03-31T07:13:57.7540007Z test_fake_tensor.py::FakeTensorTest::test_fake_grad_copy PASSED [ 18%] 2023-03-31T07:13:57.7540416Z test_fake_tensor.py::FakeTensorTest::test_fake_mode_error PASSED [ 20%] 2023-03-31T07:13:57.7540869Z test_fake_tensor.py::FakeTensorTest::test_fallback_memory_prop SKIPPED (requires cuda) [ 21%] 2023-03-31T07:13:57.7541328Z test_fake_tensor.py::FakeTensorTest::test_from_numpy PASSED [ 23%] 2023-03-31T07:13:57.7541762Z test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu SKIPPED (requires cuda) [ 25%] 2023-03-31T07:13:57.7542546Z test_fake_tensor.py::FakeTensorTest::test_like_constructor SKIPPED (requires cuda) [ 26%] 2023-03-31T07:13:57.7543024Z test_fake_tensor.py::FakeTensorTest::test_mode PASSED [ 28%] 2023-03-31T07:13:57.7543611Z test_fake_tensor.py::FakeTensorTest::test_nan_to_num PASSED [ 29%] 2023-03-31T07:13:57.7544027Z test_fake_tensor.py::FakeTensorTest::test_new SKIPPED (requires cuda) [ 31%] 2023-03-31T07:13:57.7544489Z test_fake_tensor.py::FakeTensorTest::test_non_kwarg_device SKIPPED (requires cuda) [ 32%] 2023-03-31T07:13:57.7544968Z test_fake_tensor.py::FakeTensorTest::test_non_overlapping_stride_zero PASSED [ 34%] 2023-03-31T07:13:57.7545422Z test_fake_tensor.py::FakeTensorTest::test_non_parameter_grad PASSED [ 35%] 2023-03-31T07:13:57.7545864Z test_fake_tensor.py::FakeTensorTest::test_normalize_device SKIPPED (requires cuda) [ 37%] 2023-03-31T07:13:57.7546327Z test_fake_tensor.py::FakeTensorTest::test_parameter_instantiation PASSED [ 39%] 2023-03-31T07:13:57.7546781Z test_fake_tensor.py::FakeTensorTest::test_print_in_fake_mode PASSED [ 40%] 2023-03-31T07:13:57.7547248Z test_fake_tensor.py::FakeTensorTest::test_randperm PASSED [ 42%] 2023-03-31T07:13:57.7547663Z test_fake_tensor.py::FakeTensorTest::test_recursive_invocation PASSED [ 43%] 2023-03-31T07:13:57.7548086Z test_fake_tensor.py::FakeTensorTest::test_scalar_inputs PASSED [ 45%] 2023-03-31T07:13:57.7548524Z test_fake_tensor.py::FakeTensorTest::test_setitem SKIPPED (requires cuda) [ 46%] 2023-03-31T07:13:57.7548983Z test_fake_tensor.py::FakeTensorTest::test_shape_take_not_device SKIPPED (requires cuda) [ 48%] 2023-03-31T07:13:57.7549441Z test_fake_tensor.py::FakeTensorTest::test_throw SKIPPED (requires cuda) [ 50%] 2023-03-31T07:13:57.7549878Z test_fake_tensor.py::FakeTensorTest::test_type_as SKIPPED (requires cuda) [ 51%] 2023-03-31T07:13:57.7550365Z test_fake_tensor.py::FakeTensorTest::test_upsample_bilinear_small_channels SKIPPED (requires cuda) [ 53%] 2023-03-31T07:13:57.7550839Z test_fake_tensor.py::FakeTensorTest::test_zero_dim SKIPPED (requires cuda) [ 54%] 2023-03-31T07:13:57.7551307Z test_fake_tensor.py::FakeTensorConstHandling::test_aliased_const_write PASSED [ 56%] 2023-03-31T07:13:57.7551797Z test_fake_tensor.py::FakeTensorConstHandling::test_constant_invalidation PASSED [ 57%] 2023-03-31T07:13:57.7552297Z test_fake_tensor.py::FakeTensorConstHandling::test_constant_propagate_through_functions PASSED [ 59%] 2023-03-31T07:13:57.7553021Z test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_batch_norm_cpu PASSED [ 60%] 2023-03-31T07:13:57.7553581Z test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_in_intlist_repro PASSED [ 62%] 2023-03-31T07:13:57.7554062Z test_fake_tensor.py::FakeTensorConstHandling::test_inplace_add PASSED [ 64%] 2023-03-31T07:13:57.7554528Z test_fake_tensor.py::FakeTensorConstHandling::test_inplace_view_invalidation PASSED [ 65%] 2023-03-31T07:13:57.7555029Z test_fake_tensor.py::FakeTensorConstHandling::test_shared_storage_invalidation PASSED [ 67%] 2023-03-31T07:13:57.7555525Z test_fake_tensor.py::FakeTensorConstHandling::test_shared_storages PASSED [ 68%] 2023-03-31T07:13:57.7555981Z test_fake_tensor.py::FakeTensorConstHandling::test_simple PASSED [ 70%] 2023-03-31T07:13:57.7556430Z test_fake_tensor.py::FakeTensorConverterTest::test_dead_key PASSED [ 71%] 2023-03-31T07:13:57.7556872Z test_fake_tensor.py::FakeTensorConverterTest::test_dead_weak_ref PASSED [ 73%] 2023-03-31T07:13:57.7557367Z test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_from_meta PASSED [ 75%] 2023-03-31T07:13:57.7557859Z test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_to_meta PASSED [ 76%] 2023-03-31T07:13:57.7558345Z test_fake_tensor.py::FakeTensorConverterTest::test_no_active_mode PASSED [ 78%] 2023-03-31T07:13:57.7558808Z test_fake_tensor.py::FakeTensorConverterTest::test_no_ref_cycle PASSED [ 79%] 2023-03-31T07:13:57.7559260Z test_fake_tensor.py::FakeTensorConverterTest::test_separate_mode_error PASSED [ 81%] 2023-03-31T07:13:57.7559817Z test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_non_view PASSED [ 82%] 2023-03-31T07:13:57.7560365Z test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_view PASSED [ 84%] 2023-03-31T07:13:57.7561099Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_conv_c1_backward SKIPPED (requires cuda) [ 85%] 2023-03-31T07:13:57.7561635Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_embedding_bag_private PASSED [ 87%] 2023-03-31T07:13:57.7562153Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_like_ops PASSED [ 89%] 2023-03-31T07:13:57.7562710Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_no_dispatch_with_like_function PASSED [ 90%] 2023-03-31T07:13:57.7563414Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_non_kwarg_only_device PASSED [ 92%] 2023-03-31T07:13:57.7563891Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_sparse_new XFAIL [ 93%] 2023-03-31T07:13:57.7564439Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_constructors_all_have_kwarg_device PASSED [ 95%] 2023-03-31T07:13:57.7565010Z test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_new PASSED [ 96%] 2023-03-31T07:13:57.7565479Z test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module PASSED [ 98%] 2023-03-31T07:13:57.7565998Z test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module_with_optional_args PASSED [100%] 2023-03-31T07:13:57.7566291Z 2023-03-31T07:13:57.7566880Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fake_tensor/test_fake_tensor-76f6a1f7b24b7983.xml - 2023-03-31T07:13:57.7567412Z ================== 46 passed, 17 skipped, 1 xfailed in 0.67s =================== 2023-03-31T07:13:57.7567829Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:13:57.7568097Z 2023-03-31T07:13:57.7568503Z ##[endgroup] 2023-03-31T07:13:57.7569068Z FINISHED PRINTING LOG FILE of test_fake_tensor (/var/lib/jenkins/workspace/test/test-reports/test_fake_tensor_tywhji31.log) 2023-03-31T07:13:57.7569386Z 2023-03-31T07:13:57.7569596Z Running test_prims ... [2023-03-31 07:13:57.751541] 2023-03-31T07:13:57.7570607Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_prims.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:13:57.751945] 2023-03-31T07:14:01.2391357Z 2023-03-31T07:14:01.2391951Z Expand the folded group to see the log file of test_prims 2023-03-31T07:14:01.2393155Z ##[group]PRINTING LOG FILE of test_prims (/var/lib/jenkins/workspace/test/test-reports/test_prims_qh6t5u4_.log) 2023-03-31T07:14:01.2393877Z Test results will be stored in test-reports/python-pytest/test_prims/test_prims-16c3db175ecf734a.xml 2023-03-31T07:14:01.2394217Z ============================= test session starts ============================== 2023-03-31T07:14:01.2394626Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:01.2394926Z cachedir: .pytest_cache 2023-03-31T07:14:01.2395353Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:01.2395732Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:01.2396243Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:01.2396598Z collecting ... collected 58 items 2023-03-31T07:14:01.2402026Z Running 58 items in this shard: test/test_prims.py::TestPrimsBasic::test_mul_complex, test/test_prims.py::TestPrimsBasic::test_torch_ops, test/test_prims.py::TestPrimsCPU::test_aten_overload_to_prims_cpu, test/test_prims.py::TestPrimsCPU::test_batch_norm_backward_nvprims_cpu_float16, test/test_prims.py::TestPrimsCPU::test_batch_norm_backward_nvprims_cpu_float32, test/test_prims.py::TestPrimsCPU::test_broadcast_in_dim_cpu_float32, test/test_prims.py::TestPrimsCPU::test_broadcast_in_dim_sum_cpu_float32, test/test_prims.py::TestPrimsCPU::test_cbrt_prim_cpu_float64, test/test_prims.py::TestPrimsCPU::test_cbrt_prim_cpu_int64, test/test_prims.py::TestPrimsCPU::test_collapse_cpu_float32, test/test_prims.py::TestPrimsCPU::test_cpu_tensor_cpu_float16, test/test_prims.py::TestPrimsCPU::test_cpu_tensor_cpu_float32, test/test_prims.py::TestPrimsCPU::test_cudnn_batch_norm_nvprims_cpu_float32, test/test_prims.py::TestPrimsCPU::test_cudnn_batch_norm_nvprims_cpu_float64, test/test_prims.py::TestPrimsCPU::test_full_cpu_float32, test/test_prims.py::TestPrimsCPU::test_memory_format_strides_cpu_float32, test/test_prims.py::TestPrimsCPU::test_native_batch_norm_nvprims_cpu_float32, test/test_prims.py::TestPrimsCPU::test_native_batch_norm_nvprims_cpu_float64, test/test_prims.py::TestPrimsCPU::test_nvfuser_capability_context_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_constant_tensors_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_empty_fusion_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_executor_cached_noncontiguous_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_executor_parameters_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_executor_partitioned_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_executor_partitioned_no_partitions_error_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_impl_is_used_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_no_args_cpu, test/test_prims.py::TestPrimsCPU::test_nvfuser_rand_like_fusion_cpu, test/test_prims.py::TestPrimsCPU::test_nvprim_convert_element_type_cpu_float16, test/test_prims.py::TestPrimsCPU::test_nvprim_convert_element_type_cpu_uint8, test/test_prims.py::TestPrimsCPU::test_nvprims_cpu, test/test_prims.py::TestPrimsCPU::test_nvprims_view_cpu_float16, test/test_prims.py::TestPrimsCPU::test_nvprims_view_cpu_float32, test/test_prims.py::TestPrimsCPU::test_nvprims_view_partitioner_cpu_float16, test/test_prims.py::TestPrimsCPU::test_nvprims_view_partitioner_cpu_float32, test/test_prims.py::TestPrimsCPU::test_partitioner_tuple_output_cpu, test/test_prims.py::TestPrimsCPU::test_pytree_input_output_cpu_float32, test/test_prims.py::TestPrimsCPU::test_reshape_view_method_cpu_float32, test/test_prims.py::TestPrimsCPU::test_silu_backward_no_filled_tensor_cpu_float32, test/test_prims.py::TestPrimsCPU::test_skip_ops_nvfuser_capability_mode_cpu, test/test_prims.py::TestPrimsCPU::test_skip_ops_nvfuser_prims_mode_cpu, test/test_prims.py::TestPrimsCPU::test_var_correction_0_cpu_float32, test/test_prims.py::TestPrimsCPU::test_var_correction_1_cpu_float32, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_False_cpu_float16, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_False_cpu_float32, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_True_cpu_float16, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_True_cpu_float32, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_False_cpu_float16, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_False_cpu_float32, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_True_cpu_float16, test/test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_True_cpu_float32, test/test_prims.py::TestRefsCPU::test_constant_pad_nd_memory_format_cpu_float32, test/test_prims.py::TestDecompCPU::test_decomposition_method_vararg_ones_cpu_float32, test/test_prims.py::TestDecompCPU::test_decomposition_method_vararg_permute_cpu_float32, test/test_prims.py::TestDecompCPU::test_decomposition_type_promotion_nvprim_amp_cpu_float16, test/test_prims.py::TestDecompCPU::test_decomposition_type_promotion_nvprim_amp_cpu_float32, test/test_prims.py::TestDecompCPU::test_masked_fill_decomposition_under_nvprim_context_cpu_float16, test/test_prims.py::TestDecompCPU::test_masked_fill_decomposition_under_nvprim_context_cpu_float32 2023-03-31T07:14:01.2407089Z 2023-03-31T07:14:01.2407251Z test_prims.py::TestPrimsBasic::test_mul_complex PASSED [ 1%] 2023-03-31T07:14:01.2407630Z test_prims.py::TestPrimsBasic::test_torch_ops PASSED [ 3%] 2023-03-31T07:14:01.2407943Z test_prims.py::TestPrimsCPU::test_aten_overload_to_prims_cpu PASSED [ 5%] 2023-03-31T07:14:01.2408352Z test_prims.py::TestPrimsCPU::test_batch_norm_backward_nvprims_cpu_float16 SKIPPED (Only runs on cuda) [ 6%] 2023-03-31T07:14:01.2408764Z test_prims.py::TestPrimsCPU::test_batch_norm_backward_nvprims_cpu_float32 SKIPPED (Only runs on cuda) [ 8%] 2023-03-31T07:14:01.2409148Z test_prims.py::TestPrimsCPU::test_broadcast_in_dim_cpu_float32 SKIPPED (Only runs on cuda) [ 10%] 2023-03-31T07:14:01.2409512Z test_prims.py::TestPrimsCPU::test_broadcast_in_dim_sum_cpu_float32 SKIPPED (Only runs on cuda) [ 12%] 2023-03-31T07:14:01.2409858Z test_prims.py::TestPrimsCPU::test_cbrt_prim_cpu_float64 PASSED [ 13%] 2023-03-31T07:14:01.2410174Z test_prims.py::TestPrimsCPU::test_cbrt_prim_cpu_int64 PASSED [ 15%] 2023-03-31T07:14:01.2410494Z test_prims.py::TestPrimsCPU::test_collapse_cpu_float32 PASSED [ 17%] 2023-03-31T07:14:01.2410822Z test_prims.py::TestPrimsCPU::test_cpu_tensor_cpu_float16 SKIPPED (Only runs on cuda) [ 18%] 2023-03-31T07:14:01.2411175Z test_prims.py::TestPrimsCPU::test_cpu_tensor_cpu_float32 SKIPPED (Only runs on cuda) [ 20%] 2023-03-31T07:14:01.2411542Z test_prims.py::TestPrimsCPU::test_cudnn_batch_norm_nvprims_cpu_float32 SKIPPED (Only runs on cuda) [ 22%] 2023-03-31T07:14:01.2411915Z test_prims.py::TestPrimsCPU::test_cudnn_batch_norm_nvprims_cpu_float64 SKIPPED (Only runs on cuda) [ 24%] 2023-03-31T07:14:01.2412274Z test_prims.py::TestPrimsCPU::test_full_cpu_float32 SKIPPED (Only runs on cuda) [ 25%] 2023-03-31T07:14:01.2412620Z test_prims.py::TestPrimsCPU::test_memory_format_strides_cpu_float32 PASSED [ 27%] 2023-03-31T07:14:01.2412993Z test_prims.py::TestPrimsCPU::test_native_batch_norm_nvprims_cpu_float32 SKIPPED (Only runs on cuda) [ 29%] 2023-03-31T07:14:01.2413369Z test_prims.py::TestPrimsCPU::test_native_batch_norm_nvprims_cpu_float64 SKIPPED (Only runs on cuda) [ 31%] 2023-03-31T07:14:01.2413735Z test_prims.py::TestPrimsCPU::test_nvfuser_capability_context_cpu PASSED [ 32%] 2023-03-31T07:14:01.2414096Z test_prims.py::TestPrimsCPU::test_nvfuser_constant_tensors_cpu SKIPPED (Only runs on cuda) [ 34%] 2023-03-31T07:14:01.2414445Z test_prims.py::TestPrimsCPU::test_nvfuser_empty_fusion_cpu SKIPPED (Only runs on cuda) [ 36%] 2023-03-31T07:14:01.2414886Z test_prims.py::TestPrimsCPU::test_nvfuser_executor_cached_noncontiguous_cpu SKIPPED (Only runs on cuda) [ 37%] 2023-03-31T07:14:01.2415321Z test_prims.py::TestPrimsCPU::test_nvfuser_executor_parameters_cpu SKIPPED (Only runs on cuda) [ 39%] 2023-03-31T07:14:01.2415702Z test_prims.py::TestPrimsCPU::test_nvfuser_executor_partitioned_cpu SKIPPED (Only runs on cuda) [ 41%] 2023-03-31T07:14:01.2416100Z test_prims.py::TestPrimsCPU::test_nvfuser_executor_partitioned_no_partitions_error_cpu SKIPPED (Only runs on cuda) [ 43%] 2023-03-31T07:14:01.2416495Z test_prims.py::TestPrimsCPU::test_nvfuser_impl_is_used_cpu SKIPPED (Only runs on cuda) [ 44%] 2023-03-31T07:14:01.2416850Z test_prims.py::TestPrimsCPU::test_nvfuser_no_args_cpu SKIPPED (Only runs on cuda) [ 46%] 2023-03-31T07:14:01.2417212Z test_prims.py::TestPrimsCPU::test_nvfuser_rand_like_fusion_cpu SKIPPED (Only runs on cuda) [ 48%] 2023-03-31T07:14:01.2417582Z test_prims.py::TestPrimsCPU::test_nvprim_convert_element_type_cpu_float16 SKIPPED (Only runs on cuda) [ 50%] 2023-03-31T07:14:01.2417978Z test_prims.py::TestPrimsCPU::test_nvprim_convert_element_type_cpu_uint8 SKIPPED (Only runs on cuda) [ 51%] 2023-03-31T07:14:01.2418328Z test_prims.py::TestPrimsCPU::test_nvprims_cpu PASSED [ 53%] 2023-03-31T07:14:01.2418649Z test_prims.py::TestPrimsCPU::test_nvprims_view_cpu_float16 SKIPPED (Only runs on cuda) [ 55%] 2023-03-31T07:14:01.2419004Z test_prims.py::TestPrimsCPU::test_nvprims_view_cpu_float32 SKIPPED (Only runs on cuda) [ 56%] 2023-03-31T07:14:01.2419374Z test_prims.py::TestPrimsCPU::test_nvprims_view_partitioner_cpu_float16 SKIPPED (Only runs on cuda) [ 58%] 2023-03-31T07:14:01.2419789Z test_prims.py::TestPrimsCPU::test_nvprims_view_partitioner_cpu_float32 SKIPPED (Only runs on cuda) [ 60%] 2023-03-31T07:14:01.2420164Z test_prims.py::TestPrimsCPU::test_partitioner_tuple_output_cpu PASSED [ 62%] 2023-03-31T07:14:01.2420524Z test_prims.py::TestPrimsCPU::test_pytree_input_output_cpu_float32 SKIPPED (Only runs on cuda) [ 63%] 2023-03-31T07:14:01.2420881Z test_prims.py::TestPrimsCPU::test_reshape_view_method_cpu_float32 PASSED [ 65%] 2023-03-31T07:14:01.2421257Z test_prims.py::TestPrimsCPU::test_silu_backward_no_filled_tensor_cpu_float32 SKIPPED (Only runs on cuda) [ 67%] 2023-03-31T07:14:01.2421619Z test_prims.py::TestPrimsCPU::test_skip_ops_nvfuser_capability_mode_cpu PASSED [ 68%] 2023-03-31T07:14:01.2421966Z test_prims.py::TestPrimsCPU::test_skip_ops_nvfuser_prims_mode_cpu PASSED [ 70%] 2023-03-31T07:14:01.2422400Z test_prims.py::TestPrimsCPU::test_var_correction_0_cpu_float32 SKIPPED (Only runs on cuda) [ 72%] 2023-03-31T07:14:01.2422757Z test_prims.py::TestPrimsCPU::test_var_correction_1_cpu_float32 SKIPPED (Only runs on cuda) [ 74%] 2023-03-31T07:14:01.2423153Z test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_False_cpu_float16 SKIPPED (Only runs on cuda) [ 75%] 2023-03-31T07:14:01.2423569Z test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_False_cpu_float32 SKIPPED (Only runs on cuda) [ 77%] 2023-03-31T07:14:01.2423981Z test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_True_cpu_float16 SKIPPED (Only runs on cuda) [ 79%] 2023-03-31T07:14:01.2424376Z test_prims.py::TestPrimsCPU::test_var_mean_correction_0_keepdim_True_cpu_float32 SKIPPED (Only runs on cuda) [ 81%] 2023-03-31T07:14:01.2424781Z test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_False_cpu_float16 SKIPPED (Only runs on cuda) [ 82%] 2023-03-31T07:14:01.2425189Z test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_False_cpu_float32 SKIPPED (Only runs on cuda) [ 84%] 2023-03-31T07:14:01.2425594Z test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_True_cpu_float16 SKIPPED (Only runs on cuda) [ 86%] 2023-03-31T07:14:01.2425985Z test_prims.py::TestPrimsCPU::test_var_mean_correction_1_keepdim_True_cpu_float32 SKIPPED (Only runs on cuda) [ 87%] 2023-03-31T07:14:01.2426364Z test_prims.py::TestRefsCPU::test_constant_pad_nd_memory_format_cpu_float32 PASSED [ 89%] 2023-03-31T07:14:01.2426769Z test_prims.py::TestDecompCPU::test_decomposition_method_vararg_ones_cpu_float32 PASSED [ 91%] 2023-03-31T07:14:01.2427184Z test_prims.py::TestDecompCPU::test_decomposition_method_vararg_permute_cpu_float32 PASSED [ 93%] 2023-03-31T07:14:01.2427580Z test_prims.py::TestDecompCPU::test_decomposition_type_promotion_nvprim_amp_cpu_float16 SKIPPED (Only runs on cuda) [ 94%] 2023-03-31T07:14:01.2428002Z test_prims.py::TestDecompCPU::test_decomposition_type_promotion_nvprim_amp_cpu_float32 SKIPPED (Only runs on cuda) [ 96%] 2023-03-31T07:14:01.2428434Z test_prims.py::TestDecompCPU::test_masked_fill_decomposition_under_nvprim_context_cpu_float16 SKIPPED (Only runs on cuda) [ 98%] 2023-03-31T07:14:01.2428873Z test_prims.py::TestDecompCPU::test_masked_fill_decomposition_under_nvprim_context_cpu_float32 SKIPPED (Only runs on cuda) [100%] 2023-03-31T07:14:01.2429097Z 2023-03-31T07:14:01.2429499Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_prims/test_prims-16c3db175ecf734a.xml - 2023-03-31T07:14:01.2429866Z ======================== 16 passed, 42 skipped in 0.52s ======================== 2023-03-31T07:14:01.2430197Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:01.2430404Z 2023-03-31T07:14:01.2430687Z ##[endgroup] 2023-03-31T07:14:01.2431084Z FINISHED PRINTING LOG FILE of test_prims (/var/lib/jenkins/workspace/test/test-reports/test_prims_qh6t5u4_.log) 2023-03-31T07:14:01.2431306Z 2023-03-31T07:14:01.2431543Z Running functorch/test_memory_efficient_fusion ... [2023-03-31 07:14:01.239510] 2023-03-31T07:14:01.2432298Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'functorch/test_memory_efficient_fusion.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:01.239845] 2023-03-31T07:14:03.7459315Z 2023-03-31T07:14:03.7459882Z Expand the folded group to see the log file of functorch/test_memory_efficient_fusion 2023-03-31T07:14:03.7460885Z ##[group]PRINTING LOG FILE of functorch/test_memory_efficient_fusion (/var/lib/jenkins/workspace/test/test-reports/functorch-test_memory_efficient_fusion_e8qaj1pn.log) 2023-03-31T07:14:03.7461701Z Test results will be stored in test-reports/python-pytest/functorch.test_memory_efficient_fusion/functorch.test_memory_efficient_fusion-a953279d56a60dbd.xml 2023-03-31T07:14:03.7462079Z ============================= test session starts ============================== 2023-03-31T07:14:03.7462578Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:03.7462939Z cachedir: .pytest_cache 2023-03-31T07:14:03.7463417Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:03.7463897Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:03.7464402Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:03.7464715Z collecting ... collected 21 items 2023-03-31T07:14:03.7467327Z Running 21 items in this shard: test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_gelu_bias, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_hard_sigmoid, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_hard_swish, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_layer_norm, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_mish, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_rmsnorm, test/functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_swish, test/functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_empty, test/functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_nochange, test/functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_rand_like, test/functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_rand_n, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_immutable_list_multiple_entries, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_immutable_list_type, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_kwarg, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_nested_immutable_list_type, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple_2, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple_multiple_same_ops, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_two_args, test/functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_two_args_default, test/functorch/test_memory_efficient_fusion.py::RandomOpTestCase::test_random 2023-03-31T07:14:03.7469640Z 2023-03-31T07:14:03.7469886Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_gelu_bias SKIPPED (CUDA is unavailable) [ 4%] 2023-03-31T07:14:03.7470358Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_hard_sigmoid SKIPPED (CUDA is unavailable) [ 9%] 2023-03-31T07:14:03.7470837Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_hard_swish SKIPPED (CUDA is unavailable) [ 14%] 2023-03-31T07:14:03.7471309Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_layer_norm SKIPPED (CUDA is unavailable) [ 19%] 2023-03-31T07:14:03.7471829Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_mish SKIPPED (CUDA is unavailable) [ 23%] 2023-03-31T07:14:03.7472315Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_rmsnorm SKIPPED (CUDA is unavailable) [ 28%] 2023-03-31T07:14:03.7472786Z functorch/test_memory_efficient_fusion.py::TestMemoryEfficientOpAuthoring::test_swish SKIPPED (CUDA is unavailable) [ 33%] 2023-03-31T07:14:03.7473195Z functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_empty PASSED [ 38%] 2023-03-31T07:14:03.7473568Z functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_nochange PASSED [ 42%] 2023-03-31T07:14:03.7473920Z functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_rand_like PASSED [ 47%] 2023-03-31T07:14:03.7474281Z functorch/test_memory_efficient_fusion.py::NoChangeTestCase::test_rand_n PASSED [ 52%] 2023-03-31T07:14:03.7474668Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_immutable_list_multiple_entries PASSED [ 57%] 2023-03-31T07:14:03.7475053Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_immutable_list_type PASSED [ 61%] 2023-03-31T07:14:03.7475419Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_kwarg PASSED [ 66%] 2023-03-31T07:14:03.7475796Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_nested_immutable_list_type PASSED [ 71%] 2023-03-31T07:14:03.7476174Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple PASSED [ 76%] 2023-03-31T07:14:03.7476516Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple_2 PASSED [ 80%] 2023-03-31T07:14:03.7476889Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_simple_multiple_same_ops PASSED [ 85%] 2023-03-31T07:14:03.7477260Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_two_args PASSED [ 90%] 2023-03-31T07:14:03.7477622Z functorch/test_memory_efficient_fusion.py::ReduceTestCase::test_two_args_default PASSED [ 95%] 2023-03-31T07:14:03.7477975Z functorch/test_memory_efficient_fusion.py::RandomOpTestCase::test_random PASSED [100%] 2023-03-31T07:14:03.7478171Z 2023-03-31T07:14:03.7478610Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_memory_efficient_fusion/functorch.test_memory_efficient_fusion-a953279d56a60dbd.xml - 2023-03-31T07:14:03.7479065Z ======================== 14 passed, 7 skipped in 0.46s ========================= 2023-03-31T07:14:03.7479424Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:03.7479617Z 2023-03-31T07:14:03.7479881Z ##[endgroup] 2023-03-31T07:14:03.7480378Z FINISHED PRINTING LOG FILE of functorch/test_memory_efficient_fusion (/var/lib/jenkins/workspace/test/test-reports/functorch-test_memory_efficient_fusion_e8qaj1pn.log) 2023-03-31T07:14:03.7480659Z 2023-03-31T07:14:03.7480845Z Running test_show_pickle ... [2023-03-31 07:14:03.746228] 2023-03-31T07:14:03.7481458Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_show_pickle.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:03.746551] 2023-03-31T07:14:05.8857497Z 2023-03-31T07:14:05.8858060Z Expand the folded group to see the log file of test_show_pickle 2023-03-31T07:14:05.8859182Z ##[group]PRINTING LOG FILE of test_show_pickle (/var/lib/jenkins/workspace/test/test-reports/test_show_pickle_l6jmbv57.log) 2023-03-31T07:14:05.8860142Z Test results will be stored in test-reports/python-pytest/test_show_pickle/test_show_pickle-5361e547855f3713.xml 2023-03-31T07:14:05.8860487Z ============================= test session starts ============================== 2023-03-31T07:14:05.8860957Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:05.8861242Z cachedir: .pytest_cache 2023-03-31T07:14:05.8861738Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:05.8862507Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:05.8863015Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:05.8863328Z collecting ... collected 1 item 2023-03-31T07:14:05.8863636Z Running 1 items in this shard: test/test_show_pickle.py::TestShowPickle::test_scripted_model 2023-03-31T07:14:05.8863840Z 2023-03-31T07:14:05.8863984Z test_show_pickle.py::TestShowPickle::test_scripted_model PASSED [100%] 2023-03-31T07:14:05.8864168Z 2023-03-31T07:14:05.8864512Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_show_pickle/test_show_pickle-5361e547855f3713.xml - 2023-03-31T07:14:05.8864880Z ============================== 1 passed in 0.29s =============================== 2023-03-31T07:14:05.8865199Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:05.8865404Z 2023-03-31T07:14:05.8865635Z ##[endgroup] 2023-03-31T07:14:05.8866046Z FINISHED PRINTING LOG FILE of test_show_pickle (/var/lib/jenkins/workspace/test/test-reports/test_show_pickle_l6jmbv57.log) 2023-03-31T07:14:05.8866282Z 2023-03-31T07:14:05.8866636Z Running test_cpp_extensions_aot_ninja ... [2023-03-31 07:14:05.886045] 2023-03-31T07:14:07.4858176Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:14:07.6113182Z running install 2023-03-31T07:14:07.6121138Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. 2023-03-31T07:14:07.6121566Z warnings.warn( 2023-03-31T07:14:07.6227948Z running build 2023-03-31T07:14:07.6228200Z running build_py 2023-03-31T07:14:07.6293375Z creating build 2023-03-31T07:14:07.6293886Z creating build/lib.linux-x86_64-cpython-38 2023-03-31T07:14:07.6294408Z creating build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension 2023-03-31T07:14:07.6295134Z copying torch_test_cpp_extension/__init__.py -> build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension 2023-03-31T07:14:07.6297161Z running build_ext 2023-03-31T07:14:07.6724855Z building 'torch_test_cpp_extension.cpp' extension 2023-03-31T07:14:07.6725666Z creating /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38 2023-03-31T07:14:07.7105511Z Emitting ninja build file /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/build.ninja... 2023-03-31T07:14:07.7106691Z Compiling objects... 2023-03-31T07:14:07.7106974Z Using envvar MAX_JOBS (6) as the number of workers... 2023-03-31T07:14:08.8768377Z [1/1] c++ -MMD -MF /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/extension.o.d -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/var/lib/jenkins/workspace/test/cpp_extensions/self_compiler_include_dirs_test -I/opt/conda/envs/py_3.8/include/python3.8 -c -c /var/lib/jenkins/workspace/test/cpp_extensions/extension.cpp -o /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/extension.o -g -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_clang"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1002"' -DTORCH_EXTENSION_NAME=cpp -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++17 2023-03-31T07:14:08.8772106Z cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ 2023-03-31T07:14:08.8772649Z In file included from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/Exceptions.h:14:0, 2023-03-31T07:14:08.8773388Z from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include/torch/python.h:11, 2023-03-31T07:14:08.8773932Z from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/extension.h:6, 2023-03-31T07:14:08.8774293Z from /var/lib/jenkins/workspace/test/cpp_extensions/extension.cpp:1: 2023-03-31T07:14:08.8774947Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/pybind11/pybind11.h: In instantiation of ‘class pybind11::class_’: 2023-03-31T07:14:08.8775356Z /var/lib/jenkins/workspace/test/cpp_extensions/extension.cpp:40:53: required from here 2023-03-31T07:14:08.8776140Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/pybind11/pybind11.h:1479:7: warning: ‘pybind11::class_’ declared with greater visibility than the type of its field ‘pybind11::class_::’ [-Wattributes] 2023-03-31T07:14:08.8776628Z class class_ : public detail::generic_type { 2023-03-31T07:14:08.8776837Z ^~~~~~ 2023-03-31T07:14:08.8777468Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/pybind11/pybind11.h:1479:7: warning: ‘pybind11::class_’ declared with greater visibility than its base ‘pybind11::detail::generic_type’ [-Wattributes] 2023-03-31T07:14:08.8870381Z g++ -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -pthread -shared -B /opt/conda/envs/py_3.8/compiler_compat -L/opt/conda/envs/py_3.8/lib -Wl,-rpath=/opt/conda/envs/py_3.8/lib -Wl,--no-as-needed -Wl,--sysroot=/ /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/extension.o -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch -ltorch_cpu -ltorch_python -o build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/cpp.cpython-38-x86_64-linux-gnu.so 2023-03-31T07:14:09.3398610Z building 'torch_test_cpp_extension.ort' extension 2023-03-31T07:14:09.3792921Z Emitting ninja build file /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/build.ninja... 2023-03-31T07:14:09.3798977Z Compiling objects... 2023-03-31T07:14:09.3799334Z Using envvar MAX_JOBS (6) as the number of workers... 2023-03-31T07:14:10.6300258Z [1/1] c++ -MMD -MF /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/ort_extension.o.d -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/var/lib/jenkins/workspace/test/cpp_extensions/self_compiler_include_dirs_test -I/opt/conda/envs/py_3.8/include/python3.8 -c -c /var/lib/jenkins/workspace/test/cpp_extensions/ort_extension.cpp -o /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/ort_extension.o -g -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_clang"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1002"' -DTORCH_EXTENSION_NAME=ort -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++17 2023-03-31T07:14:10.6302175Z cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ 2023-03-31T07:14:10.6356704Z g++ -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -pthread -shared -B /opt/conda/envs/py_3.8/compiler_compat -L/opt/conda/envs/py_3.8/lib -Wl,-rpath=/opt/conda/envs/py_3.8/lib -Wl,--no-as-needed -Wl,--sysroot=/ /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/ort_extension.o -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch -ltorch_cpu -ltorch_python -o build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/ort.cpython-38-x86_64-linux-gnu.so 2023-03-31T07:14:11.0363980Z building 'torch_test_cpp_extension.rng' extension 2023-03-31T07:14:11.0749509Z Emitting ninja build file /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/build.ninja... 2023-03-31T07:14:11.0750721Z Compiling objects... 2023-03-31T07:14:11.0751010Z Using envvar MAX_JOBS (6) as the number of workers... 2023-03-31T07:14:12.4022274Z [1/1] c++ -MMD -MF /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/rng_extension.o.d -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/var/lib/jenkins/workspace/test/cpp_extensions/self_compiler_include_dirs_test -I/opt/conda/envs/py_3.8/include/python3.8 -c -c /var/lib/jenkins/workspace/test/cpp_extensions/rng_extension.cpp -o /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/rng_extension.o -g -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_clang"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1002"' -DTORCH_EXTENSION_NAME=rng -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++17 2023-03-31T07:14:12.4024096Z cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ 2023-03-31T07:14:12.4024634Z In file included from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/ATen/cpu/vec/vec256/vec256.h:8:0, 2023-03-31T07:14:12.4025120Z from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/ATen/cpu/vec/vec.h:6, 2023-03-31T07:14:12.4025591Z from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/ATen/native/cpu/Loops.h:37, 2023-03-31T07:14:12.4026093Z from /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/ATen/native/cpu/DistributionTemplates.h:8, 2023-03-31T07:14:12.4026478Z from /var/lib/jenkins/workspace/test/cpp_extensions/rng_extension.cpp:6: 2023-03-31T07:14:12.4027016Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/ATen/cpu/vec/vec_base.h:1025:0: warning: ignoring #pragma unroll [-Wunknown-pragmas] 2023-03-31T07:14:12.4027501Z # pragma unroll 2023-03-31T07:14:12.4027733Z 2023-03-31T07:14:12.4077606Z g++ -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -pthread -shared -B /opt/conda/envs/py_3.8/compiler_compat -L/opt/conda/envs/py_3.8/lib -Wl,-rpath=/opt/conda/envs/py_3.8/lib -Wl,--no-as-needed -Wl,--sysroot=/ /var/lib/jenkins/workspace/test/cpp_extensions/build/temp.linux-x86_64-cpython-38/rng_extension.o -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch -ltorch_cpu -ltorch_python -o build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/rng.cpython-38-x86_64-linux-gnu.so 2023-03-31T07:14:12.8098418Z running install_lib 2023-03-31T07:14:12.8159348Z creating install 2023-03-31T07:14:12.8159595Z creating install/opt 2023-03-31T07:14:12.8159889Z creating install/opt/conda 2023-03-31T07:14:12.8160259Z creating install/opt/conda/envs 2023-03-31T07:14:12.8160940Z creating install/opt/conda/envs/py_3.8 2023-03-31T07:14:12.8161570Z creating install/opt/conda/envs/py_3.8/lib 2023-03-31T07:14:12.8162395Z creating install/opt/conda/envs/py_3.8/lib/python3.8 2023-03-31T07:14:12.8162940Z creating install/opt/conda/envs/py_3.8/lib/python3.8/site-packages 2023-03-31T07:14:12.8164382Z creating install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension 2023-03-31T07:14:12.8165125Z copying build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/__init__.py -> ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension 2023-03-31T07:14:12.8166750Z copying build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/cpp.cpython-38-x86_64-linux-gnu.so -> ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension 2023-03-31T07:14:12.8251177Z copying build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/ort.cpython-38-x86_64-linux-gnu.so -> ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension 2023-03-31T07:14:12.8333988Z copying build/lib.linux-x86_64-cpython-38/torch_test_cpp_extension/rng.cpython-38-x86_64-linux-gnu.so -> ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension 2023-03-31T07:14:12.8431976Z byte-compiling ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension/__init__.py to __init__.cpython-38.pyc 2023-03-31T07:14:12.8433518Z running install_egg_info 2023-03-31T07:14:12.8604049Z running egg_info 2023-03-31T07:14:12.8604524Z creating torch_test_cpp_extension.egg-info 2023-03-31T07:14:12.8640662Z writing torch_test_cpp_extension.egg-info/PKG-INFO 2023-03-31T07:14:12.8643858Z writing dependency_links to torch_test_cpp_extension.egg-info/dependency_links.txt 2023-03-31T07:14:12.8646562Z writing top-level names to torch_test_cpp_extension.egg-info/top_level.txt 2023-03-31T07:14:12.8648169Z writing manifest file 'torch_test_cpp_extension.egg-info/SOURCES.txt' 2023-03-31T07:14:12.8690119Z reading manifest file 'torch_test_cpp_extension.egg-info/SOURCES.txt' 2023-03-31T07:14:12.8696465Z writing manifest file 'torch_test_cpp_extension.egg-info/SOURCES.txt' 2023-03-31T07:14:12.8697941Z Copying torch_test_cpp_extension.egg-info to ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch_test_cpp_extension-0.0.0-py3.8.egg-info 2023-03-31T07:14:12.8702767Z running install_scripts 2023-03-31T07:14:14.8398332Z No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:14:14.8630606Z running install 2023-03-31T07:14:14.8632419Z /opt/conda/envs/py_3.8/lib/python3.8/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. 2023-03-31T07:14:14.8632859Z warnings.warn( 2023-03-31T07:14:14.8734742Z running build 2023-03-31T07:14:14.8735099Z running build_ext 2023-03-31T07:14:14.9140305Z building 'no_python_abi_suffix_test' extension 2023-03-31T07:14:14.9140916Z creating /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build 2023-03-31T07:14:14.9141665Z creating /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build/temp.linux-x86_64-cpython-38 2023-03-31T07:14:14.9525466Z Emitting ninja build file /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build/temp.linux-x86_64-cpython-38/build.ninja... 2023-03-31T07:14:14.9526905Z Compiling objects... 2023-03-31T07:14:14.9527154Z Using envvar MAX_JOBS (6) as the number of workers... 2023-03-31T07:14:15.0220636Z [1/1] c++ -MMD -MF /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build/temp.linux-x86_64-cpython-38/no_python_abi_suffix_test.o.d -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/TH -I/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/include/THC -I/opt/conda/envs/py_3.8/include/python3.8 -c -c /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/no_python_abi_suffix_test.cpp -o /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build/temp.linux-x86_64-cpython-38/no_python_abi_suffix_test.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_clang"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1002"' -DTORCH_EXTENSION_NAME=no_python_abi_suffix_test -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++17 2023-03-31T07:14:15.0222358Z cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ 2023-03-31T07:14:15.0272947Z creating build/lib.linux-x86_64-cpython-38 2023-03-31T07:14:15.0275392Z g++ -pthread -B /opt/conda/envs/py_3.8/compiler_compat -Wl,--sysroot=/ -pthread -shared -B /opt/conda/envs/py_3.8/compiler_compat -L/opt/conda/envs/py_3.8/lib -Wl,-rpath=/opt/conda/envs/py_3.8/lib -Wl,--no-as-needed -Wl,--sysroot=/ /var/lib/jenkins/workspace/test/cpp_extensions/no_python_abi_suffix_test/build/temp.linux-x86_64-cpython-38/no_python_abi_suffix_test.o -L/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib -lc10 -ltorch -ltorch_cpu -ltorch_python -o build/lib.linux-x86_64-cpython-38/no_python_abi_suffix_test.so 2023-03-31T07:14:15.2376718Z running install_lib 2023-03-31T07:14:15.2427525Z creating install 2023-03-31T07:14:15.2427807Z creating install/opt 2023-03-31T07:14:15.2428598Z creating install/opt/conda 2023-03-31T07:14:15.2428962Z creating install/opt/conda/envs 2023-03-31T07:14:15.2429505Z creating install/opt/conda/envs/py_3.8 2023-03-31T07:14:15.2430010Z creating install/opt/conda/envs/py_3.8/lib 2023-03-31T07:14:15.2430472Z creating install/opt/conda/envs/py_3.8/lib/python3.8 2023-03-31T07:14:15.2431263Z creating install/opt/conda/envs/py_3.8/lib/python3.8/site-packages 2023-03-31T07:14:15.2432598Z copying build/lib.linux-x86_64-cpython-38/no_python_abi_suffix_test.so -> ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages 2023-03-31T07:14:15.2438865Z running install_egg_info 2023-03-31T07:14:15.2550101Z running egg_info 2023-03-31T07:14:15.2550606Z creating no_python_abi_suffix_test.egg-info 2023-03-31T07:14:15.2585311Z writing no_python_abi_suffix_test.egg-info/PKG-INFO 2023-03-31T07:14:15.2588404Z writing dependency_links to no_python_abi_suffix_test.egg-info/dependency_links.txt 2023-03-31T07:14:15.2591111Z writing top-level names to no_python_abi_suffix_test.egg-info/top_level.txt 2023-03-31T07:14:15.2592678Z writing manifest file 'no_python_abi_suffix_test.egg-info/SOURCES.txt' 2023-03-31T07:14:15.2631777Z reading manifest file 'no_python_abi_suffix_test.egg-info/SOURCES.txt' 2023-03-31T07:14:15.2637508Z writing manifest file 'no_python_abi_suffix_test.egg-info/SOURCES.txt' 2023-03-31T07:14:15.2638371Z Copying no_python_abi_suffix_test.egg-info to ./install/opt/conda/envs/py_3.8/lib/python3.8/site-packages/no_python_abi_suffix_test-0.0.0-py3.8.egg-info 2023-03-31T07:14:15.2643275Z running install_scripts 2023-03-31T07:14:15.6436730Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cpp_extensions_aot_ninja.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:15.643252] 2023-03-31T07:14:18.2581712Z 2023-03-31T07:14:18.2582213Z Expand the folded group to see the log file of test_cpp_extensions_aot_ninja 2023-03-31T07:14:18.2583414Z ##[group]PRINTING LOG FILE of test_cpp_extensions_aot_ninja (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_aot_ninja_9p3er0ub.log) 2023-03-31T07:14:18.2584228Z Test results will be stored in test-reports/python-pytest/test_cpp_extensions_aot_ninja/test_cpp_extensions_aot_ninja-d75d40c2ceabefce.xml 2023-03-31T07:14:18.2584612Z ============================= test session starts ============================== 2023-03-31T07:14:18.2585083Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:18.2585364Z cachedir: .pytest_cache 2023-03-31T07:14:18.2585878Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:18.2586257Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:18.2586763Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:18.2587098Z collecting ... collected 16 items 2023-03-31T07:14:18.2589270Z Running 16 items in this shard: test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_backward, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cublas_extension, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cuda_dlink_libs, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cuda_extension, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cusolver_extension, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_extension_function, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_extension_module, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_no_python_abi_suffix_sets_the_correct_library_name, test/test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_optional, test/test_cpp_extensions_aot_ninja.py::TestPybindTypeCasters::test_pybind_return_types, test/test_cpp_extensions_aot_ninja.py::TestORTTensor::test_add, test/test_cpp_extensions_aot_ninja.py::TestORTTensor::test_conv_backend_override, test/test_cpp_extensions_aot_ninja.py::TestORTTensor::test_unregistered, test/test_cpp_extensions_aot_ninja.py::TestORTTensor::test_zeros, test/test_cpp_extensions_aot_ninja.py::TestRNGExtension::test_rng, test/test_cpp_extensions_aot_ninja.py::TestTorchLibrary::test_torch_library 2023-03-31T07:14:18.2591179Z 2023-03-31T07:14:18.2591366Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_backward PASSED [ 6%] 2023-03-31T07:14:18.2591835Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cublas_extension SKIPPED (CUDA not found) [ 12%] 2023-03-31T07:14:18.2592240Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cuda_dlink_libs SKIPPED (CUDA not found) [ 18%] 2023-03-31T07:14:18.2592711Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cuda_extension SKIPPED (CUDA not found) [ 25%] 2023-03-31T07:14:18.2593123Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_cusolver_extension SKIPPED (CUDA not found) [ 31%] 2023-03-31T07:14:18.2593579Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_extension_function PASSED [ 37%] 2023-03-31T07:14:18.2593958Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_extension_module PASSED [ 43%] 2023-03-31T07:14:18.2594440Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_no_python_abi_suffix_sets_the_correct_library_name PASSED [ 50%] 2023-03-31T07:14:18.2594906Z test_cpp_extensions_aot_ninja.py::TestCppExtensionAOT::test_optional PASSED [ 56%] 2023-03-31T07:14:18.2595342Z test_cpp_extensions_aot_ninja.py::TestPybindTypeCasters::test_pybind_return_types PASSED [ 62%] 2023-03-31T07:14:18.2595826Z test_cpp_extensions_aot_ninja.py::TestORTTensor::test_add PASSED [ 68%] 2023-03-31T07:14:18.2596170Z test_cpp_extensions_aot_ninja.py::TestORTTensor::test_conv_backend_override PASSED [ 75%] 2023-03-31T07:14:18.2596588Z test_cpp_extensions_aot_ninja.py::TestORTTensor::test_unregistered PASSED [ 81%] 2023-03-31T07:14:18.2596914Z test_cpp_extensions_aot_ninja.py::TestORTTensor::test_zeros PASSED [ 87%] 2023-03-31T07:14:18.2597308Z test_cpp_extensions_aot_ninja.py::TestRNGExtension::test_rng PASSED [ 93%] 2023-03-31T07:14:18.2597910Z test_cpp_extensions_aot_ninja.py::TestTorchLibrary::test_torch_library SKIPPED (CUDA not found) [100%]No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda' 2023-03-31T07:14:18.2598193Z 2023-03-31T07:14:18.2598198Z 2023-03-31T07:14:18.2598590Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cpp_extensions_aot_ninja/test_cpp_extensions_aot_ninja-d75d40c2ceabefce.xml - 2023-03-31T07:14:18.2599056Z ======================== 11 passed, 5 skipped in 0.08s ========================= 2023-03-31T07:14:18.2599382Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:18.2599648Z 2023-03-31T07:14:18.2599922Z ##[endgroup] 2023-03-31T07:14:18.2600434Z FINISHED PRINTING LOG FILE of test_cpp_extensions_aot_ninja (/var/lib/jenkins/workspace/test/test-reports/test_cpp_extensions_aot_ninja_9p3er0ub.log) 2023-03-31T07:14:18.2600697Z 2023-03-31T07:14:18.2601278Z Running test_jit_cuda_fuser ... [2023-03-31 07:14:18.258671] 2023-03-31T07:14:18.2601970Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_jit_cuda_fuser.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:18.258973] 2023-03-31T07:14:21.5535141Z 2023-03-31T07:14:21.5535707Z Expand the folded group to see the log file of test_jit_cuda_fuser 2023-03-31T07:14:21.5536850Z ##[group]PRINTING LOG FILE of test_jit_cuda_fuser (/var/lib/jenkins/workspace/test/test-reports/test_jit_cuda_fuser_37coh5vu.log) 2023-03-31T07:14:21.5537875Z Test results will be stored in test-reports/python-pytest/test_jit_cuda_fuser/test_jit_cuda_fuser-d23d3053563b1b2f.xml 2023-03-31T07:14:21.5538495Z ============================= test session starts ============================== 2023-03-31T07:14:21.5538891Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:21.5539181Z cachedir: .pytest_cache 2023-03-31T07:14:21.5539618Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:21.5539993Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:21.5540435Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:21.5540748Z collecting ... collected 158 items 2023-03-31T07:14:21.5555021Z Running 158 items in this shard: test/test_jit_cuda_fuser.py::TestCudaFuser::test__softmax_function, test/test_jit_cuda_fuser.py::TestCudaFuser::test__softmax_function_half_to_float, test/test_jit_cuda_fuser.py::TestCudaFuser::test_addcmul_ops, test/test_jit_cuda_fuser.py::TestCudaFuser::test_alias_pass_fix, test/test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_1, test/test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_1_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_2, test/test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_2_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_backward_type, test/test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_half, test/test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_impl_index_correctness, test/test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_impl_index_inner_bcast, test/test_jit_cuda_fuser.py::TestCudaFuser::test_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_binary_bitwise, test/test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops, test/test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_channels_last_with_bcast, test/test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_complex, test/test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_permutation, test/test_jit_cuda_fuser.py::TestCudaFuser::test_branches, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_0, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_1, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_2, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_3, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_multiple_output, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_multiple_output_shape, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_partition_logic_0, test/test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_partition_logic_1, test/test_jit_cuda_fuser.py::TestCudaFuser::test_build_shape_expression_native_dropout, test/test_jit_cuda_fuser.py::TestCudaFuser::test_category_rule, test/test_jit_cuda_fuser.py::TestCudaFuser::test_channels_last_with_broadcast, test/test_jit_cuda_fuser.py::TestCudaFuser::test_chunk, test/test_jit_cuda_fuser.py::TestCudaFuser::test_clamp, test/test_jit_cuda_fuser.py::TestCudaFuser::test_clamp_reversed_bound, test/test_jit_cuda_fuser.py::TestCudaFuser::test_clean_profile_ivalue, test/test_jit_cuda_fuser.py::TestCudaFuser::test_const, test/test_jit_cuda_fuser.py::TestCudaFuser::test_contiguous_on_broadcasted, test/test_jit_cuda_fuser.py::TestCudaFuser::test_conv2d_bias, test/test_jit_cuda_fuser.py::TestCudaFuser::test_conv2d_symbolic_shapes, test/test_jit_cuda_fuser.py::TestCudaFuser::test_cpu_scalar, test/test_jit_cuda_fuser.py::TestCudaFuser::test_cuda_fusion_guard, test/test_jit_cuda_fuser.py::TestCudaFuser::test_cuda_fusion_guard_backward, test/test_jit_cuda_fuser.py::TestCudaFuser::test_device_constant, test/test_jit_cuda_fuser.py::TestCudaFuser::test_disable_const_chunk_propagation_for_normalization, test/test_jit_cuda_fuser.py::TestCudaFuser::test_disable_sibling_fuse, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_inference_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_train_nograd_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_train_nograd_prob_check, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_training_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_training_prob_check, test/test_jit_cuda_fuser.py::TestCudaFuser::test_dynamic_size, test/test_jit_cuda_fuser.py::TestCudaFuser::test_expand, test/test_jit_cuda_fuser.py::TestCudaFuser::test_fix_shape_expression_bn, test/test_jit_cuda_fuser.py::TestCudaFuser::test_flatten, test/test_jit_cuda_fuser.py::TestCudaFuser::test_gelu, test/test_jit_cuda_fuser.py::TestCudaFuser::test_grad_sum_to_size, test/test_jit_cuda_fuser.py::TestCudaFuser::test_graph_for_with_missing_optimized_engine, test/test_jit_cuda_fuser.py::TestCudaFuser::test_graph_rng, test/test_jit_cuda_fuser.py::TestCudaFuser::test_half, test/test_jit_cuda_fuser.py::TestCudaFuser::test_high_rank_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_inf_quick_patch, test/test_jit_cuda_fuser.py::TestCudaFuser::test_inplace_removal, test/test_jit_cuda_fuser.py::TestCudaFuser::test_input_output_passthrough, test/test_jit_cuda_fuser.py::TestCudaFuser::test_int_tensor_input, test/test_jit_cuda_fuser.py::TestCudaFuser::test_issue1445_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_issue_1785, test/test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_autodiff, test/test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_parser, test/test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_trivial_reduce_dim, test/test_jit_cuda_fuser.py::TestCudaFuser::test_linear, test/test_jit_cuda_fuser.py::TestCudaFuser::test_linear_symbolic_shapes, test/test_jit_cuda_fuser.py::TestCudaFuser::test_multiple_device_pw, test/test_jit_cuda_fuser.py::TestCudaFuser::test_native_batch_norm_backward, test/test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm, test/test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm_half, test/test_jit_cuda_fuser.py::TestCudaFuser::test_nested_view, test/test_jit_cuda_fuser.py::TestCudaFuser::test_no_tensor_input, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm_channels_last, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm_half, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm_half_layer, test/test_jit_cuda_fuser.py::TestCudaFuser::test_norm_large, test/test_jit_cuda_fuser.py::TestCudaFuser::test_normalization_partition, test/test_jit_cuda_fuser.py::TestCudaFuser::test_nvfuser_comparison_callbacks_with_fallback, test/test_jit_cuda_fuser.py::TestCudaFuser::test_nvfuser_comparison_callbacks_without_fallback, test/test_jit_cuda_fuser.py::TestCudaFuser::test_overlapped_input, test/test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation, test/test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_0, test/test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_1_broken, test/test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_2, test/test_jit_cuda_fuser.py::TestCudaFuser::test_permute, test/test_jit_cuda_fuser.py::TestCudaFuser::test_pointwise_reference_tensor, test/test_jit_cuda_fuser.py::TestCudaFuser::test_profile_ivalue, test/test_jit_cuda_fuser.py::TestCudaFuser::test_profile_ivalue_multiple_profiles, test/test_jit_cuda_fuser.py::TestCudaFuser::test_profiling_node, test/test_jit_cuda_fuser.py::TestCudaFuser::test_pw_single_reduction_partition, test/test_jit_cuda_fuser.py::TestCudaFuser::test_random_topo, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_dtypes_axis, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_empty_axes, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_multiple_output, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_permutation, test/test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_sizes_op, test/test_jit_cuda_fuser.py::TestCudaFuser::test_remove_output_used_only_in_dtype, test/test_jit_cuda_fuser.py::TestCudaFuser::test_rsub, test/test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_cuda_tensor, test/test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_input, test/test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_tensor, test/test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_tensor_permuted, test/test_jit_cuda_fuser.py::TestCudaFuser::test_scheduler_with_polymorphic_broadcast, test/test_jit_cuda_fuser.py::TestCudaFuser::test_shape_expression, test/test_jit_cuda_fuser.py::TestCudaFuser::test_sibling_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_sibling_fusion_no_scalar_inputs, test/test_jit_cuda_fuser.py::TestCudaFuser::test_single_reduction_broadcast, test/test_jit_cuda_fuser.py::TestCudaFuser::test_singleton_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_skip_parser, test/test_jit_cuda_fuser.py::TestCudaFuser::test_softmax, test/test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_bfloat, test/test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_dtype, test/test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_half, test/test_jit_cuda_fuser.py::TestCudaFuser::test_softplus_fuser, test/test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze, test/test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze_negative_dim, test/test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze_zero, test/test_jit_cuda_fuser.py::TestCudaFuser::test_strict_fusion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_sum_to_one, test/test_jit_cuda_fuser.py::TestCudaFuser::test_sum_to_size, test/test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops, test/test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops_integer_compatibility, test/test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops_type_promotion, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_boolean, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_copy, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_bf16_to_bf16, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_bf16_to_fp32, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp16_to_fp16, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp16_to_fp32, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp32_to_bf16, test/test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp32_to_fp16, test/test_jit_cuda_fuser.py::TestCudaFuser::test_transpose, test/test_jit_cuda_fuser.py::TestCudaFuser::test_transpose_default, test/test_jit_cuda_fuser.py::TestCudaFuser::test_trivial_reduction, test/test_jit_cuda_fuser.py::TestCudaFuser::test_type_as_op, test/test_jit_cuda_fuser.py::TestCudaFuser::test_type_inference, test/test_jit_cuda_fuser.py::TestCudaFuser::test_unary_bitwise, test/test_jit_cuda_fuser.py::TestCudaFuser::test_unary_ops, test/test_jit_cuda_fuser.py::TestCudaFuser::test_unsqueeze, test/test_jit_cuda_fuser.py::TestCudaFuser::test_variance, test/test_jit_cuda_fuser.py::TestCudaFuser::test_variance_profiling, test/test_jit_cuda_fuser.py::TestCudaFuser::test_view, test/test_jit_cuda_fuser.py::TestCudaFuser::test_view_before_permute, test/test_jit_cuda_fuser.py::TestCudaFuser::test_view_copy_graph_guard, test/test_jit_cuda_fuser.py::TestCudaFuser::test_view_copy_graph_guard_double_fusion, test/test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_can_be_enabled_nvfuser, test/test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_context_manager_test, test/test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser, test/test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser_cpu, test/test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser_rocm 2023-03-31T07:14:21.5568402Z 2023-03-31T07:14:21.5568820Z test_jit_cuda_fuser.py::TestCudaFuser::test__softmax_function <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 0%] 2023-03-31T07:14:21.5569463Z test_jit_cuda_fuser.py::TestCudaFuser::test__softmax_function_half_to_float <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 1%] 2023-03-31T07:14:21.5570059Z test_jit_cuda_fuser.py::TestCudaFuser::test_addcmul_ops <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 1%] 2023-03-31T07:14:21.5570668Z test_jit_cuda_fuser.py::TestCudaFuser::test_alias_pass_fix <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since unsqueeze is disabled now) [ 2%] 2023-03-31T07:14:21.5571301Z test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_1 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 3%] 2023-03-31T07:14:21.5571831Z test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_1_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 3%] 2023-03-31T07:14:21.5572414Z test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_2 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 4%] 2023-03-31T07:14:21.5572953Z test_jit_cuda_fuser.py::TestCudaFuser::test_autocast_2_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 5%] 2023-03-31T07:14:21.5573489Z test_jit_cuda_fuser.py::TestCudaFuser::test_backward_type <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 5%] 2023-03-31T07:14:21.5574008Z test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_half <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 6%] 2023-03-31T07:14:21.5574632Z test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_impl_index_correctness <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 6%] 2023-03-31T07:14:21.5575221Z test_jit_cuda_fuser.py::TestCudaFuser::test_batch_norm_impl_index_inner_bcast <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 7%] 2023-03-31T07:14:21.5575795Z test_jit_cuda_fuser.py::TestCudaFuser::test_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (device does not support BFloat16) [ 8%] 2023-03-31T07:14:21.5576399Z test_jit_cuda_fuser.py::TestCudaFuser::test_binary_bitwise <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 8%] 2023-03-31T07:14:21.5576977Z test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 9%] 2023-03-31T07:14:21.5577541Z test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_channels_last_with_bcast <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 10%] 2023-03-31T07:14:21.5578204Z test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_complex <- test/_nvfuser/test_torchscript.py SKIPPED (see issue https://github.com/csarofeen/pytorch/issues/1730) [ 10%] 2023-03-31T07:14:21.5578871Z test_jit_cuda_fuser.py::TestCudaFuser::test_binary_ops_permutation <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 11%] 2023-03-31T07:14:21.5579408Z test_jit_cuda_fuser.py::TestCudaFuser::test_branches <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 12%] 2023-03-31T07:14:21.5579945Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_0 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 12%] 2023-03-31T07:14:21.5580484Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_1 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 13%] 2023-03-31T07:14:21.5580999Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_2 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 13%] 2023-03-31T07:14:21.5581516Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_3 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 14%] 2023-03-31T07:14:21.5582187Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_multiple_output <- test/_nvfuser/test_torchscript.py SKIPPED (broadcast on branches can't be resolved yet) [ 15%] 2023-03-31T07:14:21.5582990Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_multiple_output_shape <- test/_nvfuser/test_torchscript.py SKIPPED (Broadcast with different output not supported yet) [ 15%] 2023-03-31T07:14:21.5583613Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_partition_logic_0 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 16%] 2023-03-31T07:14:21.5584198Z test_jit_cuda_fuser.py::TestCudaFuser::test_broadcasting_partition_logic_1 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 17%] 2023-03-31T07:14:21.5584791Z test_jit_cuda_fuser.py::TestCudaFuser::test_build_shape_expression_native_dropout <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 17%] 2023-03-31T07:14:21.5585426Z test_jit_cuda_fuser.py::TestCudaFuser::test_category_rule <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 18%] 2023-03-31T07:14:21.5585977Z test_jit_cuda_fuser.py::TestCudaFuser::test_channels_last_with_broadcast <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 18%] 2023-03-31T07:14:21.5586510Z test_jit_cuda_fuser.py::TestCudaFuser::test_chunk <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 19%] 2023-03-31T07:14:21.5587025Z test_jit_cuda_fuser.py::TestCudaFuser::test_clamp <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 20%] 2023-03-31T07:14:21.5587566Z test_jit_cuda_fuser.py::TestCudaFuser::test_clamp_reversed_bound <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 20%] 2023-03-31T07:14:21.5588103Z test_jit_cuda_fuser.py::TestCudaFuser::test_clean_profile_ivalue <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 21%] 2023-03-31T07:14:21.5588628Z test_jit_cuda_fuser.py::TestCudaFuser::test_const <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 22%] 2023-03-31T07:14:21.5589172Z test_jit_cuda_fuser.py::TestCudaFuser::test_contiguous_on_broadcasted <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 22%] 2023-03-31T07:14:21.5589713Z test_jit_cuda_fuser.py::TestCudaFuser::test_conv2d_bias <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 23%] 2023-03-31T07:14:21.5590255Z test_jit_cuda_fuser.py::TestCudaFuser::test_conv2d_symbolic_shapes <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 24%] 2023-03-31T07:14:21.5590832Z test_jit_cuda_fuser.py::TestCudaFuser::test_cpu_scalar <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 24%] 2023-03-31T07:14:21.5591400Z test_jit_cuda_fuser.py::TestCudaFuser::test_cuda_fusion_guard <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 25%] 2023-03-31T07:14:21.5591964Z test_jit_cuda_fuser.py::TestCudaFuser::test_cuda_fusion_guard_backward <- test/_nvfuser/test_torchscript.py SKIPPED (requires NVFuser) [ 25%] 2023-03-31T07:14:21.5592501Z test_jit_cuda_fuser.py::TestCudaFuser::test_device_constant <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 26%] 2023-03-31T07:14:21.5593095Z test_jit_cuda_fuser.py::TestCudaFuser::test_disable_const_chunk_propagation_for_normalization <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 27%] 2023-03-31T07:14:21.5593672Z test_jit_cuda_fuser.py::TestCudaFuser::test_disable_sibling_fuse <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 27%] 2023-03-31T07:14:21.5594227Z test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_inference_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 28%] 2023-03-31T07:14:21.5608165Z test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_train_nograd_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (not enough memory) [ 29%] 2023-03-31T07:14:21.5608780Z test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_train_nograd_prob_check <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 29%] 2023-03-31T07:14:21.5609345Z test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_training_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 30%] 2023-03-31T07:14:21.5610078Z test_jit_cuda_fuser.py::TestCudaFuser::test_dropout_training_prob_check <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 31%] 2023-03-31T07:14:21.5610681Z test_jit_cuda_fuser.py::TestCudaFuser::test_dynamic_size <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 31%] 2023-03-31T07:14:21.5611209Z test_jit_cuda_fuser.py::TestCudaFuser::test_expand <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 32%] 2023-03-31T07:14:21.5611748Z test_jit_cuda_fuser.py::TestCudaFuser::test_fix_shape_expression_bn <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 32%] 2023-03-31T07:14:21.5612345Z test_jit_cuda_fuser.py::TestCudaFuser::test_flatten <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since flatten is disabled now) [ 33%] 2023-03-31T07:14:21.5612897Z test_jit_cuda_fuser.py::TestCudaFuser::test_gelu <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 34%] 2023-03-31T07:14:21.5613432Z test_jit_cuda_fuser.py::TestCudaFuser::test_grad_sum_to_size <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 34%] 2023-03-31T07:14:21.5613997Z test_jit_cuda_fuser.py::TestCudaFuser::test_graph_for_with_missing_optimized_engine <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 35%] 2023-03-31T07:14:21.5614554Z test_jit_cuda_fuser.py::TestCudaFuser::test_graph_rng <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 36%] 2023-03-31T07:14:21.5615063Z test_jit_cuda_fuser.py::TestCudaFuser::test_half <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 36%] 2023-03-31T07:14:21.5615601Z test_jit_cuda_fuser.py::TestCudaFuser::test_high_rank_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 37%] 2023-03-31T07:14:21.5616119Z test_jit_cuda_fuser.py::TestCudaFuser::test_inf_quick_patch <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 37%] 2023-03-31T07:14:21.5616653Z test_jit_cuda_fuser.py::TestCudaFuser::test_inplace_removal <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 38%] 2023-03-31T07:14:21.5617212Z test_jit_cuda_fuser.py::TestCudaFuser::test_input_output_passthrough <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 39%] 2023-03-31T07:14:21.5617767Z test_jit_cuda_fuser.py::TestCudaFuser::test_int_tensor_input <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 39%] 2023-03-31T07:14:21.5618340Z test_jit_cuda_fuser.py::TestCudaFuser::test_issue1445_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 40%] 2023-03-31T07:14:21.5618906Z test_jit_cuda_fuser.py::TestCudaFuser::test_issue_1785 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 41%] 2023-03-31T07:14:21.5619437Z test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_autodiff <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 41%] 2023-03-31T07:14:21.5619979Z test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_parser <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 42%] 2023-03-31T07:14:21.5620531Z test_jit_cuda_fuser.py::TestCudaFuser::test_layer_norm_trivial_reduce_dim <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 43%] 2023-03-31T07:14:21.5621072Z test_jit_cuda_fuser.py::TestCudaFuser::test_linear <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 43%] 2023-03-31T07:14:21.5621616Z test_jit_cuda_fuser.py::TestCudaFuser::test_linear_symbolic_shapes <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 44%] 2023-03-31T07:14:21.5622166Z test_jit_cuda_fuser.py::TestCudaFuser::test_multiple_device_pw <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 44%] 2023-03-31T07:14:21.5622707Z test_jit_cuda_fuser.py::TestCudaFuser::test_native_batch_norm_backward <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 45%] 2023-03-31T07:14:21.5623353Z test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 46%] 2023-03-31T07:14:21.5623943Z test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 46%] 2023-03-31T07:14:21.5624525Z test_jit_cuda_fuser.py::TestCudaFuser::test_native_layer_norm_half <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 47%] 2023-03-31T07:14:21.5625189Z test_jit_cuda_fuser.py::TestCudaFuser::test_nested_view <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since view is disabled now) [ 48%] 2023-03-31T07:14:21.5625758Z test_jit_cuda_fuser.py::TestCudaFuser::test_no_tensor_input <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 48%] 2023-03-31T07:14:21.5626273Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 49%] 2023-03-31T07:14:21.5626791Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 50%] 2023-03-31T07:14:21.5627312Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm_channels_last <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 50%] 2023-03-31T07:14:21.5627838Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm_half <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 51%] 2023-03-31T07:14:21.5628361Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm_half_layer <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 51%] 2023-03-31T07:14:21.5628884Z test_jit_cuda_fuser.py::TestCudaFuser::test_norm_large <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 52%] 2023-03-31T07:14:21.5629423Z test_jit_cuda_fuser.py::TestCudaFuser::test_normalization_partition <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 53%] 2023-03-31T07:14:21.5630015Z test_jit_cuda_fuser.py::TestCudaFuser::test_nvfuser_comparison_callbacks_with_fallback <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 53%] 2023-03-31T07:14:21.5630632Z test_jit_cuda_fuser.py::TestCudaFuser::test_nvfuser_comparison_callbacks_without_fallback <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 54%] 2023-03-31T07:14:21.5631203Z test_jit_cuda_fuser.py::TestCudaFuser::test_overlapped_input <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 55%] 2023-03-31T07:14:21.5631750Z test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 55%] 2023-03-31T07:14:21.5632371Z test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_0 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 56%] 2023-03-31T07:14:21.5633010Z test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_1_broken <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 56%] 2023-03-31T07:14:21.5633614Z test_jit_cuda_fuser.py::TestCudaFuser::test_permutation_preservation_edge_case_2 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 57%] 2023-03-31T07:14:21.5634144Z test_jit_cuda_fuser.py::TestCudaFuser::test_permute <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 58%] 2023-03-31T07:14:21.5634692Z test_jit_cuda_fuser.py::TestCudaFuser::test_pointwise_reference_tensor <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 58%] 2023-03-31T07:14:21.5635239Z test_jit_cuda_fuser.py::TestCudaFuser::test_profile_ivalue <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 59%] 2023-03-31T07:14:21.5635808Z test_jit_cuda_fuser.py::TestCudaFuser::test_profile_ivalue_multiple_profiles <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 60%] 2023-03-31T07:14:21.5636411Z test_jit_cuda_fuser.py::TestCudaFuser::test_profiling_node <- test/_nvfuser/test_torchscript.py SKIPPED (Skipped due to rand_like behavior change) [ 60%] 2023-03-31T07:14:21.5636987Z test_jit_cuda_fuser.py::TestCudaFuser::test_pw_single_reduction_partition <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 61%] 2023-03-31T07:14:21.5637527Z test_jit_cuda_fuser.py::TestCudaFuser::test_random_topo <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 62%] 2023-03-31T07:14:21.5638081Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 62%] 2023-03-31T07:14:21.5638645Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_dtypes_axis <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 63%] 2023-03-31T07:14:21.5639181Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_empty_axes <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 63%] 2023-03-31T07:14:21.5639740Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_multiple_output <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 64%] 2023-03-31T07:14:21.5640294Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_permutation <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 65%] 2023-03-31T07:14:21.5640840Z test_jit_cuda_fuser.py::TestCudaFuser::test_reduction_sizes_op <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 65%] 2023-03-31T07:14:21.5641398Z test_jit_cuda_fuser.py::TestCudaFuser::test_remove_output_used_only_in_dtype <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 66%] 2023-03-31T07:14:21.5641932Z test_jit_cuda_fuser.py::TestCudaFuser::test_rsub <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 67%] 2023-03-31T07:14:21.5642461Z test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_cuda_tensor <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 67%] 2023-03-31T07:14:21.5642992Z test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_input <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 68%] 2023-03-31T07:14:21.5643699Z test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_tensor <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 68%] 2023-03-31T07:14:21.5644245Z test_jit_cuda_fuser.py::TestCudaFuser::test_scalar_tensor_permuted <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 69%] 2023-03-31T07:14:21.5644829Z test_jit_cuda_fuser.py::TestCudaFuser::test_scheduler_with_polymorphic_broadcast <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 70%] 2023-03-31T07:14:21.5645465Z test_jit_cuda_fuser.py::TestCudaFuser::test_shape_expression <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 70%] 2023-03-31T07:14:21.5646033Z test_jit_cuda_fuser.py::TestCudaFuser::test_sibling_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 71%] 2023-03-31T07:14:21.5646662Z test_jit_cuda_fuser.py::TestCudaFuser::test_sibling_fusion_no_scalar_inputs <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 72%] 2023-03-31T07:14:21.5647276Z test_jit_cuda_fuser.py::TestCudaFuser::test_single_reduction_broadcast <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 72%] 2023-03-31T07:14:21.5647826Z test_jit_cuda_fuser.py::TestCudaFuser::test_singleton_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 73%] 2023-03-31T07:14:21.5648339Z test_jit_cuda_fuser.py::TestCudaFuser::test_skip_parser <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 74%] 2023-03-31T07:14:21.5648864Z test_jit_cuda_fuser.py::TestCudaFuser::test_softmax <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 74%] 2023-03-31T07:14:21.5649393Z test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_bfloat <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 75%] 2023-03-31T07:14:21.5649927Z test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_dtype <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 75%] 2023-03-31T07:14:21.5650445Z test_jit_cuda_fuser.py::TestCudaFuser::test_softmax_half <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 76%] 2023-03-31T07:14:21.5650976Z test_jit_cuda_fuser.py::TestCudaFuser::test_softplus_fuser <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 77%] 2023-03-31T07:14:21.5651577Z test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 77%] 2023-03-31T07:14:21.5652299Z test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze_negative_dim <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 78%] 2023-03-31T07:14:21.5652991Z test_jit_cuda_fuser.py::TestCudaFuser::test_squeeze_zero <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 79%] 2023-03-31T07:14:21.5653569Z test_jit_cuda_fuser.py::TestCudaFuser::test_strict_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 79%] 2023-03-31T07:14:21.5654091Z test_jit_cuda_fuser.py::TestCudaFuser::test_sum_to_one <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 80%] 2023-03-31T07:14:21.5654606Z test_jit_cuda_fuser.py::TestCudaFuser::test_sum_to_size <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 81%] 2023-03-31T07:14:21.5655108Z test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 81%] 2023-03-31T07:14:21.5655674Z test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops_integer_compatibility <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 82%] 2023-03-31T07:14:21.5656256Z test_jit_cuda_fuser.py::TestCudaFuser::test_ternary_ops_type_promotion <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 82%] 2023-03-31T07:14:21.5656798Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_boolean <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 83%] 2023-03-31T07:14:21.5657303Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_copy <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 84%] 2023-03-31T07:14:21.5657837Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_bf16_to_bf16 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 84%] 2023-03-31T07:14:21.5658373Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_bf16_to_fp32 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 85%] 2023-03-31T07:14:21.5658904Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp16_to_fp16 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 86%] 2023-03-31T07:14:21.5659418Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp16_to_fp32 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 86%] 2023-03-31T07:14:21.5659952Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp32_to_bf16 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 87%] 2023-03-31T07:14:21.5660526Z test_jit_cuda_fuser.py::TestCudaFuser::test_to_dtype_fp32_to_fp16 <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 87%] 2023-03-31T07:14:21.5661079Z test_jit_cuda_fuser.py::TestCudaFuser::test_transpose <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 88%] 2023-03-31T07:14:21.5661606Z test_jit_cuda_fuser.py::TestCudaFuser::test_transpose_default <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 89%] 2023-03-31T07:14:21.5662146Z test_jit_cuda_fuser.py::TestCudaFuser::test_trivial_reduction <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 89%] 2023-03-31T07:14:21.5662672Z test_jit_cuda_fuser.py::TestCudaFuser::test_type_as_op <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 90%] 2023-03-31T07:14:21.5663281Z test_jit_cuda_fuser.py::TestCudaFuser::test_type_inference <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 91%] 2023-03-31T07:14:21.5663800Z test_jit_cuda_fuser.py::TestCudaFuser::test_unary_bitwise <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 91%] 2023-03-31T07:14:21.5664318Z test_jit_cuda_fuser.py::TestCudaFuser::test_unary_ops <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 92%] 2023-03-31T07:14:21.5664915Z test_jit_cuda_fuser.py::TestCudaFuser::test_unsqueeze <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 93%] 2023-03-31T07:14:21.5665477Z test_jit_cuda_fuser.py::TestCudaFuser::test_variance <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 93%] 2023-03-31T07:14:21.5666039Z test_jit_cuda_fuser.py::TestCudaFuser::test_variance_profiling <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 94%] 2023-03-31T07:14:21.5666645Z test_jit_cuda_fuser.py::TestCudaFuser::test_view <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since view is disabled now) [ 94%] 2023-03-31T07:14:21.5667217Z test_jit_cuda_fuser.py::TestCudaFuser::test_view_before_permute <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 95%] 2023-03-31T07:14:21.5667823Z test_jit_cuda_fuser.py::TestCudaFuser::test_view_copy_graph_guard <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since reshape is disabled now) [ 96%] 2023-03-31T07:14:21.5668470Z test_jit_cuda_fuser.py::TestCudaFuser::test_view_copy_graph_guard_double_fusion <- test/_nvfuser/test_torchscript.py SKIPPED (skipping this test since view is disabled now) [ 96%] 2023-03-31T07:14:21.5669078Z test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_can_be_enabled_nvfuser <- test/_nvfuser/test_torchscript.py PASSED [ 97%] 2023-03-31T07:14:21.5669679Z test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_context_manager_test <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 98%] 2023-03-31T07:14:21.5670277Z test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [ 98%] 2023-03-31T07:14:21.5670850Z test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser_cpu <- test/_nvfuser/test_torchscript.py PASSED [ 99%] 2023-03-31T07:14:21.5671428Z test_jit_cuda_fuser.py::TestEnableDisableCudaFuser::test_register_fuser_rocm <- test/_nvfuser/test_torchscript.py SKIPPED (requires CUDA) [100%] 2023-03-31T07:14:21.5671693Z 2023-03-31T07:14:21.5672039Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit_cuda_fuser/test_jit_cuda_fuser-d23d3053563b1b2f.xml - 2023-03-31T07:14:21.5672414Z ======================== 2 passed, 156 skipped in 0.22s ======================== 2023-03-31T07:14:21.5672744Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:21.5672935Z 2023-03-31T07:14:21.5673268Z ##[endgroup] 2023-03-31T07:14:21.5673694Z FINISHED PRINTING LOG FILE of test_jit_cuda_fuser (/var/lib/jenkins/workspace/test/test-reports/test_jit_cuda_fuser_37coh5vu.log) 2023-03-31T07:14:21.5673927Z 2023-03-31T07:14:21.5674181Z Running _nvfuser/test_torchscript ... [2023-03-31 07:14:21.554053] 2023-03-31T07:14:21.5674851Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', '_nvfuser/test_torchscript.py', '-v', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:21.554382] 2023-03-31T07:14:24.8090795Z 2023-03-31T07:14:24.8091665Z Expand the folded group to see the log file of _nvfuser/test_torchscript 2023-03-31T07:14:24.8092515Z ##[group]PRINTING LOG FILE of _nvfuser/test_torchscript (/var/lib/jenkins/workspace/test/test-reports/_nvfuser-test_torchscript_nyfqim5r.log) 2023-03-31T07:14:24.8093285Z Test results will be stored in test-reports/python-pytest/_nvfuser.test_torchscript/_nvfuser.test_torchscript-5ab7f0515610d7c1.xml 2023-03-31T07:14:24.8093763Z ============================= test session starts ============================== 2023-03-31T07:14:24.8094160Z platform linux -- Python 3.8.16, pytest-7.2.2, pluggy-1.0.0 -- /opt/conda/envs/py_3.8/bin/python 2023-03-31T07:14:24.8094456Z cachedir: .pytest_cache 2023-03-31T07:14:24.8094897Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2023-03-31T07:14:24.8095259Z rootdir: /var/lib/jenkins/workspace, configfile: pytest.ini 2023-03-31T07:14:24.8095709Z plugins: hypothesis-5.35.1, flakefinder-1.1.0, rerunfailures-11.1.2, shard-0.1.2, xdist-3.2.1, xdoctest-1.1.0 2023-03-31T07:14:24.8096022Z collecting ... collected 158 items 2023-03-31T07:14:24.8110797Z Running 158 items in this shard: test/_nvfuser/test_torchscript.py::TestCudaFuser::test__softmax_function, test/_nvfuser/test_torchscript.py::TestCudaFuser::test__softmax_function_half_to_float, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_addcmul_ops, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_alias_pass_fix, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_1, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_1_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_2, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_2_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_backward_type, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_half, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_impl_index_correctness, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_impl_index_inner_bcast, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_binary_bitwise, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_channels_last_with_bcast, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_complex, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_permutation, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_branches, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_0, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_1, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_2, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_3, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_multiple_output, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_multiple_output_shape, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_partition_logic_0, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_partition_logic_1, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_build_shape_expression_native_dropout, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_category_rule, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_channels_last_with_broadcast, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_chunk, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_clamp, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_clamp_reversed_bound, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_clean_profile_ivalue, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_const, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_contiguous_on_broadcasted, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_conv2d_bias, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_conv2d_symbolic_shapes, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_cpu_scalar, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_cuda_fusion_guard, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_cuda_fusion_guard_backward, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_device_constant, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_disable_const_chunk_propagation_for_normalization, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_disable_sibling_fuse, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_inference_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_train_nograd_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_train_nograd_prob_check, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_training_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_training_prob_check, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_dynamic_size, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_expand, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_fix_shape_expression_bn, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_flatten, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_gelu, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_grad_sum_to_size, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_graph_for_with_missing_optimized_engine, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_graph_rng, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_half, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_high_rank_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_inf_quick_patch, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_inplace_removal, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_input_output_passthrough, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_int_tensor_input, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_issue1445_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_issue_1785, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_autodiff, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_parser, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_trivial_reduce_dim, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_linear, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_linear_symbolic_shapes, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_multiple_device_pw, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_native_batch_norm_backward, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm_half, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_nested_view, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_no_tensor_input, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm_channels_last, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm_half, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm_half_layer, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_norm_large, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_normalization_partition, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_nvfuser_comparison_callbacks_with_fallback, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_nvfuser_comparison_callbacks_without_fallback, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_overlapped_input, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_0, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_1_broken, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_2, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_permute, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_pointwise_reference_tensor, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_profile_ivalue, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_profile_ivalue_multiple_profiles, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_profiling_node, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_pw_single_reduction_partition, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_random_topo, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_dtypes_axis, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_empty_axes, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_multiple_output, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_permutation, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_sizes_op, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_remove_output_used_only_in_dtype, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_rsub, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_cuda_tensor, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_input, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_tensor, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_tensor_permuted, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_scheduler_with_polymorphic_broadcast, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_shape_expression, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_sibling_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_sibling_fusion_no_scalar_inputs, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_single_reduction_broadcast, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_singleton_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_skip_parser, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_softmax, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_bfloat, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_dtype, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_half, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_softplus_fuser, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze_negative_dim, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze_zero, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_strict_fusion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_sum_to_one, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_sum_to_size, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops_integer_compatibility, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops_type_promotion, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_boolean, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_copy, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_bf16_to_bf16, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_bf16_to_fp32, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp16_to_fp16, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp16_to_fp32, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp32_to_bf16, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp32_to_fp16, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_transpose, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_transpose_default, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_trivial_reduction, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_type_as_op, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_type_inference, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_unary_bitwise, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_unary_ops, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_unsqueeze, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_variance, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_variance_profiling, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_view, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_view_before_permute, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_view_copy_graph_guard, test/_nvfuser/test_torchscript.py::TestCudaFuser::test_view_copy_graph_guard_double_fusion, test/_nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_can_be_enabled_nvfuser, test/_nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_context_manager_test, test/_nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser, test/_nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser_cpu, test/_nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser_rocm 2023-03-31T07:14:24.8124668Z 2023-03-31T07:14:24.8124871Z _nvfuser/test_torchscript.py::TestCudaFuser::test__softmax_function SKIPPED (requires CUDA) [ 0%] 2023-03-31T07:14:24.8125275Z _nvfuser/test_torchscript.py::TestCudaFuser::test__softmax_function_half_to_float SKIPPED (requires CUDA) [ 1%] 2023-03-31T07:14:24.8125672Z _nvfuser/test_torchscript.py::TestCudaFuser::test_addcmul_ops SKIPPED (requires CUDA) [ 1%] 2023-03-31T07:14:24.8126162Z _nvfuser/test_torchscript.py::TestCudaFuser::test_alias_pass_fix SKIPPED (skipping this test since unsqueeze is disabled now) [ 2%] 2023-03-31T07:14:24.8126624Z _nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_1 SKIPPED (requires CUDA) [ 3%] 2023-03-31T07:14:24.8127006Z _nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_1_bfloat SKIPPED (requires CUDA) [ 3%] 2023-03-31T07:14:24.8127384Z _nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_2 SKIPPED (requires CUDA) [ 4%] 2023-03-31T07:14:24.8127759Z _nvfuser/test_torchscript.py::TestCudaFuser::test_autocast_2_bfloat SKIPPED (requires CUDA) [ 5%] 2023-03-31T07:14:24.8128121Z _nvfuser/test_torchscript.py::TestCudaFuser::test_backward_type SKIPPED (requires CUDA) [ 5%] 2023-03-31T07:14:24.8128497Z _nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_half SKIPPED (requires CUDA) [ 6%] 2023-03-31T07:14:24.8128898Z _nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_impl_index_correctness SKIPPED (requires CUDA) [ 6%] 2023-03-31T07:14:24.8129320Z _nvfuser/test_torchscript.py::TestCudaFuser::test_batch_norm_impl_index_inner_bcast SKIPPED (requires CUDA) [ 7%] 2023-03-31T07:14:24.8129712Z _nvfuser/test_torchscript.py::TestCudaFuser::test_bfloat SKIPPED (device does not support BFloat16) [ 8%] 2023-03-31T07:14:24.8130102Z _nvfuser/test_torchscript.py::TestCudaFuser::test_binary_bitwise SKIPPED (requires CUDA) [ 8%] 2023-03-31T07:14:24.8130476Z _nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops SKIPPED (requires CUDA) [ 9%] 2023-03-31T07:14:24.8130861Z _nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_channels_last_with_bcast SKIPPED (requires CUDA) [ 10%] 2023-03-31T07:14:24.8131339Z _nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_complex SKIPPED (see issue https://github.com/csarofeen/pytorch/issues/1730) [ 10%] 2023-03-31T07:14:24.8131779Z _nvfuser/test_torchscript.py::TestCudaFuser::test_binary_ops_permutation SKIPPED (requires CUDA) [ 11%] 2023-03-31T07:14:24.8132160Z _nvfuser/test_torchscript.py::TestCudaFuser::test_branches SKIPPED (requires CUDA) [ 12%] 2023-03-31T07:14:24.8132519Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_0 SKIPPED (requires CUDA) [ 12%] 2023-03-31T07:14:24.8132889Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_1 SKIPPED (requires CUDA) [ 13%] 2023-03-31T07:14:24.8133306Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_2 SKIPPED (requires CUDA) [ 13%] 2023-03-31T07:14:24.8133714Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_3 SKIPPED (requires CUDA) [ 14%] 2023-03-31T07:14:24.8134275Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_multiple_output SKIPPED (broadcast on branches can't be resolved yet) [ 15%] 2023-03-31T07:14:24.8134766Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_multiple_output_shape SKIPPED (Broadcast with different output not supported yet) [ 15%] 2023-03-31T07:14:24.8135239Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_partition_logic_0 SKIPPED (requires CUDA) [ 16%] 2023-03-31T07:14:24.8135651Z _nvfuser/test_torchscript.py::TestCudaFuser::test_broadcasting_partition_logic_1 SKIPPED (requires CUDA) [ 17%] 2023-03-31T07:14:24.8136063Z _nvfuser/test_torchscript.py::TestCudaFuser::test_build_shape_expression_native_dropout SKIPPED (requires CUDA) [ 17%] 2023-03-31T07:14:24.8136461Z _nvfuser/test_torchscript.py::TestCudaFuser::test_category_rule SKIPPED (requires CUDA) [ 18%] 2023-03-31T07:14:24.8136852Z _nvfuser/test_torchscript.py::TestCudaFuser::test_channels_last_with_broadcast SKIPPED (requires CUDA) [ 18%] 2023-03-31T07:14:24.8137232Z _nvfuser/test_torchscript.py::TestCudaFuser::test_chunk SKIPPED (requires CUDA) [ 19%] 2023-03-31T07:14:24.8137579Z _nvfuser/test_torchscript.py::TestCudaFuser::test_clamp SKIPPED (requires CUDA) [ 20%] 2023-03-31T07:14:24.8137953Z _nvfuser/test_torchscript.py::TestCudaFuser::test_clamp_reversed_bound SKIPPED (requires CUDA) [ 20%] 2023-03-31T07:14:24.8138379Z _nvfuser/test_torchscript.py::TestCudaFuser::test_clean_profile_ivalue SKIPPED (requires CUDA) [ 21%] 2023-03-31T07:14:24.8138729Z _nvfuser/test_torchscript.py::TestCudaFuser::test_const SKIPPED (requires CUDA) [ 22%] 2023-03-31T07:14:24.8139167Z _nvfuser/test_torchscript.py::TestCudaFuser::test_contiguous_on_broadcasted SKIPPED (requires CUDA) [ 22%] 2023-03-31T07:14:24.8139558Z _nvfuser/test_torchscript.py::TestCudaFuser::test_conv2d_bias SKIPPED (requires CUDA) [ 23%] 2023-03-31T07:14:24.8139945Z _nvfuser/test_torchscript.py::TestCudaFuser::test_conv2d_symbolic_shapes SKIPPED (requires CUDA) [ 24%] 2023-03-31T07:14:24.8140312Z _nvfuser/test_torchscript.py::TestCudaFuser::test_cpu_scalar SKIPPED (requires CUDA) [ 24%] 2023-03-31T07:14:24.8140687Z _nvfuser/test_torchscript.py::TestCudaFuser::test_cuda_fusion_guard SKIPPED (requires CUDA) [ 25%] 2023-03-31T07:14:24.8141091Z _nvfuser/test_torchscript.py::TestCudaFuser::test_cuda_fusion_guard_backward SKIPPED (requires NVFuser) [ 25%] 2023-03-31T07:14:24.8141487Z _nvfuser/test_torchscript.py::TestCudaFuser::test_device_constant SKIPPED (requires CUDA) [ 26%] 2023-03-31T07:14:24.8141900Z _nvfuser/test_torchscript.py::TestCudaFuser::test_disable_const_chunk_propagation_for_normalization SKIPPED (requires CUDA) [ 27%] 2023-03-31T07:14:24.8142317Z _nvfuser/test_torchscript.py::TestCudaFuser::test_disable_sibling_fuse SKIPPED (requires CUDA) [ 27%] 2023-03-31T07:14:24.8142793Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_inference_fusion SKIPPED (requires CUDA) [ 28%] 2023-03-31T07:14:24.8143191Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_train_nograd_fusion SKIPPED (not enough memory) [ 29%] 2023-03-31T07:14:24.8143598Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_train_nograd_prob_check SKIPPED (requires CUDA) [ 29%] 2023-03-31T07:14:24.8143997Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_training_fusion SKIPPED (requires CUDA) [ 30%] 2023-03-31T07:14:24.8144399Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dropout_training_prob_check SKIPPED (requires CUDA) [ 31%] 2023-03-31T07:14:24.8144770Z _nvfuser/test_torchscript.py::TestCudaFuser::test_dynamic_size SKIPPED (requires CUDA) [ 31%] 2023-03-31T07:14:24.8145137Z _nvfuser/test_torchscript.py::TestCudaFuser::test_expand SKIPPED (requires CUDA) [ 32%] 2023-03-31T07:14:24.8145560Z _nvfuser/test_torchscript.py::TestCudaFuser::test_fix_shape_expression_bn SKIPPED (requires CUDA) [ 32%] 2023-03-31T07:14:24.8146006Z _nvfuser/test_torchscript.py::TestCudaFuser::test_flatten SKIPPED (skipping this test since flatten is disabled now) [ 33%] 2023-03-31T07:14:24.8146390Z _nvfuser/test_torchscript.py::TestCudaFuser::test_gelu SKIPPED (requires CUDA) [ 34%] 2023-03-31T07:14:24.8146754Z _nvfuser/test_torchscript.py::TestCudaFuser::test_grad_sum_to_size SKIPPED (requires CUDA) [ 34%] 2023-03-31T07:14:24.8147158Z _nvfuser/test_torchscript.py::TestCudaFuser::test_graph_for_with_missing_optimized_engine SKIPPED (requires CUDA) [ 35%] 2023-03-31T07:14:24.8147547Z _nvfuser/test_torchscript.py::TestCudaFuser::test_graph_rng SKIPPED (requires CUDA) [ 36%] 2023-03-31T07:14:24.8147903Z _nvfuser/test_torchscript.py::TestCudaFuser::test_half SKIPPED (requires CUDA) [ 36%] 2023-03-31T07:14:24.8148271Z _nvfuser/test_torchscript.py::TestCudaFuser::test_high_rank_fusion SKIPPED (requires CUDA) [ 37%] 2023-03-31T07:14:24.8148648Z _nvfuser/test_torchscript.py::TestCudaFuser::test_inf_quick_patch SKIPPED (requires CUDA) [ 37%] 2023-03-31T07:14:24.8149013Z _nvfuser/test_torchscript.py::TestCudaFuser::test_inplace_removal SKIPPED (requires CUDA) [ 38%] 2023-03-31T07:14:24.8149398Z _nvfuser/test_torchscript.py::TestCudaFuser::test_input_output_passthrough SKIPPED (requires CUDA) [ 39%] 2023-03-31T07:14:24.8149788Z _nvfuser/test_torchscript.py::TestCudaFuser::test_int_tensor_input SKIPPED (requires CUDA) [ 39%] 2023-03-31T07:14:24.8150160Z _nvfuser/test_torchscript.py::TestCudaFuser::test_issue1445_fusion SKIPPED (requires CUDA) [ 40%] 2023-03-31T07:14:24.8150554Z _nvfuser/test_torchscript.py::TestCudaFuser::test_issue_1785 SKIPPED (requires CUDA) [ 41%] 2023-03-31T07:14:24.8150931Z _nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_autodiff SKIPPED (requires CUDA) [ 41%] 2023-03-31T07:14:24.8151346Z _nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_parser SKIPPED (requires CUDA) [ 42%] 2023-03-31T07:14:24.8151733Z _nvfuser/test_torchscript.py::TestCudaFuser::test_layer_norm_trivial_reduce_dim SKIPPED (requires CUDA) [ 43%] 2023-03-31T07:14:24.8152117Z _nvfuser/test_torchscript.py::TestCudaFuser::test_linear SKIPPED (requires CUDA) [ 43%] 2023-03-31T07:14:24.8152492Z _nvfuser/test_torchscript.py::TestCudaFuser::test_linear_symbolic_shapes SKIPPED (requires CUDA) [ 44%] 2023-03-31T07:14:24.8152884Z _nvfuser/test_torchscript.py::TestCudaFuser::test_multiple_device_pw SKIPPED (requires CUDA) [ 44%] 2023-03-31T07:14:24.8153262Z _nvfuser/test_torchscript.py::TestCudaFuser::test_native_batch_norm_backward SKIPPED (requires CUDA) [ 45%] 2023-03-31T07:14:24.8153653Z _nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm SKIPPED (requires CUDA) [ 46%] 2023-03-31T07:14:24.8154040Z _nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm_bfloat SKIPPED (requires CUDA) [ 46%] 2023-03-31T07:14:24.8154433Z _nvfuser/test_torchscript.py::TestCudaFuser::test_native_layer_norm_half SKIPPED (requires CUDA) [ 47%] 2023-03-31T07:14:24.8154832Z _nvfuser/test_torchscript.py::TestCudaFuser::test_nested_view SKIPPED (skipping this test since view is disabled now) [ 48%] 2023-03-31T07:14:24.8155241Z _nvfuser/test_torchscript.py::TestCudaFuser::test_no_tensor_input SKIPPED (requires CUDA) [ 48%] 2023-03-31T07:14:24.8155607Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm SKIPPED (requires CUDA) [ 49%] 2023-03-31T07:14:24.8155967Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm_bfloat SKIPPED (requires CUDA) [ 50%] 2023-03-31T07:14:24.8156326Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm_channels_last SKIPPED (requires CUDA) [ 50%] 2023-03-31T07:14:24.8156704Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm_half SKIPPED (requires CUDA) [ 51%] 2023-03-31T07:14:24.8157076Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm_half_layer SKIPPED (requires CUDA) [ 51%] 2023-03-31T07:14:24.8157433Z _nvfuser/test_torchscript.py::TestCudaFuser::test_norm_large SKIPPED (requires CUDA) [ 52%] 2023-03-31T07:14:24.8157867Z _nvfuser/test_torchscript.py::TestCudaFuser::test_normalization_partition SKIPPED (requires CUDA) [ 53%] 2023-03-31T07:14:24.8158319Z _nvfuser/test_torchscript.py::TestCudaFuser::test_nvfuser_comparison_callbacks_with_fallback SKIPPED (requires CUDA) [ 53%] 2023-03-31T07:14:24.8158758Z _nvfuser/test_torchscript.py::TestCudaFuser::test_nvfuser_comparison_callbacks_without_fallback SKIPPED (requires CUDA) [ 54%] 2023-03-31T07:14:24.8159155Z _nvfuser/test_torchscript.py::TestCudaFuser::test_overlapped_input SKIPPED (requires CUDA) [ 55%] 2023-03-31T07:14:24.8159551Z _nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation SKIPPED (requires CUDA) [ 55%] 2023-03-31T07:14:24.8159967Z _nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_0 SKIPPED (requires CUDA) [ 56%] 2023-03-31T07:14:24.8160402Z _nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_1_broken SKIPPED (requires CUDA) [ 56%] 2023-03-31T07:14:24.8160819Z _nvfuser/test_torchscript.py::TestCudaFuser::test_permutation_preservation_edge_case_2 SKIPPED (requires CUDA) [ 57%] 2023-03-31T07:14:24.8161213Z _nvfuser/test_torchscript.py::TestCudaFuser::test_permute SKIPPED (requires CUDA) [ 58%] 2023-03-31T07:14:24.8161600Z _nvfuser/test_torchscript.py::TestCudaFuser::test_pointwise_reference_tensor SKIPPED (requires CUDA) [ 58%] 2023-03-31T07:14:24.8161973Z _nvfuser/test_torchscript.py::TestCudaFuser::test_profile_ivalue SKIPPED (requires CUDA) [ 59%] 2023-03-31T07:14:24.8162367Z _nvfuser/test_torchscript.py::TestCudaFuser::test_profile_ivalue_multiple_profiles SKIPPED (requires CUDA) [ 60%] 2023-03-31T07:14:24.8162828Z _nvfuser/test_torchscript.py::TestCudaFuser::test_profiling_node SKIPPED (Skipped due to rand_like behavior change) [ 60%] 2023-03-31T07:14:24.8163578Z _nvfuser/test_torchscript.py::TestCudaFuser::test_pw_single_reduction_partition SKIPPED (requires CUDA) [ 61%] 2023-03-31T07:14:24.8163966Z _nvfuser/test_torchscript.py::TestCudaFuser::test_random_topo SKIPPED (requires CUDA) [ 62%] 2023-03-31T07:14:24.8164335Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction SKIPPED (requires CUDA) [ 62%] 2023-03-31T07:14:24.8164720Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_dtypes_axis SKIPPED (requires CUDA) [ 63%] 2023-03-31T07:14:24.8165112Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_empty_axes SKIPPED (requires CUDA) [ 63%] 2023-03-31T07:14:24.8165495Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_multiple_output SKIPPED (requires CUDA) [ 64%] 2023-03-31T07:14:24.8165892Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_permutation SKIPPED (requires CUDA) [ 65%] 2023-03-31T07:14:24.8166285Z _nvfuser/test_torchscript.py::TestCudaFuser::test_reduction_sizes_op SKIPPED (requires CUDA) [ 65%] 2023-03-31T07:14:24.8166686Z _nvfuser/test_torchscript.py::TestCudaFuser::test_remove_output_used_only_in_dtype SKIPPED (requires CUDA) [ 66%] 2023-03-31T07:14:24.8167053Z _nvfuser/test_torchscript.py::TestCudaFuser::test_rsub SKIPPED (requires CUDA) [ 67%] 2023-03-31T07:14:24.8167422Z _nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_cuda_tensor SKIPPED (requires CUDA) [ 67%] 2023-03-31T07:14:24.8167800Z _nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_input SKIPPED (requires CUDA) [ 68%] 2023-03-31T07:14:24.8168157Z _nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_tensor SKIPPED (requires CUDA) [ 68%] 2023-03-31T07:14:24.8168541Z _nvfuser/test_torchscript.py::TestCudaFuser::test_scalar_tensor_permuted SKIPPED (requires CUDA) [ 69%] 2023-03-31T07:14:24.8168950Z _nvfuser/test_torchscript.py::TestCudaFuser::test_scheduler_with_polymorphic_broadcast SKIPPED (requires CUDA) [ 70%] 2023-03-31T07:14:24.8169408Z _nvfuser/test_torchscript.py::TestCudaFuser::test_shape_expression SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 70%] 2023-03-31T07:14:24.8169821Z _nvfuser/test_torchscript.py::TestCudaFuser::test_sibling_fusion SKIPPED (requires CUDA) [ 71%] 2023-03-31T07:14:24.8170265Z _nvfuser/test_torchscript.py::TestCudaFuser::test_sibling_fusion_no_scalar_inputs SKIPPED (requires CUDA) [ 72%] 2023-03-31T07:14:24.8170728Z _nvfuser/test_torchscript.py::TestCudaFuser::test_single_reduction_broadcast SKIPPED (requires CUDA) [ 72%] 2023-03-31T07:14:24.8171116Z _nvfuser/test_torchscript.py::TestCudaFuser::test_singleton_fusion SKIPPED (requires CUDA) [ 73%] 2023-03-31T07:14:24.8171477Z _nvfuser/test_torchscript.py::TestCudaFuser::test_skip_parser SKIPPED (requires CUDA) [ 74%] 2023-03-31T07:14:24.8171843Z _nvfuser/test_torchscript.py::TestCudaFuser::test_softmax SKIPPED (requires CUDA) [ 74%] 2023-03-31T07:14:24.8172215Z _nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_bfloat SKIPPED (requires CUDA) [ 75%] 2023-03-31T07:14:24.8172574Z _nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_dtype SKIPPED (requires CUDA) [ 75%] 2023-03-31T07:14:24.8172949Z _nvfuser/test_torchscript.py::TestCudaFuser::test_softmax_half SKIPPED (requires CUDA) [ 76%] 2023-03-31T07:14:24.8173320Z _nvfuser/test_torchscript.py::TestCudaFuser::test_softplus_fuser SKIPPED (requires CUDA) [ 77%] 2023-03-31T07:14:24.8173745Z _nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 77%] 2023-03-31T07:14:24.8174204Z _nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze_negative_dim SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 78%] 2023-03-31T07:14:24.8174681Z _nvfuser/test_torchscript.py::TestCudaFuser::test_squeeze_zero SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 79%] 2023-03-31T07:14:24.8175144Z _nvfuser/test_torchscript.py::TestCudaFuser::test_strict_fusion SKIPPED (requires CUDA) [ 79%] 2023-03-31T07:14:24.8175511Z _nvfuser/test_torchscript.py::TestCudaFuser::test_sum_to_one SKIPPED (requires CUDA) [ 80%] 2023-03-31T07:14:24.8175899Z _nvfuser/test_torchscript.py::TestCudaFuser::test_sum_to_size SKIPPED (requires CUDA) [ 81%] 2023-03-31T07:14:24.8176269Z _nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops SKIPPED (requires CUDA) [ 81%] 2023-03-31T07:14:24.8176666Z _nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops_integer_compatibility SKIPPED (requires CUDA) [ 82%] 2023-03-31T07:14:24.8177076Z _nvfuser/test_torchscript.py::TestCudaFuser::test_ternary_ops_type_promotion SKIPPED (requires CUDA) [ 82%] 2023-03-31T07:14:24.8177446Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_boolean SKIPPED (requires CUDA) [ 83%] 2023-03-31T07:14:24.8177804Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_copy SKIPPED (requires CUDA) [ 84%] 2023-03-31T07:14:24.8178180Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_bf16_to_bf16 SKIPPED (requires CUDA) [ 84%] 2023-03-31T07:14:24.8178550Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_bf16_to_fp32 SKIPPED (requires CUDA) [ 85%] 2023-03-31T07:14:24.8178928Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp16_to_fp16 SKIPPED (requires CUDA) [ 86%] 2023-03-31T07:14:24.8179311Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp16_to_fp32 SKIPPED (requires CUDA) [ 86%] 2023-03-31T07:14:24.8179690Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp32_to_bf16 SKIPPED (requires CUDA) [ 87%] 2023-03-31T07:14:24.8180050Z _nvfuser/test_torchscript.py::TestCudaFuser::test_to_dtype_fp32_to_fp16 SKIPPED (requires CUDA) [ 87%] 2023-03-31T07:14:24.8180419Z _nvfuser/test_torchscript.py::TestCudaFuser::test_transpose SKIPPED (requires CUDA) [ 88%] 2023-03-31T07:14:24.8180798Z _nvfuser/test_torchscript.py::TestCudaFuser::test_transpose_default SKIPPED (requires CUDA) [ 89%] 2023-03-31T07:14:24.8181185Z _nvfuser/test_torchscript.py::TestCudaFuser::test_trivial_reduction SKIPPED (requires CUDA) [ 89%] 2023-03-31T07:14:24.8181543Z _nvfuser/test_torchscript.py::TestCudaFuser::test_type_as_op SKIPPED (requires CUDA) [ 90%] 2023-03-31T07:14:24.8181916Z _nvfuser/test_torchscript.py::TestCudaFuser::test_type_inference SKIPPED (requires CUDA) [ 91%] 2023-03-31T07:14:24.8182319Z _nvfuser/test_torchscript.py::TestCudaFuser::test_unary_bitwise SKIPPED (requires CUDA) [ 91%] 2023-03-31T07:14:24.8182774Z _nvfuser/test_torchscript.py::TestCudaFuser::test_unary_ops SKIPPED (requires CUDA) [ 92%] 2023-03-31T07:14:24.8183198Z _nvfuser/test_torchscript.py::TestCudaFuser::test_unsqueeze SKIPPED (skipping this test since squeeze/unsqueeze is disabled now) [ 93%] 2023-03-31T07:14:24.8183610Z _nvfuser/test_torchscript.py::TestCudaFuser::test_variance SKIPPED (requires CUDA) [ 93%] 2023-03-31T07:14:24.8183987Z _nvfuser/test_torchscript.py::TestCudaFuser::test_variance_profiling SKIPPED (requires CUDA) [ 94%] 2023-03-31T07:14:24.8184383Z _nvfuser/test_torchscript.py::TestCudaFuser::test_view SKIPPED (skipping this test since view is disabled now) [ 94%] 2023-03-31T07:14:24.8184783Z _nvfuser/test_torchscript.py::TestCudaFuser::test_view_before_permute SKIPPED (requires CUDA) [ 95%] 2023-03-31T07:14:24.8185214Z _nvfuser/test_torchscript.py::TestCudaFuser::test_view_copy_graph_guard SKIPPED (skipping this test since reshape is disabled now) [ 96%] 2023-03-31T07:14:24.8185692Z _nvfuser/test_torchscript.py::TestCudaFuser::test_view_copy_graph_guard_double_fusion SKIPPED (skipping this test since view is disabled now) [ 96%] 2023-03-31T07:14:24.8186130Z _nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_can_be_enabled_nvfuser PASSED [ 97%] 2023-03-31T07:14:24.8186561Z _nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_context_manager_test SKIPPED (requires CUDA) [ 98%] 2023-03-31T07:14:24.8186999Z _nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser SKIPPED (requires CUDA) [ 98%] 2023-03-31T07:14:24.8187458Z _nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser_cpu PASSED [ 99%] 2023-03-31T07:14:24.8187861Z _nvfuser/test_torchscript.py::TestEnableDisableCudaFuser::test_register_fuser_rocm SKIPPED (requires CUDA) [100%] 2023-03-31T07:14:24.8188094Z 2023-03-31T07:14:24.8188563Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/_nvfuser.test_torchscript/_nvfuser.test_torchscript-5ab7f0515610d7c1.xml - 2023-03-31T07:14:24.8188973Z ======================== 2 passed, 156 skipped in 0.21s ======================== 2023-03-31T07:14:24.8189301Z If in CI, skip info is located in the xml test reports, please either go to s3 or the hud to download them 2023-03-31T07:14:24.8189496Z 2023-03-31T07:14:24.8189962Z ##[endgroup] 2023-03-31T07:14:24.8190422Z FINISHED PRINTING LOG FILE of _nvfuser/test_torchscript (/var/lib/jenkins/workspace/test/test-reports/_nvfuser-test_torchscript_nyfqim5r.log) 2023-03-31T07:14:24.8190678Z 2023-03-31T07:14:24.8190897Z Running test_cuda_nvml_based_avail ... [2023-03-31 07:14:24.809617] 2023-03-31T07:14:24.8191562Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cuda_nvml_based_avail.py', '-v', '--subprocess', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:24.809941] 2023-03-31T07:14:26.6123969Z 2023-03-31T07:14:26.6124582Z Expand the folded group to see the log file of test_cuda_nvml_based_avail 2023-03-31T07:14:26.6125587Z ##[group]PRINTING LOG FILE of test_cuda_nvml_based_avail (/var/lib/jenkins/workspace/test/test-reports/test_cuda_nvml_based_avail_mp2ce410.log) 2023-03-31T07:14:26.6125958Z CUDA not available, skipping tests 2023-03-31T07:14:26.6126181Z Running 0 items in this shard 2023-03-31T07:14:26.6126310Z 2023-03-31T07:14:26.6126384Z no tests collected in 0.00s 2023-03-31T07:14:26.6126516Z 2023-03-31T07:14:26.6126740Z ##[endgroup] 2023-03-31T07:14:26.6127296Z FINISHED PRINTING LOG FILE of test_cuda_nvml_based_avail (/var/lib/jenkins/workspace/test/test-reports/test_cuda_nvml_based_avail_mp2ce410.log) 2023-03-31T07:14:26.6127686Z 2023-03-31T07:14:26.6128661Z Running test_cuda_trace ... [2023-03-31 07:14:26.612680] 2023-03-31T07:14:26.6133045Z Executing ['/opt/conda/envs/py_3.8/bin/python', '-bb', 'test_cuda_trace.py', '-v', '--subprocess', '--use-pytest', '-vv', '-rfEX', '-p', 'no:xdist', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2023-03-31 07:14:26.613030] 2023-03-31T07:14:28.3904709Z 2023-03-31T07:14:28.3905208Z Expand the folded group to see the log file of test_cuda_trace 2023-03-31T07:14:28.3906041Z ##[group]PRINTING LOG FILE of test_cuda_trace (/var/lib/jenkins/workspace/test/test-reports/test_cuda_trace_1h94gxib.log) 2023-03-31T07:14:28.3906392Z CUDA not available, skipping tests 2023-03-31T07:14:28.3906654Z Running 0 items in this shard 2023-03-31T07:14:28.3906804Z 2023-03-31T07:14:28.3906895Z no tests collected in 0.00s 2023-03-31T07:14:28.3907021Z 2023-03-31T07:14:28.3907269Z ##[endgroup] 2023-03-31T07:14:28.3907824Z FINISHED PRINTING LOG FILE of test_cuda_trace (/var/lib/jenkins/workspace/test/test-reports/test_cuda_trace_1h94gxib.log) 2023-03-31T07:14:28.3908053Z 2023-03-31T07:14:28.7879454Z 2023-03-31T07:14:28.7879877Z real 103m9.840s 2023-03-31T07:14:28.7880292Z user 274m13.057s 2023-03-31T07:14:28.7880586Z sys 10m30.980s 2023-03-31T07:14:28.7880927Z + assert_git_not_dirty 2023-03-31T07:14:28.7881523Z + [[ linux-bionic-py3.8-clang9 != *rocm* ]] 2023-03-31T07:14:28.7881895Z + [[ linux-bionic-py3.8-clang9 != *xla* ]] 2023-03-31T07:14:28.7882728Z ++ git status --porcelain 2023-03-31T07:14:38.2903394Z + git_status= 2023-03-31T07:14:38.2903971Z + [[ -n '' ]] 2023-03-31T07:14:38.2904199Z + test_aten 2023-03-31T07:14:38.2904513Z + [[ linux-bionic-py3.8-clang9 != *asan* ]] 2023-03-31T07:14:38.2904828Z + [[ linux-bionic-py3.8-clang9 != *rocm* ]] 2023-03-31T07:14:38.2905112Z + echo 'Running ATen tests with pytorch lib' 2023-03-31T07:14:38.2905355Z Running ATen tests with pytorch lib 2023-03-31T07:14:38.2905797Z + [[ -n '' ]] 2023-03-31T07:14:38.2906035Z + echo 'Running test with the build folder' 2023-03-31T07:14:38.2906270Z Running test with the build folder 2023-03-31T07:14:38.2906486Z + TEST_BASE_DIR=build/bin 2023-03-31T07:14:38.2906898Z + ln -sf /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libc10.so build/bin 2023-03-31T07:14:38.2982295Z + ln -sf '/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libcaffe2*' build/bin 2023-03-31T07:14:38.2990750Z + ln -sf '/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libmkldnn*' build/bin 2023-03-31T07:14:38.2999378Z + ln -sf '/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libnccl*' build/bin 2023-03-31T07:14:38.3008824Z + ln -sf /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtorch.so /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtorch_cpu.so /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtorch_global_deps.so /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtorch_python.so /opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtorchbind_test.so build/bin 2023-03-31T07:14:38.3017306Z + ln -sf '/opt/conda/envs/py_3.8/lib/python3.8/site-packages/torch/lib/libtbb*' build/bin 2023-03-31T07:14:38.3024859Z + ls build/bin 2023-03-31T07:14:38.3103253Z CMakeFiles example_allreduce 2023-03-31T07:14:38.3103743Z CTestTestfile.cmake extension_backend_test 2023-03-31T07:14:38.3104217Z CppSignature_test half_test 2023-03-31T07:14:38.3104632Z Dict_test inline_container_test 2023-03-31T07:14:38.3105013Z Dimname_test ivalue_test 2023-03-31T07:14:38.3105419Z FileStoreTest kernel_function_legacy_test 2023-03-31T07:14:38.3105867Z HashStoreTest kernel_function_test 2023-03-31T07:14:38.3106313Z IListRef_test kernel_lambda_legacy_test 2023-03-31T07:14:38.3106746Z KernelFunction_test kernel_lambda_test 2023-03-31T07:14:38.3107144Z List_test kernel_stackbased_test 2023-03-31T07:14:38.3107561Z Makefile lazy_tensor_test 2023-03-31T07:14:38.3107951Z MaybeOwned_test legacy_vmap_test 2023-03-31T07:14:38.3108210Z NamedTensor_test libc10.so 2023-03-31T07:14:38.3108574Z ProcessGroupGlooTest 'libcaffe2*' 2023-03-31T07:14:38.3108863Z StorageUtils_test 'libmkldnn*' 2023-03-31T07:14:38.3109111Z TCPStoreTest 'libnccl*' 2023-03-31T07:14:38.3109587Z aot_model_compiler_test 'libtbb*' 2023-03-31T07:14:38.3109918Z apply_utils_test libtorch.so 2023-03-31T07:14:38.3110121Z atest libtorch_cpu.so 2023-03-31T07:14:38.3110363Z backend_fallback_test libtorch_global_deps.so 2023-03-31T07:14:38.3110605Z basic libtorch_python.so 2023-03-31T07:14:38.3110825Z broadcast_test libtorchbind_test.so 2023-03-31T07:14:38.3111089Z c10_Array_test make_boxed_from_unboxed_functor_test 2023-03-31T07:14:38.3111337Z c10_Bitset_test math_kernel_test 2023-03-31T07:14:38.3111556Z c10_C++17_test memory_format_test 2023-03-31T07:14:38.3111836Z c10_CompileTimeFunctionPointer_test memory_overlapping_test 2023-03-31T07:14:38.3112128Z c10_ConstexprCrc_test mobile_memory_cleanup 2023-03-31T07:14:38.3112375Z c10_DeadlockDetection_test native_test 2023-03-31T07:14:38.3112625Z c10_DeviceGuard_test op_allowlist_test 2023-03-31T07:14:38.3112869Z c10_Device_test op_registration_test 2023-03-31T07:14:38.3113127Z c10_DispatchKeySet_test operator_name_test 2023-03-31T07:14:38.3113353Z c10_Half_test operators_test 2023-03-31T07:14:38.3113621Z c10_InlineDeviceGuard_test packedtensoraccessor_test 2023-03-31T07:14:38.3113908Z c10_InlineStreamGuard_test parallel_benchmark 2023-03-31T07:14:38.3114134Z c10_LeftRight_test pow_test 2023-03-31T07:14:38.3114363Z c10_Metaprogramming_test protoc 2023-03-31T07:14:38.3114674Z c10_SizesAndStrides_test protoc-3.13.0.0 2023-03-31T07:14:38.3114969Z c10_SmallVectorTest quantized_test 2023-03-31T07:14:38.3115210Z c10_StreamGuard_test reduce_ops_test 2023-03-31T07:14:38.3115457Z c10_SymInt_test reportMemoryUsage_test 2023-03-31T07:14:38.3115694Z c10_Synchronized_test scalar_tensor_test 2023-03-31T07:14:38.3115975Z c10_ThreadLocal_test scalar_test 2023-03-31T07:14:38.3116220Z c10_TypeIndex_test static_runtime_bench 2023-03-31T07:14:38.3116455Z c10_TypeList_test static_runtime_test 2023-03-31T07:14:38.3116706Z c10_TypeTraits_test stride_properties_test 2023-03-31T07:14:38.3116979Z c10_accumulate_test tensor_iterator_test 2023-03-31T07:14:38.3117207Z c10_bfloat16_test test_api 2023-03-31T07:14:38.3117417Z c10_complex_math_test test_cpp_rpc 2023-03-31T07:14:38.3117651Z c10_complex_test test_dist_autograd 2023-03-31T07:14:38.3117894Z c10_either_test test_edge_op_registration 2023-03-31T07:14:38.3118113Z c10_exception_test test_jit 2023-03-31T07:14:38.3118327Z c10_flags_test test_lazy 2023-03-31T07:14:38.3118562Z c10_intrusive_ptr_benchmark test_mobile_nnc 2023-03-31T07:14:38.3118796Z c10_intrusive_ptr_test test_parallel 2023-03-31T07:14:38.3119026Z c10_irange_test test_tensorexpr 2023-03-31T07:14:38.3119257Z c10_logging_test thread_init_test 2023-03-31T07:14:38.3119487Z c10_optional_test torch_shm_manager 2023-03-31T07:14:38.3119735Z c10_ordered_preserving_dict_test tutorial_tensorexpr 2023-03-31T07:14:38.3119993Z c10_registry_test type_ptr_test 2023-03-31T07:14:38.3120210Z c10_ssize_test type_test 2023-03-31T07:14:38.3120426Z c10_string_view_test undefined_tensor_test 2023-03-31T07:14:38.3120662Z c10_tempfile_test variant_test 2023-03-31T07:14:38.3120898Z c10_typeid_test vec_test_all_types_AVX2 2023-03-31T07:14:38.3121142Z cmake_install.cmake vec_test_all_types_AVX512 2023-03-31T07:14:38.3121407Z cpu_generator_test vec_test_all_types_DEFAULT 2023-03-31T07:14:38.3121683Z cpu_profiling_allocator_test verify_api_visibility 2023-03-31T07:14:38.3121912Z cpu_rng_test weakref_test 2023-03-31T07:14:38.3122140Z dispatch_key_set_test wrapdim_test 2023-03-31T07:14:38.3122376Z dlconvertor_test xla_tensor_test 2023-03-31T07:14:38.3122592Z + aten/tools/run_tests.sh build/bin 2023-03-31T07:14:38.3167079Z + set -e 2023-03-31T07:14:38.3169981Z ++ dirname aten/tools/run_tests.sh 2023-03-31T07:14:38.3184414Z + VALGRIND_SUP=/var/lib/jenkins/workspace/aten/tools/valgrind.sup 2023-03-31T07:14:38.3184891Z + pushd build/bin 2023-03-31T07:14:38.3185153Z ~/workspace/build/bin ~/workspace 2023-03-31T07:14:38.3185359Z + VALGRIND=OFF 2023-03-31T07:14:38.3185520Z + ./basic 2023-03-31T07:14:39.1455013Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:39.1455697Z [==========] Running 5 tests from 1 test suite. 2023-03-31T07:14:39.1456047Z [----------] Global test environment set-up. 2023-03-31T07:14:39.1456352Z [----------] 5 tests from BasicTest 2023-03-31T07:14:39.1456629Z [ RUN ] BasicTest.BasicTestCPU 2023-03-31T07:14:39.2669998Z 5 ms 2023-03-31T07:14:39.2725035Z 1 ms 2023-03-31T07:14:39.2741861Z 0 ms 2023-03-31T07:14:39.3579052Z [ OK ] BasicTest.BasicTestCPU (212 ms) 2023-03-31T07:14:39.3579565Z [ RUN ] BasicTest.BasicTestHalfCPU 2023-03-31T07:14:39.3627560Z 2 ms 2023-03-31T07:14:39.3646419Z 0 ms 2023-03-31T07:14:39.3657820Z 1 ms 2023-03-31T07:14:39.3815743Z [ OK ] BasicTest.BasicTestHalfCPU (23 ms) 2023-03-31T07:14:39.3816217Z [ RUN ] BasicTest.BasicTestCUDA 2023-03-31T07:14:39.3816671Z [ OK ] BasicTest.BasicTestCUDA (0 ms) 2023-03-31T07:14:39.3817120Z [ RUN ] BasicTest.FactoryMethodsTest 2023-03-31T07:14:39.3857969Z [ OK ] BasicTest.FactoryMethodsTest (4 ms) 2023-03-31T07:14:39.3858634Z [ RUN ] BasicTest.BasicStdTestCPU 2023-03-31T07:14:39.3858916Z Simple example: called once 2023-03-31T07:14:39.3861855Z throw: call_once will retry 2023-03-31T07:14:39.3863707Z throw: call_once will retry 2023-03-31T07:14:39.3866047Z Didn't throw, call_once will not attempt again 2023-03-31T07:14:39.3866905Z [ OK ] BasicTest.BasicStdTestCPU (1 ms) 2023-03-31T07:14:39.3867381Z [----------] 5 tests from BasicTest (241 ms total) 2023-03-31T07:14:39.3867623Z 2023-03-31T07:14:39.3867880Z [----------] Global test environment tear-down 2023-03-31T07:14:39.3868353Z [==========] 5 tests from 1 test suite ran. (241 ms total) 2023-03-31T07:14:39.3868737Z [ PASSED ] 5 tests. 2023-03-31T07:14:39.5234748Z + ./atest 2023-03-31T07:14:39.7863400Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:39.7863991Z [==========] Running 16 tests from 1 test suite. 2023-03-31T07:14:39.7864316Z [----------] Global test environment set-up. 2023-03-31T07:14:39.7864611Z [----------] 16 tests from atest 2023-03-31T07:14:39.7864872Z [ RUN ] atest.operators 2023-03-31T07:14:39.7888699Z [ OK ] atest.operators (2 ms) 2023-03-31T07:14:39.7888998Z [ RUN ] atest.logical_and_operators 2023-03-31T07:14:39.7903075Z [ OK ] atest.logical_and_operators (1 ms) 2023-03-31T07:14:39.7903389Z [ RUN ] atest.logical_or_operators 2023-03-31T07:14:39.7908614Z [ OK ] atest.logical_or_operators (0 ms) 2023-03-31T07:14:39.7909161Z [ RUN ] atest.logical_xor_operators 2023-03-31T07:14:39.7909721Z [ OK ] atest.logical_xor_operators (0 ms) 2023-03-31T07:14:39.7910107Z [ RUN ] atest.lt_operators 2023-03-31T07:14:39.7910433Z [ OK ] atest.lt_operators (0 ms) 2023-03-31T07:14:39.7910705Z [ RUN ] atest.le_operators 2023-03-31T07:14:39.7910992Z [ OK ] atest.le_operators (0 ms) 2023-03-31T07:14:39.7911249Z [ RUN ] atest.gt_operators 2023-03-31T07:14:39.7916934Z [ OK ] atest.gt_operators (0 ms) 2023-03-31T07:14:39.7917378Z [ RUN ] atest.ge_operators 2023-03-31T07:14:39.7917844Z [ OK ] atest.ge_operators (0 ms) 2023-03-31T07:14:39.7918243Z [ RUN ] atest.eq_operators 2023-03-31T07:14:39.7918802Z [ OK ] atest.eq_operators (0 ms) 2023-03-31T07:14:39.7919139Z [ RUN ] atest.ne_operators 2023-03-31T07:14:39.7919406Z [ OK ] atest.ne_operators (0 ms) 2023-03-31T07:14:39.7919678Z [ RUN ] atest.add_operators 2023-03-31T07:14:39.7919957Z [ OK ] atest.add_operators (0 ms) 2023-03-31T07:14:39.7920216Z [ RUN ] atest.max_operators 2023-03-31T07:14:39.7933273Z [ OK ] atest.max_operators (1 ms) 2023-03-31T07:14:39.7933777Z [ RUN ] atest.min_operators 2023-03-31T07:14:39.7934075Z [ OK ] atest.min_operators (0 ms) 2023-03-31T07:14:39.7934383Z [ RUN ] atest.sigmoid_backward_operator 2023-03-31T07:14:39.7943182Z [ OK ] atest.sigmoid_backward_operator (0 ms) 2023-03-31T07:14:39.7943629Z [ RUN ] atest.fmod_tensor_operators 2023-03-31T07:14:39.7943932Z [ OK ] atest.fmod_tensor_operators (0 ms) 2023-03-31T07:14:39.7944207Z [ RUN ] atest.atest 2023-03-31T07:14:39.8098991Z [ OK ] atest.atest (15 ms) 2023-03-31T07:14:39.8099517Z [----------] 16 tests from atest (23 ms total) 2023-03-31T07:14:39.8099672Z 2023-03-31T07:14:39.8099854Z [----------] Global test environment tear-down 2023-03-31T07:14:39.8100186Z [==========] 16 tests from 1 test suite ran. (23 ms total) 2023-03-31T07:14:39.8100465Z [ PASSED ] 16 tests. 2023-03-31T07:14:39.8790493Z + ./scalar_test 2023-03-31T07:14:40.1398986Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:40.1400115Z [==========] Running 4 tests from 1 test suite. 2023-03-31T07:14:40.1400444Z [----------] Global test environment set-up. 2023-03-31T07:14:40.1400838Z [----------] 4 tests from TestScalar 2023-03-31T07:14:40.1401120Z [ RUN ] TestScalar.TestScalar 2023-03-31T07:14:40.1401330Z H2: 3 257 3 1 2023-03-31T07:14:40.1526783Z [ OK ] TestScalar.TestScalar (12 ms) 2023-03-31T07:14:40.1527300Z [ RUN ] TestScalar.TestConj 2023-03-31T07:14:40.1527837Z [ OK ] TestScalar.TestConj (0 ms) 2023-03-31T07:14:40.1528154Z [ RUN ] TestScalar.TestEqual 2023-03-31T07:14:40.1528438Z [ OK ] TestScalar.TestEqual (0 ms) 2023-03-31T07:14:40.1528748Z [ RUN ] TestScalar.TestFormatting 2023-03-31T07:14:40.1529068Z [ OK ] TestScalar.TestFormatting (0 ms) 2023-03-31T07:14:40.1529393Z [----------] 4 tests from TestScalar (12 ms total) 2023-03-31T07:14:40.1529558Z 2023-03-31T07:14:40.1529734Z [----------] Global test environment tear-down 2023-03-31T07:14:40.1530061Z [==========] 4 tests from 1 test suite ran. (12 ms total) 2023-03-31T07:14:40.1530337Z [ PASSED ] 4 tests. 2023-03-31T07:14:40.2218347Z + ./broadcast_test 2023-03-31T07:14:40.4834757Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:40.4835408Z [==========] Running 1 test from 1 test suite. 2023-03-31T07:14:40.4835727Z [----------] Global test environment set-up. 2023-03-31T07:14:40.4836118Z [----------] 1 test from BroadcastTest 2023-03-31T07:14:40.4836420Z [ RUN ] BroadcastTest.Broadcast 2023-03-31T07:14:40.5449105Z [ OK ] BroadcastTest.Broadcast (61 ms) 2023-03-31T07:14:40.5449530Z [----------] 1 test from BroadcastTest (61 ms total) 2023-03-31T07:14:40.5449719Z 2023-03-31T07:14:40.5449904Z [----------] Global test environment tear-down 2023-03-31T07:14:40.5450233Z [==========] 1 test from 1 test suite ran. (61 ms total) 2023-03-31T07:14:40.5450491Z [ PASSED ] 1 test. 2023-03-31T07:14:40.6147986Z + ./wrapdim_test 2023-03-31T07:14:40.8753203Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:40.8754111Z [==========] Running 1 test from 1 test suite. 2023-03-31T07:14:40.8754564Z [----------] Global test environment set-up. 2023-03-31T07:14:40.8754880Z [----------] 1 test from TestWrapdim 2023-03-31T07:14:40.8755252Z [ RUN ] TestWrapdim.TestWrapdim 2023-03-31T07:14:40.8760529Z [ OK ] TestWrapdim.TestWrapdim (0 ms) 2023-03-31T07:14:40.8761001Z [----------] 1 test from TestWrapdim (0 ms total) 2023-03-31T07:14:40.8761166Z 2023-03-31T07:14:40.8761344Z [----------] Global test environment tear-down 2023-03-31T07:14:40.8761771Z [==========] 1 test from 1 test suite ran. (0 ms total) 2023-03-31T07:14:40.8762025Z [ PASSED ] 1 test. 2023-03-31T07:14:40.9454209Z + ./apply_utils_test 2023-03-31T07:14:41.2090272Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:41.2090934Z [==========] Running 6 tests from 1 test suite. 2023-03-31T07:14:41.2091285Z [----------] Global test environment set-up. 2023-03-31T07:14:41.2091607Z [----------] 6 tests from ApplyUtilsTest 2023-03-31T07:14:41.2091922Z [ RUN ] ApplyUtilsTest.Contiguous2D 2023-03-31T07:14:41.2136433Z [ OK ] ApplyUtilsTest.Contiguous2D (4 ms) 2023-03-31T07:14:41.2136783Z [ RUN ] ApplyUtilsTest.Small2D 2023-03-31T07:14:41.2137102Z [ OK ] ApplyUtilsTest.Small2D (0 ms) 2023-03-31T07:14:41.2137408Z [ RUN ] ApplyUtilsTest._2D 2023-03-31T07:14:41.2144993Z [ OK ] ApplyUtilsTest._2D (0 ms) 2023-03-31T07:14:41.2145455Z [ RUN ] ApplyUtilsTest._3D 2023-03-31T07:14:41.2146082Z [ OK ] ApplyUtilsTest._3D (0 ms) 2023-03-31T07:14:41.2146388Z [ RUN ] ApplyUtilsTest.Medium3D 2023-03-31T07:14:41.2156778Z [ OK ] ApplyUtilsTest.Medium3D (1 ms) 2023-03-31T07:14:41.2157097Z [ RUN ] ApplyUtilsTest._10D 2023-03-31T07:14:41.2826863Z [ OK ] ApplyUtilsTest._10D (66 ms) 2023-03-31T07:14:41.2827507Z [----------] 6 tests from ApplyUtilsTest (73 ms total) 2023-03-31T07:14:41.2827742Z 2023-03-31T07:14:41.2827927Z [----------] Global test environment tear-down 2023-03-31T07:14:41.2828257Z [==========] 6 tests from 1 test suite ran. (73 ms total) 2023-03-31T07:14:41.2828532Z [ PASSED ] 6 tests. 2023-03-31T07:14:41.3521521Z + ./dlconvertor_test 2023-03-31T07:14:41.6115338Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:41.6116108Z [==========] Running 2 tests from 1 test suite. 2023-03-31T07:14:41.6116466Z [----------] Global test environment set-up. 2023-03-31T07:14:41.6116846Z [----------] 2 tests from TestDlconvertor 2023-03-31T07:14:41.6117165Z [ RUN ] TestDlconvertor.TestDlconvertor 2023-03-31T07:14:41.6122437Z [ OK ] TestDlconvertor.TestDlconvertor (0 ms) 2023-03-31T07:14:41.6123369Z [ RUN ] TestDlconvertor.TestDlconvertorNoStrides 2023-03-31T07:14:41.6123821Z [ OK ] TestDlconvertor.TestDlconvertorNoStrides (0 ms) 2023-03-31T07:14:41.6124222Z [----------] 2 tests from TestDlconvertor (0 ms total) 2023-03-31T07:14:41.6124391Z 2023-03-31T07:14:41.6124569Z [----------] Global test environment tear-down 2023-03-31T07:14:41.6124897Z [==========] 2 tests from 1 test suite ran. (0 ms total) 2023-03-31T07:14:41.6125156Z [ PASSED ] 2 tests. 2023-03-31T07:14:41.6834823Z + ./native_test 2023-03-31T07:14:41.9437611Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:41.9438500Z [==========] Running 2 tests from 1 test suite. 2023-03-31T07:14:41.9438850Z [----------] Global test environment set-up. 2023-03-31T07:14:41.9439151Z [----------] 2 tests from TestNative 2023-03-31T07:14:41.9439688Z [ RUN ] TestNative.NativeTestCPU 2023-03-31T07:14:41.9639601Z [W TensorCompare.cpp:493] Warning: where received a uint8 condition tensor. This behavior is deprecated and will be removed in a future version of PyTorch. Use a boolean condition instead. (function operator()) 2023-03-31T07:14:41.9656384Z [ OK ] TestNative.NativeTestCPU (21 ms) 2023-03-31T07:14:41.9656892Z [ RUN ] TestNative.NativeTestGPU 2023-03-31T07:14:41.9657387Z [ OK ] TestNative.NativeTestGPU (0 ms) 2023-03-31T07:14:41.9657906Z [----------] 2 tests from TestNative (21 ms total) 2023-03-31T07:14:41.9658152Z 2023-03-31T07:14:41.9658364Z [----------] Global test environment tear-down 2023-03-31T07:14:41.9658786Z [==========] 2 tests from 1 test suite ran. (21 ms total) 2023-03-31T07:14:41.9659140Z [ PASSED ] 2 tests. 2023-03-31T07:14:42.0367244Z + ./scalar_tensor_test 2023-03-31T07:14:42.2963454Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:42.2964247Z [==========] Running 3 tests from 1 test suite. 2023-03-31T07:14:42.2964795Z [----------] Global test environment set-up. 2023-03-31T07:14:42.2965399Z [----------] 3 tests from TestScalarTensor 2023-03-31T07:14:42.2966054Z [ RUN ] TestScalarTensor.TestScalarTensorCPU 2023-03-31T07:14:42.5068525Z [ OK ] TestScalarTensor.TestScalarTensorCPU (210 ms) 2023-03-31T07:14:42.5069170Z [ RUN ] TestScalarTensor.TestScalarTensorCUDA 2023-03-31T07:14:42.5070040Z [ OK ] TestScalarTensor.TestScalarTensorCUDA (0 ms) 2023-03-31T07:14:42.5070524Z [ RUN ] TestScalarTensor.TestScalarTensorMPS 2023-03-31T07:14:42.5071017Z [ OK ] TestScalarTensor.TestScalarTensorMPS (0 ms) 2023-03-31T07:14:42.5071635Z [----------] 3 tests from TestScalarTensor (210 ms total) 2023-03-31T07:14:42.5071857Z 2023-03-31T07:14:42.5072109Z [----------] Global test environment tear-down 2023-03-31T07:14:42.5072561Z [==========] 3 tests from 1 test suite ran. (210 ms total) 2023-03-31T07:14:42.5072917Z [ PASSED ] 3 tests. 2023-03-31T07:14:42.5758318Z + [[ -x ./tensor_interop_test ]] 2023-03-31T07:14:42.5758582Z + ./undefined_tensor_test 2023-03-31T07:14:42.8415531Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:42.8416322Z [==========] Running 1 test from 1 test suite. 2023-03-31T07:14:42.8416648Z [----------] Global test environment set-up. 2023-03-31T07:14:42.8416973Z [----------] 1 test from TestUndefined 2023-03-31T07:14:42.8417283Z [ RUN ] TestUndefined.UndefinedTest 2023-03-31T07:14:42.8689281Z [ OK ] TestUndefined.UndefinedTest (27 ms) 2023-03-31T07:14:42.8689972Z [----------] 1 test from TestUndefined (27 ms total) 2023-03-31T07:14:42.8690160Z 2023-03-31T07:14:42.8690349Z [----------] Global test environment tear-down 2023-03-31T07:14:42.8690674Z [==========] 1 test from 1 test suite ran. (27 ms total) 2023-03-31T07:14:42.8690944Z [ PASSED ] 1 test. 2023-03-31T07:14:42.9384877Z + ./extension_backend_test 2023-03-31T07:14:43.1939495Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:43.1940282Z [==========] Running 1 test from 1 test suite. 2023-03-31T07:14:43.1940738Z [----------] Global test environment set-up. 2023-03-31T07:14:43.1941109Z [----------] 1 test from BackendExtensionTest 2023-03-31T07:14:43.1941460Z [ RUN ] BackendExtensionTest.TestRegisterOp 2023-03-31T07:14:43.1941848Z [ OK ] BackendExtensionTest.TestRegisterOp (0 ms) 2023-03-31T07:14:43.1942236Z [----------] 1 test from BackendExtensionTest (0 ms total) 2023-03-31T07:14:43.1942416Z 2023-03-31T07:14:43.1942908Z [----------] Global test environment tear-down 2023-03-31T07:14:43.1943303Z [==========] 1 test from 1 test suite ran. (0 ms total) 2023-03-31T07:14:43.1943577Z [ PASSED ] 1 test. 2023-03-31T07:14:43.2628507Z + ./lazy_tensor_test 2023-03-31T07:14:43.5212325Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:43.5213002Z [==========] Running 2 tests from 2 test suites. 2023-03-31T07:14:43.5213429Z [----------] Global test environment set-up. 2023-03-31T07:14:43.5213838Z [----------] 1 test from XlaTensorTest 2023-03-31T07:14:43.5214256Z [ RUN ] XlaTensorTest.TestNoStorage 2023-03-31T07:14:43.5214682Z [ OK ] XlaTensorTest.TestNoStorage (0 ms) 2023-03-31T07:14:43.5215115Z [----------] 1 test from XlaTensorTest (0 ms total) 2023-03-31T07:14:43.5215350Z 2023-03-31T07:14:43.5215563Z [----------] 1 test from LazyTensorTest 2023-03-31T07:14:43.5215987Z [ RUN ] LazyTensorTest.TestNoStorage 2023-03-31T07:14:43.5216463Z [ OK ] LazyTensorTest.TestNoStorage (0 ms) 2023-03-31T07:14:43.5216991Z [----------] 1 test from LazyTensorTest (0 ms total) 2023-03-31T07:14:43.5217260Z 2023-03-31T07:14:43.5217534Z [----------] Global test environment tear-down 2023-03-31T07:14:43.5218029Z [==========] 2 tests from 2 test suites ran. (0 ms total) 2023-03-31T07:14:43.5218416Z [ PASSED ] 2 tests. 2023-03-31T07:14:43.5905572Z + ./tensor_iterator_test 2023-03-31T07:14:43.8574709Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:43.8575446Z [==========] Running 65 tests from 1 test suite. 2023-03-31T07:14:43.8575790Z [----------] Global test environment set-up. 2023-03-31T07:14:43.8576249Z [----------] 65 tests from TensorIteratorTest 2023-03-31T07:14:43.8576585Z [ RUN ] TensorIteratorTest.CPUScalar 2023-03-31T07:14:43.8577011Z [ OK ] TensorIteratorTest.CPUScalar (0 ms) 2023-03-31T07:14:43.8577364Z [ RUN ] TensorIteratorTest.CPUScalarInputs 2023-03-31T07:14:43.8577816Z [ OK ] TensorIteratorTest.CPUScalarInputs (0 ms) 2023-03-31T07:14:43.8578178Z [ RUN ] TensorIteratorTest.MixedDevices 2023-03-31T07:14:43.8578592Z [ OK ] TensorIteratorTest.MixedDevices (0 ms) 2023-03-31T07:14:43.8578968Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Byte 2023-03-31T07:14:43.8818796Z [ OK ] TensorIteratorTest.SerialLoopUnary_Byte (24 ms) 2023-03-31T07:14:43.8819262Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Char 2023-03-31T07:14:43.8834677Z [ OK ] TensorIteratorTest.SerialLoopUnary_Char (1 ms) 2023-03-31T07:14:43.8835315Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Short 2023-03-31T07:14:43.8852319Z [ OK ] TensorIteratorTest.SerialLoopUnary_Short (1 ms) 2023-03-31T07:14:43.8852852Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Int 2023-03-31T07:14:43.8869737Z [ OK ] TensorIteratorTest.SerialLoopUnary_Int (1 ms) 2023-03-31T07:14:43.8870330Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Long 2023-03-31T07:14:43.8886812Z [ OK ] TensorIteratorTest.SerialLoopUnary_Long (1 ms) 2023-03-31T07:14:43.8887364Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Float 2023-03-31T07:14:43.8905458Z [ OK ] TensorIteratorTest.SerialLoopUnary_Float (1 ms) 2023-03-31T07:14:43.8906235Z [ RUN ] TensorIteratorTest.SerialLoopUnary_Double 2023-03-31T07:14:43.8923598Z [ OK ] TensorIteratorTest.SerialLoopUnary_Double (1 ms) 2023-03-31T07:14:43.8924330Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Byte 2023-03-31T07:14:43.8941224Z [ OK ] TensorIteratorTest.SerialLoopBinary_Byte (1 ms) 2023-03-31T07:14:43.8941961Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Char 2023-03-31T07:14:43.8959209Z [ OK ] TensorIteratorTest.SerialLoopBinary_Char (1 ms) 2023-03-31T07:14:43.8960031Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Short 2023-03-31T07:14:43.8976724Z [ OK ] TensorIteratorTest.SerialLoopBinary_Short (1 ms) 2023-03-31T07:14:43.8977464Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Int 2023-03-31T07:14:43.8994298Z [ OK ] TensorIteratorTest.SerialLoopBinary_Int (1 ms) 2023-03-31T07:14:43.8995013Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Long 2023-03-31T07:14:43.9012052Z [ OK ] TensorIteratorTest.SerialLoopBinary_Long (1 ms) 2023-03-31T07:14:43.9012803Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Float 2023-03-31T07:14:43.9029909Z [ OK ] TensorIteratorTest.SerialLoopBinary_Float (1 ms) 2023-03-31T07:14:43.9030648Z [ RUN ] TensorIteratorTest.SerialLoopBinary_Double 2023-03-31T07:14:43.9047785Z [ OK ] TensorIteratorTest.SerialLoopBinary_Double (1 ms) 2023-03-31T07:14:43.9048549Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Byte 2023-03-31T07:14:43.9065556Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Byte (1 ms) 2023-03-31T07:14:43.9066308Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Char 2023-03-31T07:14:43.9083249Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Char (1 ms) 2023-03-31T07:14:43.9084048Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Short 2023-03-31T07:14:43.9100732Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Short (1 ms) 2023-03-31T07:14:43.9101490Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Int 2023-03-31T07:14:43.9118603Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Int (1 ms) 2023-03-31T07:14:43.9119585Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Long 2023-03-31T07:14:43.9136341Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Long (1 ms) 2023-03-31T07:14:43.9137155Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Float 2023-03-31T07:14:43.9154291Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Float (1 ms) 2023-03-31T07:14:43.9155065Z [ RUN ] TensorIteratorTest.SerialLoopPointwise_Double 2023-03-31T07:14:43.9172464Z [ OK ] TensorIteratorTest.SerialLoopPointwise_Double (1 ms) 2023-03-31T07:14:43.9173189Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Byte 2023-03-31T07:14:43.9173920Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Byte (0 ms) 2023-03-31T07:14:43.9174612Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Char 2023-03-31T07:14:43.9175312Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Char (0 ms) 2023-03-31T07:14:43.9176053Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Short 2023-03-31T07:14:43.9176820Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Short (0 ms) 2023-03-31T07:14:43.9177557Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Int 2023-03-31T07:14:43.9178281Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Int (0 ms) 2023-03-31T07:14:43.9178993Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Long 2023-03-31T07:14:43.9179802Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Long (0 ms) 2023-03-31T07:14:43.9180373Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Float 2023-03-31T07:14:43.9180927Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Float (0 ms) 2023-03-31T07:14:43.9181370Z [ RUN ] TensorIteratorTest.SerialLoopUnaryNoOutput_Double 2023-03-31T07:14:43.9181799Z [ OK ] TensorIteratorTest.SerialLoopUnaryNoOutput_Double (0 ms) 2023-03-31T07:14:43.9182243Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Byte 2023-03-31T07:14:43.9183079Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Byte (0 ms) 2023-03-31T07:14:43.9183970Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Char 2023-03-31T07:14:43.9184469Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Char (0 ms) 2023-03-31T07:14:43.9184909Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Short 2023-03-31T07:14:43.9185411Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Short (0 ms) 2023-03-31T07:14:43.9185941Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Int 2023-03-31T07:14:43.9186368Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Int (0 ms) 2023-03-31T07:14:43.9186807Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Long 2023-03-31T07:14:43.9187248Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Long (0 ms) 2023-03-31T07:14:43.9187677Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Float 2023-03-31T07:14:43.9188125Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Float (0 ms) 2023-03-31T07:14:43.9188567Z [ RUN ] TensorIteratorTest.SerialLoopBinaryNoOutput_Double 2023-03-31T07:14:43.9189018Z [ OK ] TensorIteratorTest.SerialLoopBinaryNoOutput_Double (0 ms) 2023-03-31T07:14:43.9189451Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Byte 2023-03-31T07:14:43.9189909Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Byte (0 ms) 2023-03-31T07:14:43.9190359Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Char 2023-03-31T07:14:43.9190879Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Char (0 ms) 2023-03-31T07:14:43.9191338Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Short 2023-03-31T07:14:43.9191858Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Short (0 ms) 2023-03-31T07:14:43.9192316Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Int 2023-03-31T07:14:43.9192758Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Int (0 ms) 2023-03-31T07:14:43.9193210Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Long 2023-03-31T07:14:43.9193671Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Long (0 ms) 2023-03-31T07:14:43.9194107Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Float 2023-03-31T07:14:43.9194565Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Float (0 ms) 2023-03-31T07:14:43.9195021Z [ RUN ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Double 2023-03-31T07:14:43.9195493Z [ OK ] TensorIteratorTest.SerialLoopPoinwiseNoOutput_Double (0 ms) 2023-03-31T07:14:43.9195918Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Byte 2023-03-31T07:14:43.9196338Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Byte (0 ms) 2023-03-31T07:14:43.9196755Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Char 2023-03-31T07:14:43.9197160Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Char (0 ms) 2023-03-31T07:14:43.9197581Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Short 2023-03-31T07:14:43.9198002Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Short (0 ms) 2023-03-31T07:14:43.9198400Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Int 2023-03-31T07:14:43.9198815Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Int (0 ms) 2023-03-31T07:14:43.9199239Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Long 2023-03-31T07:14:43.9199658Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Long (0 ms) 2023-03-31T07:14:43.9200061Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Float 2023-03-31T07:14:43.9200484Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Float (0 ms) 2023-03-31T07:14:43.9200897Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Double 2023-03-31T07:14:43.9201353Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Double (0 ms) 2023-03-31T07:14:43.9201802Z [ RUN ] TensorIteratorTest.ComparisonLoopBinary_Bool 2023-03-31T07:14:43.9202218Z [ OK ] TensorIteratorTest.ComparisonLoopBinary_Bool (0 ms) 2023-03-31T07:14:43.9202626Z [ RUN ] TensorIteratorTest.SerialLoopSingleThread 2023-03-31T07:14:43.9221503Z [ OK ] TensorIteratorTest.SerialLoopSingleThread (3 ms) 2023-03-31T07:14:43.9222186Z [ RUN ] TensorIteratorTest.InputDType 2023-03-31T07:14:43.9223365Z [ OK ] TensorIteratorTest.InputDType (0 ms) 2023-03-31T07:14:43.9224122Z [ RUN ] TensorIteratorTest.ComputeCommonDTypeInputOnly 2023-03-31T07:14:43.9224955Z [ OK ] TensorIteratorTest.ComputeCommonDTypeInputOnly (0 ms) 2023-03-31T07:14:43.9225813Z [ RUN ] TensorIteratorTest.DoNotComputeCommonDTypeInputOnly 2023-03-31T07:14:43.9226718Z [ OK ] TensorIteratorTest.DoNotComputeCommonDTypeInputOnly (0 ms) 2023-03-31T07:14:43.9227531Z [ RUN ] TensorIteratorTest.FailNonPromotingBinaryOp 2023-03-31T07:14:43.9234791Z [ OK ] TensorIteratorTest.FailNonPromotingBinaryOp (1 ms) 2023-03-31T07:14:43.9235759Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Byte 2023-03-31T07:14:43.9242210Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Byte (0 ms) 2023-03-31T07:14:43.9242896Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Char 2023-03-31T07:14:43.9244009Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Char (0 ms) 2023-03-31T07:14:43.9244583Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Short 2023-03-31T07:14:43.9245288Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Short (0 ms) 2023-03-31T07:14:43.9245911Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Int 2023-03-31T07:14:43.9246597Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Int (0 ms) 2023-03-31T07:14:43.9247238Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Long 2023-03-31T07:14:43.9247879Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Long (0 ms) 2023-03-31T07:14:43.9248570Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Float 2023-03-31T07:14:43.9249282Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Float (0 ms) 2023-03-31T07:14:43.9250015Z [ RUN ] TensorIteratorTest.CpuKernelMultipleOutputs_Double 2023-03-31T07:14:43.9250756Z [ OK ] TensorIteratorTest.CpuKernelMultipleOutputs_Double (0 ms) 2023-03-31T07:14:43.9251411Z [----------] 65 tests from TensorIteratorTest (66 ms total) 2023-03-31T07:14:43.9251703Z 2023-03-31T07:14:43.9252015Z [----------] Global test environment tear-down 2023-03-31T07:14:43.9252545Z [==========] 65 tests from 1 test suite ran. (66 ms total) 2023-03-31T07:14:43.9252996Z [ PASSED ] 65 tests. 2023-03-31T07:14:43.9940797Z + ./Dimname_test 2023-03-31T07:14:44.2562725Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:44.2563423Z [==========] Running 4 tests from 1 test suite. 2023-03-31T07:14:44.2563751Z [----------] Global test environment set-up. 2023-03-31T07:14:44.2564058Z [----------] 4 tests from DimnameTest 2023-03-31T07:14:44.2564375Z [ RUN ] DimnameTest.isValidIdentifier 2023-03-31T07:14:44.2564742Z [ OK ] DimnameTest.isValidIdentifier (0 ms) 2023-03-31T07:14:44.2565066Z [ RUN ] DimnameTest.wildcardName 2023-03-31T07:14:44.2565367Z [ OK ] DimnameTest.wildcardName (0 ms) 2023-03-31T07:14:44.2565697Z [ RUN ] DimnameTest.createNormalName 2023-03-31T07:14:44.2574177Z [ OK ] DimnameTest.createNormalName (1 ms) 2023-03-31T07:14:44.2575046Z [ RUN ] DimnameTest.unifyAndMatch 2023-03-31T07:14:44.2575449Z [ OK ] DimnameTest.unifyAndMatch (0 ms) 2023-03-31T07:14:44.2575790Z [----------] 4 tests from DimnameTest (1 ms total) 2023-03-31T07:14:44.2575951Z 2023-03-31T07:14:44.2576128Z [----------] Global test environment tear-down 2023-03-31T07:14:44.2576443Z [==========] 4 tests from 1 test suite ran. (1 ms total) 2023-03-31T07:14:44.2576714Z [ PASSED ] 4 tests. 2023-03-31T07:14:44.3258202Z + ./Dict_test 2023-03-31T07:14:44.5903673Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:44.5904504Z [==========] Running 47 tests from 2 test suites. 2023-03-31T07:14:44.5905067Z [----------] Global test environment set-up. 2023-03-31T07:14:44.5905599Z [----------] 46 tests from DictTest 2023-03-31T07:14:44.5906229Z [ RUN ] DictTest.givenEmptyDict_whenCallingEmpty_thenReturnsTrue 2023-03-31T07:14:44.5906968Z [ OK ] DictTest.givenEmptyDict_whenCallingEmpty_thenReturnsTrue (0 ms) 2023-03-31T07:14:44.5907627Z [ RUN ] DictTest.givenNonemptyDict_whenCallingEmpty_thenReturnsFalse 2023-03-31T07:14:44.5908317Z [ OK ] DictTest.givenNonemptyDict_whenCallingEmpty_thenReturnsFalse (0 ms) 2023-03-31T07:14:44.5909025Z [ RUN ] DictTest.givenEmptyDict_whenCallingSize_thenReturnsZero 2023-03-31T07:14:44.5909671Z [ OK ] DictTest.givenEmptyDict_whenCallingSize_thenReturnsZero (0 ms) 2023-03-31T07:14:44.5910661Z [ RUN ] DictTest.givenNonemptyDict_whenCallingSize_thenReturnsNumberOfElements 2023-03-31T07:14:44.5911427Z [ OK ] DictTest.givenNonemptyDict_whenCallingSize_thenReturnsNumberOfElements (0 ms) 2023-03-31T07:14:44.5912228Z [ RUN ] DictTest.givenNonemptyDict_whenCallingClear_thenIsEmpty 2023-03-31T07:14:44.5912857Z [ OK ] DictTest.givenNonemptyDict_whenCallingClear_thenIsEmpty (0 ms) 2023-03-31T07:14:44.5913454Z [ RUN ] DictTest.whenInsertingNewKey_thenReturnsTrueAndIteratorToNewElement 2023-03-31T07:14:44.5914076Z [ OK ] DictTest.whenInsertingNewKey_thenReturnsTrueAndIteratorToNewElement (0 ms) 2023-03-31T07:14:44.5914762Z [ RUN ] DictTest.whenInsertingExistingKey_thenReturnsFalseAndIteratorToExistingElement 2023-03-31T07:14:44.5915502Z [ OK ] DictTest.whenInsertingExistingKey_thenReturnsFalseAndIteratorToExistingElement (0 ms) 2023-03-31T07:14:44.5916155Z [ RUN ] DictTest.whenInsertingExistingKey_thenDoesNotModifyDict 2023-03-31T07:14:44.5916790Z [ OK ] DictTest.whenInsertingExistingKey_thenDoesNotModifyDict (0 ms) 2023-03-31T07:14:44.5917590Z [ RUN ] DictTest.whenInsertOrAssigningNewKey_thenReturnsTrueAndIteratorToNewElement 2023-03-31T07:14:44.5918473Z [ OK ] DictTest.whenInsertOrAssigningNewKey_thenReturnsTrueAndIteratorToNewElement (0 ms) 2023-03-31T07:14:44.5919420Z [ RUN ] DictTest.whenInsertOrAssigningExistingKey_thenReturnsFalseAndIteratorToChangedElement 2023-03-31T07:14:44.5920414Z [ OK ] DictTest.whenInsertOrAssigningExistingKey_thenReturnsFalseAndIteratorToChangedElement (0 ms) 2023-03-31T07:14:44.5921303Z [ RUN ] DictTest.whenInsertOrAssigningExistingKey_thenDoesModifyDict 2023-03-31T07:14:44.5922057Z [ OK ] DictTest.whenInsertOrAssigningExistingKey_thenDoesModifyDict (0 ms) 2023-03-31T07:14:44.5922782Z [ RUN ] DictTest.givenEmptyDict_whenIterating_thenBeginIsEnd 2023-03-31T07:14:44.5923700Z [ OK ] DictTest.givenEmptyDict_whenIterating_thenBeginIsEnd (0 ms) 2023-03-31T07:14:44.5924349Z [ RUN ] DictTest.givenMutableDict_whenIterating_thenFindsElements 2023-03-31T07:14:44.5925046Z [ OK ] DictTest.givenMutableDict_whenIterating_thenFindsElements (0 ms) 2023-03-31T07:14:44.5940333Z [ RUN ] DictTest.givenMutableDict_whenIteratingWithForeach_thenFindsElements 2023-03-31T07:14:44.5941252Z [ OK ] DictTest.givenMutableDict_whenIteratingWithForeach_thenFindsElements (0 ms) 2023-03-31T07:14:44.5941841Z [ RUN ] DictTest.givenConstDict_whenIterating_thenFindsElements 2023-03-31T07:14:44.5942397Z [ OK ] DictTest.givenConstDict_whenIterating_thenFindsElements (0 ms) 2023-03-31T07:14:44.5943083Z [ RUN ] DictTest.givenConstDict_whenIteratingWithForeach_thenFindsElements 2023-03-31T07:14:44.5943720Z [ OK ] DictTest.givenConstDict_whenIteratingWithForeach_thenFindsElements (0 ms) 2023-03-31T07:14:44.5944279Z [ RUN ] DictTest.givenIterator_thenCanModifyValue 2023-03-31T07:14:44.5944785Z [ OK ] DictTest.givenIterator_thenCanModifyValue (0 ms) 2023-03-31T07:14:44.5945351Z [ RUN ] DictTest.givenOneElementDict_whenErasingByIterator_thenDictIsEmpty 2023-03-31T07:14:44.5945973Z [ OK ] DictTest.givenOneElementDict_whenErasingByIterator_thenDictIsEmpty (0 ms) 2023-03-31T07:14:44.5946606Z [ RUN ] DictTest.givenOneElementDict_whenErasingByKey_thenReturnsOneAndDictIsEmpty 2023-03-31T07:14:44.5947251Z [ OK ] DictTest.givenOneElementDict_whenErasingByKey_thenReturnsOneAndDictIsEmpty (0 ms) 2023-03-31T07:14:44.5947965Z [ RUN ] DictTest.givenOneElementDict_whenErasingByNonexistingKey_thenReturnsZeroAndDictIsUnchanged 2023-03-31T07:14:44.5948724Z [ OK ] DictTest.givenOneElementDict_whenErasingByNonexistingKey_thenReturnsZeroAndDictIsUnchanged (0 ms) 2023-03-31T07:14:44.5949553Z [ RUN ] DictTest.whenCallingAtWithExistingKey_thenReturnsCorrectElement 2023-03-31T07:14:44.5950178Z [ OK ] DictTest.whenCallingAtWithExistingKey_thenReturnsCorrectElement (0 ms) 2023-03-31T07:14:44.5950902Z [ RUN ] DictTest.whenCallingAtWithNonExistingKey_thenReturnsCorrectElement 2023-03-31T07:14:44.5951577Z [ OK ] DictTest.whenCallingAtWithNonExistingKey_thenReturnsCorrectElement (0 ms) 2023-03-31T07:14:44.5952252Z [ RUN ] DictTest.givenMutableDict_whenCallingFindOnExistingKey_thenFindsCorrectElement 2023-03-31T07:14:44.5952923Z [ OK ] DictTest.givenMutableDict_whenCallingFindOnExistingKey_thenFindsCorrectElement (0 ms) 2023-03-31T07:14:44.5953575Z [ RUN ] DictTest.givenMutableDict_whenCallingFindOnNonExistingKey_thenReturnsEnd 2023-03-31T07:14:44.5954222Z [ OK ] DictTest.givenMutableDict_whenCallingFindOnNonExistingKey_thenReturnsEnd (0 ms) 2023-03-31T07:14:44.5954942Z [ RUN ] DictTest.givenConstDict_whenCallingFindOnExistingKey_thenFindsCorrectElement 2023-03-31T07:14:44.5955595Z [ OK ] DictTest.givenConstDict_whenCallingFindOnExistingKey_thenFindsCorrectElement (0 ms) 2023-03-31T07:14:44.5956236Z [ RUN ] DictTest.givenConstDict_whenCallingFindOnNonExistingKey_thenReturnsEnd 2023-03-31T07:14:44.5956945Z [ OK ] DictTest.givenConstDict_whenCallingFindOnNonExistingKey_thenReturnsEnd (0 ms) 2023-03-31T07:14:44.5957641Z [ RUN ] DictTest.whenCallingContainsWithExistingKey_thenReturnsTrue 2023-03-31T07:14:44.5958361Z [ OK ] DictTest.whenCallingContainsWithExistingKey_thenReturnsTrue (0 ms) 2023-03-31T07:14:44.5959117Z [ RUN ] DictTest.whenCallingContainsWithNonExistingKey_thenReturnsFalse 2023-03-31T07:14:44.5959949Z [ OK ] DictTest.whenCallingContainsWithNonExistingKey_thenReturnsFalse (0 ms) 2023-03-31T07:14:44.5960674Z [ RUN ] DictTest.whenCallingReserve_thenDoesntCrash 2023-03-31T07:14:44.5961241Z [ OK ] DictTest.whenCallingReserve_thenDoesntCrash (0 ms) 2023-03-31T07:14:44.5961820Z [ RUN ] DictTest.whenCopyConstructingDict_thenAreEqual 2023-03-31T07:14:44.5962498Z [ OK ] DictTest.whenCopyConstructingDict_thenAreEqual (0 ms) 2023-03-31T07:14:44.5963443Z [ RUN ] DictTest.whenCopyAssigningDict_thenAreEqual 2023-03-31T07:14:44.5964234Z [ OK ] DictTest.whenCopyAssigningDict_thenAreEqual (0 ms) 2023-03-31T07:14:44.5964769Z [ RUN ] DictTest.whenCopyingDict_thenAreEqual 2023-03-31T07:14:44.5965265Z [ OK ] DictTest.whenCopyingDict_thenAreEqual (0 ms) 2023-03-31T07:14:44.5965817Z [ RUN ] DictTest.whenMoveConstructingDict_thenNewIsCorrect 2023-03-31T07:14:44.5966428Z [ OK ] DictTest.whenMoveConstructingDict_thenNewIsCorrect (0 ms) 2023-03-31T07:14:44.5967022Z [ RUN ] DictTest.whenMoveAssigningDict_thenNewIsCorrect 2023-03-31T07:14:44.5967589Z [ OK ] DictTest.whenMoveAssigningDict_thenNewIsCorrect (0 ms) 2023-03-31T07:14:44.5968196Z [ RUN ] DictTest.whenMoveConstructingDict_thenOldIsUnchanged 2023-03-31T07:14:44.5968815Z [ OK ] DictTest.whenMoveConstructingDict_thenOldIsUnchanged (0 ms) 2023-03-31T07:14:44.5969400Z [ RUN ] DictTest.whenMoveAssigningDict_thenOldIsUnchanged 2023-03-31T07:14:44.5969998Z [ OK ] DictTest.whenMoveAssigningDict_thenOldIsUnchanged (0 ms) 2023-03-31T07:14:44.5970714Z [ RUN ] DictTest.givenIterator_whenPostfixIncrementing_thenMovesToNextAndReturnsOldPosition 2023-03-31T07:14:44.5971519Z [ OK ] DictTest.givenIterator_whenPostfixIncrementing_thenMovesToNextAndReturnsOldPosition (0 ms) 2023-03-31T07:14:44.5972310Z [ RUN ] DictTest.givenIterator_whenPrefixIncrementing_thenMovesToNextAndReturnsNewPosition 2023-03-31T07:14:44.5973235Z [ OK ] DictTest.givenIterator_whenPrefixIncrementing_thenMovesToNextAndReturnsNewPosition (0 ms) 2023-03-31T07:14:44.5973884Z [ RUN ] DictTest.givenEqualIterators_thenAreEqual 2023-03-31T07:14:44.5974499Z [ OK ] DictTest.givenEqualIterators_thenAreEqual (0 ms) 2023-03-31T07:14:44.5975071Z [ RUN ] DictTest.givenDifferentIterators_thenAreNotEqual 2023-03-31T07:14:44.5975668Z [ OK ] DictTest.givenDifferentIterators_thenAreNotEqual (0 ms) 2023-03-31T07:14:44.5976316Z [ RUN ] DictTest.givenIterator_whenDereferencing_thenPointsToCorrectElement 2023-03-31T07:14:44.5976989Z [ OK ] DictTest.givenIterator_whenDereferencing_thenPointsToCorrectElement (0 ms) 2023-03-31T07:14:44.5977639Z [ RUN ] DictTest.givenIterator_whenWritingToValue_thenChangesValue 2023-03-31T07:14:44.5978256Z [ OK ] DictTest.givenIterator_whenWritingToValue_thenChangesValue (0 ms) 2023-03-31T07:14:44.5978756Z [ RUN ] DictTest.isReferenceType 2023-03-31T07:14:44.5979201Z [ OK ] DictTest.isReferenceType (0 ms) 2023-03-31T07:14:44.5979677Z [ RUN ] DictTest.copyHasSeparateStorage 2023-03-31T07:14:44.5980198Z [ OK ] DictTest.copyHasSeparateStorage (0 ms) 2023-03-31T07:14:44.5980653Z [ RUN ] DictTest.dictTensorAsKey 2023-03-31T07:14:44.5981115Z [ OK ] DictTest.dictTensorAsKey (0 ms) 2023-03-31T07:14:44.5981541Z [ RUN ] DictTest.dictEquality 2023-03-31T07:14:44.5981954Z [ OK ] DictTest.dictEquality (0 ms) 2023-03-31T07:14:44.5982419Z [----------] 46 tests from DictTest (0 ms total) 2023-03-31T07:14:44.5982741Z 2023-03-31T07:14:44.5983000Z [----------] 1 test from ListTest_IValueBasedList 2023-03-31T07:14:44.5983648Z [ RUN ] ListTest_IValueBasedList.givenIterator_whenWritingToValueFromIterator_thenChangesValue 2023-03-31T07:14:44.5984478Z [ OK ] ListTest_IValueBasedList.givenIterator_whenWritingToValueFromIterator_thenChangesValue (0 ms) 2023-03-31T07:14:44.5985148Z [----------] 1 test from ListTest_IValueBasedList (0 ms total) 2023-03-31T07:14:44.5985407Z 2023-03-31T07:14:44.5985674Z [----------] Global test environment tear-down 2023-03-31T07:14:44.5986147Z [==========] 47 tests from 2 test suites ran. (0 ms total) 2023-03-31T07:14:44.5986643Z [ PASSED ] 47 tests. 2023-03-31T07:14:44.6663776Z + ./NamedTensor_test 2023-03-31T07:14:44.9298712Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:44.9299506Z [==========] Running 10 tests from 1 test suite. 2023-03-31T07:14:44.9300014Z [----------] Global test environment set-up. 2023-03-31T07:14:44.9300541Z [----------] 10 tests from NamedTensorTest 2023-03-31T07:14:44.9301033Z [ RUN ] NamedTensorTest.isNamed 2023-03-31T07:14:44.9301890Z [W TensorImpl.h:1791] Warning: Named tensors and all their associated APIs are an experimental feature and subject to change. Please do not use them for anything important until they are released as stable. (function operator()) 2023-03-31T07:14:44.9303008Z [ OK ] NamedTensorTest.isNamed (0 ms) 2023-03-31T07:14:44.9303584Z [ RUN ] NamedTensorTest.attachMetadata 2023-03-31T07:14:44.9304193Z [ OK ] NamedTensorTest.attachMetadata (0 ms) 2023-03-31T07:14:44.9304864Z [ RUN ] NamedTensorTest.internalSetNamesInplace 2023-03-31T07:14:44.9305575Z [ OK ] NamedTensorTest.internalSetNamesInplace (0 ms) 2023-03-31T07:14:44.9306172Z [ RUN ] NamedTensorTest.empty 2023-03-31T07:14:44.9339442Z [ OK ] NamedTensorTest.empty (3 ms) 2023-03-31T07:14:44.9340106Z [ RUN ] NamedTensorTest.dimnameToPosition 2023-03-31T07:14:44.9345156Z [ OK ] NamedTensorTest.dimnameToPosition (0 ms) 2023-03-31T07:14:44.9346143Z [ RUN ] NamedTensorTest.unifyFromRight 2023-03-31T07:14:44.9402873Z [ OK ] NamedTensorTest.unifyFromRight (5 ms) 2023-03-31T07:14:44.9403736Z [ RUN ] NamedTensorTest.alias 2023-03-31T07:14:44.9404586Z [ OK ] NamedTensorTest.alias (0 ms) 2023-03-31T07:14:44.9405169Z [ RUN ] NamedTensorTest.NoNamesGuard 2023-03-31T07:14:44.9405786Z [ OK ] NamedTensorTest.NoNamesGuard (0 ms) 2023-03-31T07:14:44.9406397Z [ RUN ] NamedTensorTest.TensorNamePrint 2023-03-31T07:14:44.9407011Z [ OK ] NamedTensorTest.TensorNamePrint (0 ms) 2023-03-31T07:14:44.9407638Z [ RUN ] NamedTensorTest.TensorNamesCheckUnique 2023-03-31T07:14:44.9409473Z [ OK ] NamedTensorTest.TensorNamesCheckUnique (0 ms) 2023-03-31T07:14:44.9410022Z [----------] 10 tests from NamedTensorTest (11 ms total) 2023-03-31T07:14:44.9410280Z 2023-03-31T07:14:44.9410561Z [----------] Global test environment tear-down 2023-03-31T07:14:44.9411087Z [==========] 10 tests from 1 test suite ran. (11 ms total) 2023-03-31T07:14:44.9411502Z [ PASSED ] 10 tests. 2023-03-31T07:14:45.0107320Z + ./cpu_generator_test 2023-03-31T07:14:45.2707569Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:45.2708131Z [==========] Running 16 tests from 1 test suite. 2023-03-31T07:14:45.2708459Z [----------] Global test environment set-up. 2023-03-31T07:14:45.2708888Z [----------] 16 tests from CPUGeneratorImpl 2023-03-31T07:14:45.2709339Z [ RUN ] CPUGeneratorImpl.TestGeneratorDynamicCast 2023-03-31T07:14:45.2709966Z [ OK ] CPUGeneratorImpl.TestGeneratorDynamicCast (0 ms) 2023-03-31T07:14:45.2710576Z [ RUN ] CPUGeneratorImpl.TestDefaultGenerator 2023-03-31T07:14:45.2711188Z [ OK ] CPUGeneratorImpl.TestDefaultGenerator (0 ms) 2023-03-31T07:14:45.2711748Z [ RUN ] CPUGeneratorImpl.TestCloning 2023-03-31T07:14:45.2712251Z [ OK ] CPUGeneratorImpl.TestCloning (0 ms) 2023-03-31T07:14:45.2712985Z [ RUN ] CPUGeneratorImpl.TestMultithreadingGetEngineOperator 2023-03-31T07:14:45.2713504Z [ OK ] CPUGeneratorImpl.TestMultithreadingGetEngineOperator (0 ms) 2023-03-31T07:14:45.2714151Z [ RUN ] CPUGeneratorImpl.TestGetSetCurrentSeed 2023-03-31T07:14:45.2714600Z [ OK ] CPUGeneratorImpl.TestGetSetCurrentSeed (0 ms) 2023-03-31T07:14:45.2715041Z [ RUN ] CPUGeneratorImpl.TestMultithreadingGetSetCurrentSeed 2023-03-31T07:14:45.2715531Z [ OK ] CPUGeneratorImpl.TestMultithreadingGetSetCurrentSeed (0 ms) 2023-03-31T07:14:45.2715923Z [ RUN ] CPUGeneratorImpl.TestRNGForking 2023-03-31T07:14:45.2899607Z [ OK ] CPUGeneratorImpl.TestRNGForking (18 ms) 2023-03-31T07:14:45.2900388Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineReproducibility 2023-03-31T07:14:45.2901103Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineReproducibility (0 ms) 2023-03-31T07:14:45.2901756Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineOffset1 2023-03-31T07:14:45.2902546Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineOffset1 (0 ms) 2023-03-31T07:14:45.2903186Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineOffset2 2023-03-31T07:14:45.2903817Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineOffset2 (0 ms) 2023-03-31T07:14:45.2904444Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineOffset3 2023-03-31T07:14:45.2904987Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineOffset3 (0 ms) 2023-03-31T07:14:45.2905490Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineIndex 2023-03-31T07:14:45.2905977Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineIndex (0 ms) 2023-03-31T07:14:45.2906519Z [ RUN ] CPUGeneratorImpl.TestMT19937EngineReproducibility 2023-03-31T07:14:45.2907375Z [ OK ] CPUGeneratorImpl.TestMT19937EngineReproducibility (0 ms) 2023-03-31T07:14:45.2907974Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineReproducibilityRandN 2023-03-31T07:14:45.2908735Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineReproducibilityRandN (0 ms) 2023-03-31T07:14:45.2909414Z [ RUN ] CPUGeneratorImpl.TestPhiloxEngineSeedRandN 2023-03-31T07:14:45.2910074Z [ OK ] CPUGeneratorImpl.TestPhiloxEngineSeedRandN (0 ms) 2023-03-31T07:14:45.2910658Z [ RUN ] CPUGeneratorImpl.TestPhiloxDeterministic 2023-03-31T07:14:45.2911305Z [ OK ] CPUGeneratorImpl.TestPhiloxDeterministic (0 ms) 2023-03-31T07:14:45.2911942Z [----------] 16 tests from CPUGeneratorImpl (19 ms total) 2023-03-31T07:14:45.2912222Z 2023-03-31T07:14:45.2912500Z [----------] Global test environment tear-down 2023-03-31T07:14:45.2913037Z [==========] 16 tests from 1 test suite ran. (19 ms total) 2023-03-31T07:14:45.2913479Z [ PASSED ] 16 tests. 2023-03-31T07:14:45.3597803Z + ./legacy_vmap_test 2023-03-31T07:14:45.6214916Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:45.6215539Z [==========] Running 23 tests from 1 test suite. 2023-03-31T07:14:45.6215857Z [----------] Global test environment set-up. 2023-03-31T07:14:45.6216163Z [----------] 23 tests from VmapTest 2023-03-31T07:14:45.6216467Z [ RUN ] VmapTest.TestBatchedTensor 2023-03-31T07:14:45.6241698Z [ OK ] VmapTest.TestBatchedTensor (2 ms) 2023-03-31T07:14:45.6242067Z [ RUN ] VmapTest.TestBatchedTensorMaxLevel 2023-03-31T07:14:45.6257026Z [ OK ] VmapTest.TestBatchedTensorMaxLevel (1 ms) 2023-03-31T07:14:45.6257408Z [ RUN ] VmapTest.TestBatchedTensorActualDim 2023-03-31T07:14:45.6283810Z [ OK ] VmapTest.TestBatchedTensorActualDim (2 ms) 2023-03-31T07:14:45.6284263Z [ RUN ] VmapTest.TestMultiBatchVmapTransform 2023-03-31T07:14:45.6293684Z [ OK ] VmapTest.TestMultiBatchVmapTransform (1 ms) 2023-03-31T07:14:45.6294260Z [ RUN ] VmapTest.TestVmapPhysicalViewGetPhysicalDim 2023-03-31T07:14:45.6303666Z [ OK ] VmapTest.TestVmapPhysicalViewGetPhysicalDim (0 ms) 2023-03-31T07:14:45.6304529Z [ RUN ] VmapTest.TestVmapPhysicalViewGetPhysicalDims 2023-03-31T07:14:45.6313625Z [ OK ] VmapTest.TestVmapPhysicalViewGetPhysicalDims (0 ms) 2023-03-31T07:14:45.6314266Z [ RUN ] VmapTest.TestVmapPhysicalViewNewLogicalFromPhysical 2023-03-31T07:14:45.6314765Z [ OK ] VmapTest.TestVmapPhysicalViewNewLogicalFromPhysical (0 ms) 2023-03-31T07:14:45.6315174Z [ RUN ] VmapTest.TestBatchedTensorSum 2023-03-31T07:14:45.6317179Z [ OK ] VmapTest.TestBatchedTensorSum (0 ms) 2023-03-31T07:14:45.6317596Z [ RUN ] VmapTest.TestBroadcastingVmapTransformBatchedBatched 2023-03-31T07:14:45.6323200Z [ OK ] VmapTest.TestBroadcastingVmapTransformBatchedBatched (0 ms) 2023-03-31T07:14:45.6323919Z [ RUN ] VmapTest.TestBroadcastingVmapTransformBatchedUnbatched 2023-03-31T07:14:45.6328261Z [ OK ] VmapTest.TestBroadcastingVmapTransformBatchedUnbatched (0 ms) 2023-03-31T07:14:45.6328903Z [ RUN ] VmapTest.TestBroadcastingVmapTransformMaxLevels 2023-03-31T07:14:45.6332626Z [ OK ] VmapTest.TestBroadcastingVmapTransformMaxLevels (0 ms) 2023-03-31T07:14:45.6333178Z [ RUN ] VmapTest.TestBatchedTensorMul 2023-03-31T07:14:45.6334323Z [ OK ] VmapTest.TestBatchedTensorMul (0 ms) 2023-03-31T07:14:45.6334675Z [ RUN ] VmapTest.TestBatchedTensorSize 2023-03-31T07:14:45.6340301Z [ OK ] VmapTest.TestBatchedTensorSize (0 ms) 2023-03-31T07:14:45.6340787Z [ RUN ] VmapTest.TestVmapPhysicalViewGetPhysicalShape 2023-03-31T07:14:45.6341403Z [ OK ] VmapTest.TestVmapPhysicalViewGetPhysicalShape (0 ms) 2023-03-31T07:14:45.6341806Z [ RUN ] VmapTest.TestBatchedTensorExpand 2023-03-31T07:14:45.6478123Z [ OK ] VmapTest.TestBatchedTensorExpand (13 ms) 2023-03-31T07:14:45.6479097Z [ RUN ] VmapTest.TestBatchedTensorUnsqueeze 2023-03-31T07:14:45.6479791Z [ OK ] VmapTest.TestBatchedTensorUnsqueeze (0 ms) 2023-03-31T07:14:45.6480429Z [ RUN ] VmapTest.TestBatchedTensorSqueeze 2023-03-31T07:14:45.6481088Z [ OK ] VmapTest.TestBatchedTensorSqueeze (0 ms) 2023-03-31T07:14:45.6481704Z [ RUN ] VmapTest.TestBatchedTensorTranspose 2023-03-31T07:14:45.6482606Z [ OK ] VmapTest.TestBatchedTensorTranspose (0 ms) 2023-03-31T07:14:45.6483183Z [ RUN ] VmapTest.TestBatchedTensorPermute 2023-03-31T07:14:45.6484965Z [ OK ] VmapTest.TestBatchedTensorPermute (0 ms) 2023-03-31T07:14:45.6485414Z [ RUN ] VmapTest.TestMultiBatchVmapTransformBatchedBatched 2023-03-31T07:14:45.6496946Z [ OK ] VmapTest.TestMultiBatchVmapTransformBatchedBatched (1 ms) 2023-03-31T07:14:45.6497462Z [ RUN ] VmapTest.TestMultiBatchVmapTransformBatchedUnbatched 2023-03-31T07:14:45.6501511Z [ OK ] VmapTest.TestMultiBatchVmapTransformBatchedUnbatched (0 ms) 2023-03-31T07:14:45.6502000Z [ RUN ] VmapTest.TestMultiBatchVmapTransformMaxLevels 2023-03-31T07:14:45.6506190Z [ OK ] VmapTest.TestMultiBatchVmapTransformMaxLevels (0 ms) 2023-03-31T07:14:45.6506739Z [ RUN ] VmapTest.TestMultiBatchVmapTransformMultipleTensors 2023-03-31T07:14:45.6509776Z [ OK ] VmapTest.TestMultiBatchVmapTransformMultipleTensors (0 ms) 2023-03-31T07:14:45.6510307Z [----------] 23 tests from VmapTest (29 ms total) 2023-03-31T07:14:45.6510478Z 2023-03-31T07:14:45.6510645Z [----------] Global test environment tear-down 2023-03-31T07:14:45.6510987Z [==========] 23 tests from 1 test suite ran. (29 ms total) 2023-03-31T07:14:45.6511265Z [ PASSED ] 23 tests. 2023-03-31T07:14:45.7249946Z + ./operators_test 2023-03-31T07:14:45.9854938Z Running main() from /var/lib/jenkins/workspace/third_party/googletest/googletest/src/gtest_main.cc 2023-03-31T07:14:45.9855483Z [==========] Running 4 tests from 1 test suite. 2023-03-31T07:14:45.9856019Z [----------] Global test environment set-up. 2023-03-31T07:14:45.9856391Z [----------] 4 tests from OperatorsTest 2023-03-31T07:14:45.9856728Z [ RUN ] OperatorsTest.TestFunctionDecltype 2023-03-31T07:14:45.9863855Z [ OK ] OperatorsTest.TestFunctionDecltype (0 ms) 2023-03-31T07:14:45.9864349Z [ RUN ] OperatorsTest.TestMethodOnlyDecltype 2023-03-31T07:14:45.9864884Z [ OK ] OperatorsTest.TestMethodOnlyDecltype (0 ms) 2023-03-31T07:14:45.9865244Z [ RUN ] OperatorsTest.Test_ATEN_FN 2023-03-31T07:14:45.9887275Z [ OK ] OperatorsTest.Test_ATEN_FN (2 ms) 2023-03-31T07:14:45.9887921Z [ RUN ] OperatorsTest.TestOutVariantIsFaithful 2023-03-31T07:14:45.9888556Z [ OK ] OperatorsTest.TestOutVariantIsFaithful (0 ms) 2023-03-31T07:14:45.9888968Z [----------] 4 tests from OperatorsTest (3 ms total) 2023-03-31T07:14:45.9889137Z 2023-03-31T07:14:45.9889320Z [----------] Global test environment tear-down 2023-03-31T07:14:45.9889651Z [==========] 4 tests from 1 test suite ran. (3 ms total) 2023-03-31T07:14:45.9889927Z [ PASSED ] 4 tests. 2023-03-31T07:14:46.0577285Z + [[ -x ./cudnn_test ]] 2023-03-31T07:14:46.0577777Z + [[ -x ./cuda_generator_test ]] 2023-03-31T07:14:46.0578191Z + [[ -x ./apply_test ]] 2023-03-31T07:14:46.0578573Z + [[ -x ./stream_test ]] 2023-03-31T07:14:46.0578854Z + [[ -x ./cuda_half_test ]] 2023-03-31T07:14:46.0579142Z + [[ -x ./cuda_vectorized_test ]] 2023-03-31T07:14:46.0579652Z + [[ -x ./cuda_distributions_test ]] 2023-03-31T07:14:46.0579906Z + [[ -x ./cuda_optional_test ]] 2023-03-31T07:14:46.0580147Z + [[ -x ./cuda_tensor_interop_test ]] 2023-03-31T07:14:46.0580397Z + [[ -x ./cuda_complex_test ]] 2023-03-31T07:14:46.0580653Z + [[ -x ./cuda_complex_math_test ]] 2023-03-31T07:14:46.0580936Z + [[ -x ./cuda_cub_test ]] 2023-03-31T07:14:46.0581182Z + [[ -x ./cuda_atomic_ops_test ]] 2023-03-31T07:14:46.0581407Z + '[' OFF == ON ']' 2023-03-31T07:14:46.0581581Z + popd 2023-03-31T07:14:46.0581751Z ~/workspace 2023-03-31T07:14:46.0587816Z + [[ -n '' ]] 2023-03-31T07:14:46.0588065Z + assert_git_not_dirty 2023-03-31T07:14:46.0588381Z + [[ linux-bionic-py3.8-clang9 != *rocm* ]] 2023-03-31T07:14:46.0588683Z + [[ linux-bionic-py3.8-clang9 != *xla* ]] 2023-03-31T07:14:46.0603932Z ++ git status --porcelain 2023-03-31T07:14:46.1735099Z + git_status= 2023-03-31T07:14:46.1735506Z + [[ -n '' ]] 2023-03-31T07:14:46.1959237Z ##[group]Run cat test/**/*.log || true 2023-03-31T07:14:46.1959493Z cat test/**/*.log || true 2023-03-31T07:14:46.2567631Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:46.2567871Z env: 2023-03-31T07:14:46.2568067Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:46.2568383Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:46.2568658Z ##[endgroup] 2023-03-31T07:14:46.2648038Z cat: test/**/*.log: No such file or directory 2023-03-31T07:14:46.2689973Z Prepare all required actions 2023-03-31T07:14:46.2711155Z ##[group]Run ./.github/actions/get-workflow-job-id 2023-03-31T07:14:46.2711378Z with: 2023-03-31T07:14:46.2711730Z github-token: *** 2023-03-31T07:14:46.2711891Z env: 2023-03-31T07:14:46.2712072Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:46.2712368Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:46.2712650Z ##[endgroup] 2023-03-31T07:14:46.2727197Z ##[group]Run set -eux 2023-03-31T07:14:46.2727410Z set -eux 2023-03-31T07:14:46.2727709Z GHA_WORKFLOW_JOB_ID=$(python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}") 2023-03-31T07:14:46.2728048Z echo "job-id=${GHA_WORKFLOW_JOB_ID}" >> "${GITHUB_OUTPUT}" 2023-03-31T07:14:46.2739194Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:46.2739425Z env: 2023-03-31T07:14:46.2739610Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:46.2739950Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:46.2740440Z GITHUB_TOKEN: *** 2023-03-31T07:14:46.2740626Z ##[endgroup] 2023-03-31T07:14:46.2769132Z ++ python3 .github/scripts/get_workflow_job_id.py 4571766038 i-0da0c45d38bcc5929 2023-03-31T07:14:48.8120075Z + GHA_WORKFLOW_JOB_ID=12416619924 2023-03-31T07:14:48.8120584Z + echo job-id=12416619924 2023-03-31T07:14:48.8212381Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2023-03-31T07:14:48.8212639Z kill "$MONITOR_SCRIPT_PID" 2023-03-31T07:14:48.8223730Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:48.8223960Z env: 2023-03-31T07:14:48.8224147Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:48.8224435Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:48.8224820Z MONITOR_SCRIPT_PID: 25582 2023-03-31T07:14:48.8225046Z ##[endgroup] 2023-03-31T07:14:48.8355595Z Prepare all required actions 2023-03-31T07:14:48.8356054Z Getting action download info 2023-03-31T07:14:49.0058598Z Download action repository 'actions/upload-artifact@v3' (SHA:0b7f8abb1508181956e8e162db84b466c27e18ce) 2023-03-31T07:14:49.1601512Z ##[group]Run ./.github/actions/upload-test-artifacts 2023-03-31T07:14:49.1601732Z with: 2023-03-31T07:14:49.1601964Z file-suffix: test-default-1-3-linux.2xlarge_12416619924 2023-03-31T07:14:49.1602190Z env: 2023-03-31T07:14:49.1602359Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:49.1602655Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:49.1603391Z ##[endgroup] 2023-03-31T07:14:49.1625828Z ##[group]Run # Remove any previous test jsons if they exist 2023-03-31T07:14:49.1626123Z # Remove any previous test jsons if they exist 2023-03-31T07:14:49.1626372Z rm -f test-jsons-*.zip 2023-03-31T07:14:49.1626673Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test -i '*.json' 2023-03-31T07:14:49.1637833Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:49.1638074Z env: 2023-03-31T07:14:49.1638246Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:49.1638543Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:49.1638884Z FILE_SUFFIX: test-default-1-3-linux.2xlarge_12416619924 2023-03-31T07:14:49.1639120Z ##[endgroup] 2023-03-31T07:14:49.2075308Z adding: test/allowlist_for_publicAPI.json (deflated 78%) 2023-03-31T07:14:49.2133708Z adding: test/benchmark_utils/callgrind_artifacts.json (deflated 92%) 2023-03-31T07:14:49.2149817Z adding: test/profiler/profiler_utils_mock_events.json (deflated 87%) 2023-03-31T07:14:49.2151630Z adding: test/.pytorch-slow-tests.json (deflated 78%) 2023-03-31T07:14:49.2157965Z adding: test/.pytorch-disabled-tests.json (deflated 86%) 2023-03-31T07:14:49.2199795Z ##[group]Run # Remove any previous test reports if they exist 2023-03-31T07:14:49.2200116Z # Remove any previous test reports if they exist 2023-03-31T07:14:49.2200371Z rm -f test-reports-*.zip 2023-03-31T07:14:49.2200657Z zip -r "test-reports-${FILE_SUFFIX}.zip" test -i '*.xml' -i '*.csv' 2023-03-31T07:14:49.2211891Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:49.2212119Z env: 2023-03-31T07:14:49.2212303Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:49.2212593Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:49.2212934Z FILE_SUFFIX: test-default-1-3-linux.2xlarge_12416619924 2023-03-31T07:14:49.2213170Z ##[endgroup] 2023-03-31T07:14:49.2331864Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-fca67e1907f77aa4.xml (deflated 93%) 2023-03-31T07:14:49.2337731Z adding: test/test-reports/python-pytest/inductor.test_cpp_wrapper/inductor.test_cpp_wrapper-28bf4fee1260856b.xml (deflated 82%) 2023-03-31T07:14:49.2343370Z adding: test/test-reports/python-pytest/nn.test_module_hooks/nn.test_module_hooks-cf9ef0ba863acb32.xml (deflated 85%) 2023-03-31T07:14:49.2348754Z adding: test/test-reports/python-pytest/functorch.test_functionalize/functorch.test_functionalize-93e9c96bd792f149.xml (deflated 28%) 2023-03-31T07:14:49.2353406Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7f2fa785972ff8f.xml (deflated 28%) 2023-03-31T07:14:49.2418510Z adding: test/test-reports/python-pytest/test_quantization/test_quantization-816f8ae0c160c36a.xml (deflated 92%) 2023-03-31T07:14:49.2485299Z adding: test/test-reports/python-pytest/test_jit/test_jit-47d330549a92143e.xml (deflated 90%) 2023-03-31T07:14:49.2534550Z adding: test/test-reports/python-pytest/test_foreach/test_foreach-b6d6f436a62737dc.xml (deflated 97%) 2023-03-31T07:14:49.2538441Z adding: test/test-reports/python-pytest/dynamo.test_minifier/dynamo.test_minifier-9a16a7cb30caca6b.xml (deflated 86%) 2023-03-31T07:14:49.2544842Z adding: test/test-reports/python-pytest/test_tensorboard/test_tensorboard-1df86ba04af3d664.xml (deflated 83%) 2023-03-31T07:14:49.3274288Z adding: test/test-reports/python-pytest/test_transformers/test_transformers-38a31af97a106b47.xml (deflated 99%) 2023-03-31T07:14:49.3280554Z adding: test/test-reports/python-pytest/test_stateless/test_stateless-bb1d952c6ff344c3.xml (deflated 87%) 2023-03-31T07:14:49.3294424Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-e05b502ec4c844d1.xml (deflated 86%) 2023-03-31T07:14:49.3312128Z adding: test/test-reports/python-pytest/nn.test_embedding/nn.test_embedding-970b421025a9958a.xml (deflated 94%) 2023-03-31T07:14:49.3315770Z adding: test/test-reports/python-pytest/inductor.test_standalone_compile/inductor.test_standalone_compile-12f270e795e46dc6.xml (deflated 80%) 2023-03-31T07:14:49.3321484Z adding: test/test-reports/python-pytest/test_datapipe/test_datapipe-a20c618c60f64e38.xml (deflated 90%) 2023-03-31T07:14:49.3331477Z adding: test/test-reports/python-pytest/test_functionalization/test_functionalization-c2940ed962564e0b.xml (deflated 88%) 2023-03-31T07:14:49.3335454Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-f170e6cb75f486bf.xml (deflated 80%) 2023-03-31T07:14:49.3340554Z adding: test/test-reports/python-pytest/test_weak/test_weak-1e7688e5b758402f.xml (deflated 83%) 2023-03-31T07:14:49.3351378Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-4148f53ee5bf8b62.xml (deflated 51%) 2023-03-31T07:14:49.3359972Z adding: test/test-reports/python-pytest/dynamo.test_export/dynamo.test_export-c5d005f8365c4b22.xml (deflated 91%) 2023-03-31T07:14:49.3365679Z adding: test/test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-3e275992b40418f7.xml (deflated 83%) 2023-03-31T07:14:49.3384364Z adding: test/test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-0ff7e4912c9e147c.xml (deflated 90%) 2023-03-31T07:14:49.3388406Z adding: test/test-reports/python-pytest/test_import_stats/test_import_stats-9f6b211aaa7d9826.xml (deflated 50%) 2023-03-31T07:14:49.3392056Z adding: test/test-reports/python-pytest/test_mkldnn_verbose/test_mkldnn_verbose-7df56168c4f02b69.xml (deflated 51%) 2023-03-31T07:14:49.3410079Z adding: test/test-reports/python-pytest/dynamo.test_modules/dynamo.test_modules-e361d6bdc2cfdd69.xml (deflated 90%) 2023-03-31T07:14:49.3414107Z adding: test/test-reports/python-pytest/functorch.test_control_flow/functorch.test_control_flow-d5e8f048f0857b56.xml (deflated 87%) 2023-03-31T07:14:49.3421921Z adding: test/test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-b95643fb9ab252db.xml (deflated 92%) 2023-03-31T07:14:49.3452074Z adding: test/test-reports/python-pytest/test_maskedtensor/test_maskedtensor-53d6bd0ec99cdfe0.xml (deflated 95%) 2023-03-31T07:14:49.3473500Z adding: test/test-reports/python-pytest/test_package/test_package-ee02c66f15df73e0.xml (deflated 87%) 2023-03-31T07:14:49.3523647Z adding: test/test-reports/python-pytest/functorch.test_vmap_registrations/functorch.test_vmap_registrations-1e483d8efa4950e7.xml (deflated 97%) 2023-03-31T07:14:49.3542486Z adding: test/test-reports/python-pytest/test_module_init/test_module_init-855433dfa1eb4a15.xml (deflated 95%) 2023-03-31T07:14:49.3558815Z adding: test/test-reports/python-pytest/dynamo.test_optimizers/dynamo.test_optimizers-2c37486a74067b12.xml (deflated 88%) 2023-03-31T07:14:49.3563216Z adding: test/test-reports/python-pytest/test_logging/test_logging-70826ade4cb4ade2.xml (deflated 37%) 2023-03-31T07:14:49.3571234Z adding: test/test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-009af74fcb58de5f.xml (deflated 92%) 2023-03-31T07:14:49.3574831Z adding: test/test-reports/python-pytest/test_public_bindings/test_public_bindings-46374e86c8090fb4.xml (deflated 46%) 2023-03-31T07:14:49.3579628Z adding: test/test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-e0d23d9b5160040a.xml (deflated 77%) 2023-03-31T07:14:49.3589732Z adding: test/test-reports/python-pytest/dynamo.test_functions/dynamo.test_functions-85421862794e1580.xml (deflated 93%) 2023-03-31T07:14:49.3596065Z adding: test/test-reports/python-pytest/test_python_dispatch/test_python_dispatch-0ca79c911b8030a4.xml (deflated 83%) 2023-03-31T07:14:49.3605372Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-9d35fd1a78d8eba1.xml (deflated 88%) 2023-03-31T07:14:49.3610044Z adding: test/test-reports/python-pytest/dynamo.test_backends/dynamo.test_backends-cd2960c4786ef00b.xml (deflated 83%) 2023-03-31T07:14:49.3615099Z adding: test/test-reports/python-pytest/test_bundled_inputs/test_bundled_inputs-fdaf441b142eb3fe.xml (deflated 73%) 2023-03-31T07:14:49.3619752Z adding: test/test-reports/python-pytest/test_futures/test_futures-3d74c9457cb7d0ac.xml (deflated 83%) 2023-03-31T07:14:49.3636801Z adding: test/test-reports/python-pytest/test_namedtensor/test_namedtensor-67adda638f28d19a.xml (deflated 87%) 2023-03-31T07:14:49.3643727Z adding: test/test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-c570424306e69389.xml (deflated 83%) 2023-03-31T07:14:49.3649616Z adding: test/test-reports/python-pytest/test_model_dump/test_model_dump-de886d8ecb5e46a4.xml (deflated 71%) 2023-03-31T07:14:49.3654231Z adding: test/test-reports/python-pytest/dynamo.test_python_autograd/dynamo.test_python_autograd-0520caf607d27108.xml (deflated 75%) 2023-03-31T07:14:49.3659629Z adding: test/test-reports/python-pytest/test_fx_reinplace_pass/test_fx_reinplace_pass-1d9175640db7e35d.xml (deflated 83%) 2023-03-31T07:14:49.3665264Z adding: test/test-reports/python-pytest/test_function_schema/test_function_schema-607a676d8867d6ae.xml (deflated 82%) 2023-03-31T07:14:49.3686267Z adding: test/test-reports/python-pytest/test_fx_passes/test_fx_passes-a8aac7079acdb8e3.xml (deflated 90%) 2023-03-31T07:14:49.3692129Z adding: test/test-reports/python-pytest/dynamo.test_global/dynamo.test_global-79ff908ccec5af93.xml (deflated 84%) 2023-03-31T07:14:49.3696118Z adding: test/test-reports/python-pytest/export.test_export/export.test_export-3dec0aa1ac9f692a.xml (deflated 71%) 2023-03-31T07:14:49.3702877Z adding: test/test-reports/python-pytest/dynamo.test_interop/dynamo.test_interop-d1995c4b5c6edeee.xml (deflated 65%) 2023-03-31T07:14:49.3716781Z adding: test/test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-98c50059c8acff27.xml (deflated 92%) 2023-03-31T07:14:49.3721258Z adding: test/test-reports/python-pytest/test_subclass/test_subclass-fb879363910a05f5.xml (deflated 91%) 2023-03-31T07:14:49.3729995Z adding: test/test-reports/python-pytest/test_native_functions/test_native_functions-669237d2b8594e97.xml (deflated 91%) 2023-03-31T07:14:49.3733693Z adding: test/test-reports/python-pytest/test_autocast/test_autocast-cad3792541a15294.xml (deflated 73%) 2023-03-31T07:14:49.3740469Z adding: test/test-reports/python-pytest/dynamo.test_nops/dynamo.test_nops-e432b0905306743b.xml (deflated 62%) 2023-03-31T07:14:49.3752281Z adding: test/test-reports/python-pytest/test_functional_optim/test_functional_optim-d28d870b6e4a8ab2.xml (deflated 69%) 2023-03-31T07:14:49.3761045Z adding: test/test-reports/python-pytest/nn.test_dropout/nn.test_dropout-5c381b2a31fd36ea.xml (deflated 72%) 2023-03-31T07:14:49.3774261Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-f22e01707f730048.xml (deflated 36%) 2023-03-31T07:14:49.3779036Z adding: test/test-reports/python-pytest/nn.test_parametrization/nn.test_parametrization-07c4b0561c36f470.xml (deflated 86%) 2023-03-31T07:14:49.3784105Z adding: test/test-reports/python-pytest/dynamo.test_export_mutations/dynamo.test_export_mutations-a1308598ca2531c7.xml (deflated 86%) 2023-03-31T07:14:49.3789387Z adding: test/test-reports/python-pytest/test_pytree/test_pytree-7ecb7e01843e0b86.xml (deflated 83%) 2023-03-31T07:14:49.3797000Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-b4332ef1533186c9.xml (deflated 60%) 2023-03-31T07:14:49.3805171Z adding: test/test-reports/python-pytest/test_native_mha/test_native_mha-d938d6850829e9a5.xml (deflated 90%) 2023-03-31T07:14:49.3822616Z adding: test/test-reports/python-pytest/test_dlpack/test_dlpack-8aff06badb092ae9.xml (deflated 95%) 2023-03-31T07:14:49.3839954Z adding: test/test-reports/python-pytest/test_numpy_interop/test_numpy_interop-2dbd5715b37db2e3.xml (deflated 79%) 2023-03-31T07:14:49.3847123Z adding: test/test-reports/python-pytest/distributions.test_constraints/distributions.test_constraints-c5c846935c57168a.xml (deflated 96%) 2023-03-31T07:14:49.3856427Z adding: test/test-reports/python-pytest/nn.test_pruning/nn.test_pruning-d14395771e0d5c57.xml (deflated 87%) 2023-03-31T07:14:49.3877484Z adding: test/test-reports/python-pytest/test_linalg/test_linalg-7b54af76081f9243.xml (deflated 93%) 2023-03-31T07:14:49.3900580Z adding: test/test-reports/python-pytest/test_autograd/test_autograd-e1c0d42718cf0b95.xml (deflated 89%) 2023-03-31T07:14:49.3934084Z adding: test/test-reports/python-pytest/test_torch/test_torch-e4a1d68a94426a17.xml (deflated 92%) 2023-03-31T07:14:49.4027734Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-d4f07a676c877163.xml (deflated 95%) 2023-03-31T07:14:49.4104812Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-9412a89c8914e16a.xml (deflated 95%) 2023-03-31T07:14:49.4107957Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-7f63c915f8ee55fa.xml (deflated 79%) 2023-03-31T07:14:49.4127783Z adding: test/test-reports/python-pytest/test_ops_jit/test_ops_jit-ad78045603f8ca6d.xml (deflated 92%) 2023-03-31T07:14:49.4164919Z adding: test/test-reports/python-pytest/test_ops_jit/test_ops_jit-09b7eaa136a34551.xml (deflated 93%) 2023-03-31T07:14:49.4170165Z adding: test/test-reports/python-pytest/test_ops_jit/test_ops_jit-a2be77a573a5370e.xml (deflated 66%) 2023-03-31T07:14:49.4231386Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-5d70d9a787657575.xml (deflated 95%) 2023-03-31T07:14:49.4282766Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-419d8c8bac08ffa6.xml (deflated 96%) 2023-03-31T07:14:49.4286795Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-4297aefcfc4b8bbf.xml (deflated 85%) 2023-03-31T07:14:49.4327522Z adding: test/test-reports/python-pytest/test_fx/test_fx-b0efee86074bcae7.xml (deflated 94%) 2023-03-31T07:14:49.4358318Z adding: test/test-reports/python-pytest/test_nn/test_nn-2e6c9c8f712a39c2.xml (deflated 96%) 2023-03-31T07:14:49.4359826Z adding: test/test-reports/python-pytest/test_cpp_extensions_jit/test_cpp_extensions_jit-936fb117f18bad7f.xml (deflated 88%) 2023-03-31T07:14:49.4361580Z adding: test/test-reports/python-pytest/nn.test_pooling/nn.test_pooling-29fb58834ee6cdf0.xml (deflated 92%) 2023-03-31T07:14:49.4376020Z adding: test/test-reports/python-pytest/test_cpp_api_parity/test_cpp_api_parity-e0456c88272dc2e6.xml (deflated 97%) 2023-03-31T07:14:49.4377029Z adding: test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-741c2318ad0d4c63.xml (deflated 61%) 2023-03-31T07:14:49.4378083Z adding: test/test-reports/python-pytest/test_cpp_extensions_open_device_registration/test_cpp_extensions_open_device_registration-7c212efa3ee59096.xml (deflated 55%) 2023-03-31T07:14:49.4379066Z adding: test/test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-d91e6b677e7304dd.xml (deflated 71%) 2023-03-31T07:14:49.4379812Z adding: test/test-reports/python-pytest/test_fake_tensor/test_fake_tensor-76f6a1f7b24b7983.xml (deflated 86%) 2023-03-31T07:14:49.4380581Z adding: test/test-reports/python-pytest/test_prims/test_prims-16c3db175ecf734a.xml (deflated 92%) 2023-03-31T07:14:49.4381625Z adding: test/test-reports/python-pytest/functorch.test_memory_efficient_fusion/functorch.test_memory_efficient_fusion-a953279d56a60dbd.xml (deflated 87%) 2023-03-31T07:14:49.4382726Z adding: test/test-reports/python-pytest/test_show_pickle/test_show_pickle-5361e547855f3713.xml (deflated 37%) 2023-03-31T07:14:49.4383289Z adding: test/test-reports/python-pytest/test_cpp_extensions_aot_ninja/test_cpp_extensions_aot_ninja-d75d40c2ceabefce.xml (deflated 82%) 2023-03-31T07:14:49.4385870Z adding: test/test-reports/python-pytest/test_jit_cuda_fuser/test_jit_cuda_fuser-d23d3053563b1b2f.xml (deflated 94%) 2023-03-31T07:14:49.4389419Z adding: test/test-reports/python-pytest/_nvfuser.test_torchscript/_nvfuser.test_torchscript-5ab7f0515610d7c1.xml (deflated 94%) 2023-03-31T07:14:49.4412762Z ##[group]Run # Remove any previous test reports if they exist 2023-03-31T07:14:49.4413074Z # Remove any previous test reports if they exist 2023-03-31T07:14:49.4413404Z rm -f usage-log-*.zip 2023-03-31T07:14:49.4413692Z # this workflow is also run in bazel build test, but we dont generate usage reports for it 2023-03-31T07:14:49.4413995Z # so check to see if the file exists first 2023-03-31T07:14:49.4414218Z if [ -f 'usage_log.txt' ]; then 2023-03-31T07:14:49.4414483Z  zip "usage-log-${FILE_SUFFIX}.zip" 'usage_log.txt' 2023-03-31T07:14:49.4414704Z fi 2023-03-31T07:14:49.4414896Z if ls test/**/*.log 1> /dev/null 2>&1; then 2023-03-31T07:14:49.4415157Z  zip -r "usage-log-${FILE_SUFFIX}.zip" test -i '*.log' 2023-03-31T07:14:49.4415378Z fi 2023-03-31T07:14:49.4426523Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:49.4426750Z env: 2023-03-31T07:14:49.4426933Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:49.4427220Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:49.4427566Z FILE_SUFFIX: test-default-1-3-linux.2xlarge_12416619924 2023-03-31T07:14:49.4427805Z ##[endgroup] 2023-03-31T07:14:49.5369039Z adding: usage_log.txt (deflated 98%) 2023-03-31T07:14:49.5483265Z ##[group]Run seemethere/upload-artifact-s3@v5 2023-03-31T07:14:49.5483511Z with: 2023-03-31T07:14:49.5483727Z s3-prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:49.5483956Z retention-days: 14 2023-03-31T07:14:49.5484162Z if-no-files-found: warn 2023-03-31T07:14:49.5484366Z path: test-jsons-*.zip 2023-03-31T07:14:49.5484557Z name: artifact 2023-03-31T07:14:49.5484745Z s3-bucket: gha-artifacts 2023-03-31T07:14:49.5484930Z region: us-east-1 2023-03-31T07:14:49.5485101Z env: 2023-03-31T07:14:49.5485281Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:49.5485568Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:49.5485846Z ##[endgroup] 2023-03-31T07:14:50.3150401Z NOTE: s3-prefix specified, ignoring name parameter 2023-03-31T07:14:50.3150931Z With the provided path, there will be 1 file uploaded 2023-03-31T07:14:50.3151292Z Uploading to s3 prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:50.3238509Z Starting upload of test-jsons-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:50.4827601Z Finished upload of test-jsons-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:50.4990612Z ##[group]Run seemethere/upload-artifact-s3@v5 2023-03-31T07:14:50.4990840Z with: 2023-03-31T07:14:50.4991160Z s3-prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:50.4991373Z retention-days: 14 2023-03-31T07:14:50.4991581Z if-no-files-found: error 2023-03-31T07:14:50.4991796Z path: test-reports-*.zip 2023-03-31T07:14:50.4991980Z name: artifact 2023-03-31T07:14:50.4992170Z s3-bucket: gha-artifacts 2023-03-31T07:14:50.4992366Z region: us-east-1 2023-03-31T07:14:50.4992525Z env: 2023-03-31T07:14:50.4992705Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:50.4992999Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:50.4993280Z ##[endgroup] 2023-03-31T07:14:50.8514972Z NOTE: s3-prefix specified, ignoring name parameter 2023-03-31T07:14:50.8515295Z With the provided path, there will be 1 file uploaded 2023-03-31T07:14:50.8515585Z Uploading to s3 prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:50.8523205Z Starting upload of test-reports-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:51.0619654Z Finished upload of test-reports-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:51.0773462Z ##[group]Run seemethere/upload-artifact-s3@v5 2023-03-31T07:14:51.0773688Z with: 2023-03-31T07:14:51.0773889Z s3-prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:51.0774117Z retention-days: 14 2023-03-31T07:14:51.0774324Z if-no-files-found: ignore 2023-03-31T07:14:51.0774520Z path: usage-log-*.zip 2023-03-31T07:14:51.0774712Z name: artifact 2023-03-31T07:14:51.0774903Z s3-bucket: gha-artifacts 2023-03-31T07:14:51.0775202Z region: us-east-1 2023-03-31T07:14:51.0775372Z env: 2023-03-31T07:14:51.0775552Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:51.0775837Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:51.0776116Z ##[endgroup] 2023-03-31T07:14:51.4298163Z NOTE: s3-prefix specified, ignoring name parameter 2023-03-31T07:14:51.4298577Z With the provided path, there will be 1 file uploaded 2023-03-31T07:14:51.4298942Z Uploading to s3 prefix: pytorch/pytorch/4571766038/1/artifact 2023-03-31T07:14:51.4306375Z Starting upload of usage-log-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:51.6682885Z Finished upload of usage-log-test-default-1-3-linux.2xlarge_12416619924.zip 2023-03-31T07:14:51.6827193Z ##[group]Run # shellcheck disable=SC2156 2023-03-31T07:14:51.6827460Z # shellcheck disable=SC2156 2023-03-31T07:14:51.6827788Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2023-03-31T07:14:51.6839215Z shell: /usr/bin/bash -e {0} 2023-03-31T07:14:51.6839417Z env: 2023-03-31T07:14:51.6839613Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:51.6839911Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:51.6840201Z ##[endgroup] 2023-03-31T07:14:54.1711901Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2023-03-31T07:14:54.1712162Z with: 2023-03-31T07:14:54.1712571Z env: 2023-03-31T07:14:54.1712756Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:54.1713058Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:54.1713340Z ##[endgroup] 2023-03-31T07:14:54.1729125Z ##[group]Run set -eou pipefail 2023-03-31T07:14:54.1729367Z set -eou pipefail 2023-03-31T07:14:54.1729557Z  2023-03-31T07:14:54.1729806Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2023-03-31T07:14:54.1730066Z for _ in $(seq 1440); do 2023-03-31T07:14:54.1730296Z  # Break if no ssh session exists anymore 2023-03-31T07:14:54.1730523Z  if [ "$(who)" = "" ]; then 2023-03-31T07:14:54.1730699Z  break 2023-03-31T07:14:54.1730875Z  fi 2023-03-31T07:14:54.1731094Z  echo "." 2023-03-31T07:14:54.1731264Z  sleep 5 2023-03-31T07:14:54.1731441Z done 2023-03-31T07:14:54.1742716Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:54.1743054Z env: 2023-03-31T07:14:54.1743238Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:54.1743524Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:54.1743805Z ##[endgroup] 2023-03-31T07:14:54.1768390Z Holding runner for 2 hours until all ssh sessions have logged out 2023-03-31T07:14:54.1862159Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2023-03-31T07:14:54.1862598Z # ignore expansion of "docker ps -q" since it could be empty 2023-03-31T07:14:54.1862878Z # shellcheck disable=SC2046 2023-03-31T07:14:54.1863116Z docker stop $(docker ps -q) || true 2023-03-31T07:14:54.1863343Z # Prune all of the docker images 2023-03-31T07:14:54.1863567Z docker system prune -af 2023-03-31T07:14:54.1874373Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2023-03-31T07:14:54.1874591Z env: 2023-03-31T07:14:54.1874786Z GIT_DEFAULT_BRANCH: master 2023-03-31T07:14:54.1875085Z DOCKER_CONTAINER_ID: ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:54.1875355Z ##[endgroup] 2023-03-31T07:14:54.7988447Z ad599e049a6e 2023-03-31T07:14:55.6414378Z Deleted Containers: 2023-03-31T07:14:55.6414896Z ad599e049a6e0e29c50d207c1c3dfc7ad6d0a452fec4c25745b3fdf719f141db 2023-03-31T07:14:55.6415085Z 2023-03-31T07:15:04.6674069Z Deleted Images: 2023-03-31T07:15:04.6675141Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9:51c670a3d430f7c7d9886677c30a0f7ce33e23ab 2023-03-31T07:15:04.6676096Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/pytorch-linux-bionic-py3.8-clang9@sha256:33691e310ed8c7b7d7d19369f1fbba6440bed4595f5bf54deb94711e24e235ce 2023-03-31T07:15:04.6676617Z deleted: sha256:246279e105284a11276919f54df98c866a63b3cba69133758cfe29ae3c1e1f1d 2023-03-31T07:15:04.6676954Z deleted: sha256:593293c8ed7a74920bd816966b2dbd23adabf7c4c45d7c0136d2447b4b32c7e1 2023-03-31T07:15:04.6677319Z deleted: sha256:e330116c1f411d7bded65292fda32e64c0da38a01bf7d8e9f42f32b0fb8f65a2 2023-03-31T07:15:04.6677666Z deleted: sha256:7925b3ec44f53cb1dd651f5bd7d05b42dc6043bdeec07fb6a31a2fe22681e3aa 2023-03-31T07:15:04.6678010Z deleted: sha256:06a2acc52246332a3eae71642aa0dc77542965205b1e2172b09befcf65b09e36 2023-03-31T07:15:04.6678346Z deleted: sha256:c90ee82a295423b87088d1dd1fd885b756e0f0679bbdb9593a4451c23198eeaf 2023-03-31T07:15:04.6678672Z deleted: sha256:bd6b3585b3cd8008d9b8669f21d41747d0a3dec9395cfa743157f5a1a1dc8799 2023-03-31T07:15:04.6679003Z deleted: sha256:8cba3901448495efeab1847990c30996c4a35120c8d366a75c24b8180dcbadf9 2023-03-31T07:15:04.6679322Z deleted: sha256:0b75b411663632b99b1ce2702638db1046f6e8c923a9ab0ae71f3e1ee4bc5d24 2023-03-31T07:15:04.6679751Z deleted: sha256:48400ee5c108e316a9ea676402db7063ca8c6da8537d8545b19e88f20efb9728 2023-03-31T07:15:04.6680059Z deleted: sha256:f9573546217362b36f1d178a17481d5de77b580b0e464a93eeac71c852270cb6 2023-03-31T07:15:04.6680382Z deleted: sha256:1c72ee72450d24ece5e9718dd2888f2dbe7101c28675bc2039775f7a2efff812 2023-03-31T07:15:04.6680736Z deleted: sha256:b3faef59e16637004afddf6c076ef6dfcf97398dee02ffeeae0bc77b386cb856 2023-03-31T07:15:04.6681088Z deleted: sha256:85a30a0aa029911a781dade66ac72cfbd3fde5be1459d092b13fb7092561a06e 2023-03-31T07:15:04.6681432Z deleted: sha256:376ae1ca03c3a78e3ba3760a8c3193ba4188d862099630ba440f9cdab08d5533 2023-03-31T07:15:04.6681780Z deleted: sha256:fa45ad9aa436e9ff44caeae67c3f051b476d5161040d53ae3204e5ef019117d4 2023-03-31T07:15:04.6682125Z deleted: sha256:c44358ac6966ef4f2457649d9be289746f7c006d355ef52f48139c294d139422 2023-03-31T07:15:04.6682438Z deleted: sha256:5117d0fa5640b29ef5daed93264d6c8e31467354b70dc1088a6df75dcc8279e2 2023-03-31T07:15:04.6682784Z deleted: sha256:cb1ca9b765dabdaaa91e2b6924ea25f34580f502f6214431caa340cda3fdbd5f 2023-03-31T07:15:04.6683339Z deleted: sha256:8d8ec7cc18932993efdc7112a27909c6132ee3ab4a0b52c02a4393f86e4f0f4f 2023-03-31T07:15:04.6683665Z deleted: sha256:74a7fa983ba66f725a0c74423c0c6585102e09f2e52541c96943a7c7aeaa0056 2023-03-31T07:15:04.6684077Z deleted: sha256:cb2e28563c461a086edcd5f9a3498a1f4d9bccba8aeda37519626b1f5afcf4f3 2023-03-31T07:15:04.6684417Z deleted: sha256:cc52fab4794192163dc327e974d14cd1a5794e78918a5d28a85a402d4dfcf396 2023-03-31T07:15:04.6684740Z deleted: sha256:0e96346f9649ea5a4754a356a04e1f25ebfe794826376a87267eb1923ed3f4f3 2023-03-31T07:15:04.6685060Z deleted: sha256:a4cbe5447fb09dbe64d014068388bd4519b3b695a2cdfa31adba13ee55c739e0 2023-03-31T07:15:04.6685391Z deleted: sha256:453ea647d7727770200c58c1d43d1d7ec4363891061b2c9688347b5c98e3ea0d 2023-03-31T07:15:04.6685714Z deleted: sha256:33c27ef13aac3703865ae1a7bca38fd289605f41791bfdb84016414b5f528302 2023-03-31T07:15:04.6686011Z deleted: sha256:5817327a1b7e02b325d51c90817964712022852454309376454902fa8c9a139a 2023-03-31T07:15:04.6686336Z deleted: sha256:2a6ea52e4b4993efbfde0dd25c4225363f553eb1944f8de9a49dae32f8cd9415 2023-03-31T07:15:04.6686695Z deleted: sha256:9be9956fa62c35edcbdc80f21fe726bab588170efd3cb244ebf62b26eb226945 2023-03-31T07:15:04.6687042Z deleted: sha256:569874250a53fc5916ebcc5a526213b040e477fc57cde306efeaa3cf9591631a 2023-03-31T07:15:04.6687375Z deleted: sha256:1b5ab2999c93bccadc45a2bd84acddc86b69d103b47947100ed6c7cfbfb37690 2023-03-31T07:15:04.6687728Z deleted: sha256:063d0fca1566abbbdf06091006cf7f57484eab6b7348cc31e59571903aa6489d 2023-03-31T07:15:04.6688081Z deleted: sha256:a62f3de1d873d2de12e2893d539725edb36d985d57cdee883e86ed06146ac064 2023-03-31T07:15:04.6688415Z deleted: sha256:06fe0ba99f0b99bc5b09055ec5e14f0f8958a102063db9d9eb86301002ac2edb 2023-03-31T07:15:04.6688823Z deleted: sha256:4a040a01f806a09e91bd5b91e60aa4dcfe5d19cf0dfdffd901d68448f5147825 2023-03-31T07:15:04.6689159Z deleted: sha256:542289e63ede29c34991833cac4954ff43415f63283d45f5cc29ab37e5cb625c 2023-03-31T07:15:04.6689500Z deleted: sha256:19acbad76bedcdd1a495009bdf4c181a60767b01a6d9649b35bd04c1364949ab 2023-03-31T07:15:04.6689845Z deleted: sha256:6f7537e06f45eba040bcbb964fe194bde26eaf1ad619649ddae95c2101fc53ff 2023-03-31T07:15:04.6690189Z deleted: sha256:1649df1f66b8c0410ee92250c0d124f4605f3b46a3e2c597ffee1f778b5978f7 2023-03-31T07:15:04.6690516Z deleted: sha256:1aa1a597fa94c664a9869883128445aa5f8f76e2c6d4ce5ae22377532ba98ac8 2023-03-31T07:15:04.6690827Z deleted: sha256:68aa52c114b66b9402b5759f4521af4b0ebbe602e353690309cf6503663be607 2023-03-31T07:15:04.6691151Z deleted: sha256:f434515bd729383873ff1febc16c34c99644fa5b9fe5052f660d14d048e03f00 2023-03-31T07:15:04.6691468Z deleted: sha256:2b8884747b5c948523db6a7f481afc387a284c84242965d159493f0b3523cb57 2023-03-31T07:15:04.6691770Z deleted: sha256:99110afa5c409f120f090114e2a82e165601c223fb0b6ebd194cb77d63ff6b6c 2023-03-31T07:15:04.6692086Z deleted: sha256:5be1b58b024c746172716d7753b40a20c7a0280f79903ac404ae212b5dd0acbb 2023-03-31T07:15:04.6692406Z deleted: sha256:80e71f26718fbbd9949acc1381e92167c75f8ae1e1963a257b6e694484742054 2023-03-31T07:15:04.6692826Z deleted: sha256:17c5114c5d5a11a7d5ac0cc76f397c0596a4324883146b3c8ac06d24dd369458 2023-03-31T07:15:04.6693150Z deleted: sha256:9cec12f76c60ae876d72cdb93210ef37229582049b988a2a733dabe3713a9e9e 2023-03-31T07:15:04.6693494Z deleted: sha256:0e9d8a5c5b0098af87a7685eb7484b94b9e9eab753cad2ddb35cee43ec5bfef9 2023-03-31T07:15:04.6693838Z deleted: sha256:b16caf57de4e985fab528a162175915d756370fbb5236b8e76aa789739c4b45c 2023-03-31T07:15:04.6694187Z deleted: sha256:2dbedc5e238d8deaeef4dfce0416d49e91d9bbec4e8f5651bdfdac5b75eb5446 2023-03-31T07:15:04.6694534Z deleted: sha256:6019c7a777fa065623500719c68c9515a89e848157e4aaa47871aa10ccac2c12 2023-03-31T07:15:04.6694864Z deleted: sha256:56ea1c587538ae781546ccc7dc9974ba6eb5c6f3c9a7aba4ed813c3a2e74fadf 2023-03-31T07:15:04.6695213Z deleted: sha256:0d4df873310a4cca8019a4f4e6df2ebc6ad894acb6d05092f3a82cc163d22dce 2023-03-31T07:15:04.6695538Z deleted: sha256:f986b78364149b5542e0c718c307aee74d7245d97db30c26ee4acfc5c97f3d4c 2023-03-31T07:15:04.6695873Z deleted: sha256:32aa9ef05c7a19618d27e212d2aff4d44845a32a47b9dd50c61bf0559affe348 2023-03-31T07:15:04.6696201Z deleted: sha256:e8dd7d88746654944fb8a1c9307c8ca3276c651311834d37d6c91fd9f9917299 2023-03-31T07:15:04.6696556Z deleted: sha256:87be2012697fc579646333ed3edbd5cac351283353ed2a990d6c97e69ac7692e 2023-03-31T07:15:04.6696872Z deleted: sha256:e968494681960132ed58eeb4a93a53f1d1919e69df09499e9027098f2f5c27e2 2023-03-31T07:15:04.6697193Z deleted: sha256:a0aed18bab7fc68f84878b4b3996968d1331a92a7341dd0c739e4a9d1df02939 2023-03-31T07:15:04.6697535Z deleted: sha256:f306cf67dc8548077cfa2ce78c36106f33d51295ad9a98f2efe65c0ab411b67d 2023-03-31T07:15:04.6697863Z deleted: sha256:eeeeda54a6b02dc7071e27dd6c8750b6023757a985948af3471ae2c631abbca9 2023-03-31T07:15:04.6698223Z deleted: sha256:26e7d8bae1cc2a9738a2d0dc74d3e2cb99300bcbd2b0fb394d7e5bc6b5060ebe 2023-03-31T07:15:04.6698580Z deleted: sha256:dfe1f9f9121b20f0f0f722e6879b6ba2c7d9b45370590af13bf3305fcd5059bb 2023-03-31T07:15:04.6698919Z deleted: sha256:ac3cef2d259e5eeed9a922259e04e07b1fc80a0b71258536968b10247df66173 2023-03-31T07:15:04.6699269Z deleted: sha256:cf2dc2ea141159dc72802dc1123b3f5aa2697c4ae267424a6bbcddddb4bf98bd 2023-03-31T07:15:04.6699624Z deleted: sha256:b7e0fa7bfe7f9796f1268cca2e65a8bfb1e010277652cee9a9c9d077a83db3c4 2023-03-31T07:15:04.6699822Z 2023-03-31T07:15:04.6723787Z Total reclaimed space: 14.3GB 2023-03-31T07:15:04.6792583Z Post job cleanup. 2023-03-31T07:15:04.6866970Z Post job cleanup. 2023-03-31T07:15:04.8048953Z [command]/usr/bin/git version 2023-03-31T07:15:04.8399599Z git version 2.39.2 2023-03-31T07:15:04.8442590Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/3bbebf3d-0612-4d58-ab58-0178ffb8f68d' before making global git config changes 2023-03-31T07:15:04.8443834Z Adding repository directory to the temporary git global config as a safe directory 2023-03-31T07:15:04.8447990Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2023-03-31T07:15:04.8499623Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2023-03-31T07:15:04.8534049Z [command]/usr/bin/git submodule foreach --recursive git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || : 2023-03-31T07:15:04.8977317Z Entering 'android/libs/fbjni' 2023-03-31T07:15:04.9040984Z Entering 'third_party/FP16' 2023-03-31T07:15:04.9088999Z Entering 'third_party/FXdiv' 2023-03-31T07:15:04.9138428Z Entering 'third_party/NNPACK' 2023-03-31T07:15:04.9201149Z Entering 'third_party/QNNPACK' 2023-03-31T07:15:04.9248610Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T07:15:04.9286485Z Entering 'third_party/XNNPACK' 2023-03-31T07:15:04.9445932Z Entering 'third_party/benchmark' 2023-03-31T07:15:04.9494802Z Entering 'third_party/cpuinfo' 2023-03-31T07:15:04.9559881Z Entering 'third_party/cub' 2023-03-31T07:15:04.9622004Z Entering 'third_party/cudnn_frontend' 2023-03-31T07:15:04.9665707Z Entering 'third_party/cutlass' 2023-03-31T07:15:04.9711101Z Entering 'third_party/eigen' 2023-03-31T07:15:04.9796727Z Entering 'third_party/fbgemm' 2023-03-31T07:15:04.9871735Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T07:15:04.9944440Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T07:15:05.0006011Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T07:15:05.0117597Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T07:15:05.0184701Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T07:15:05.0232791Z Entering 'third_party/flatbuffers' 2023-03-31T07:15:05.0273553Z Entering 'third_party/fmt' 2023-03-31T07:15:05.0325295Z Entering 'third_party/foxi' 2023-03-31T07:15:05.0370092Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T07:15:05.0430522Z Entering 'third_party/gloo' 2023-03-31T07:15:05.0480292Z Entering 'third_party/googletest' 2023-03-31T07:15:05.0528972Z Entering 'third_party/ideep' 2023-03-31T07:15:05.0590974Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T07:15:05.0723581Z Entering 'third_party/ios-cmake' 2023-03-31T07:15:05.0774955Z Entering 'third_party/ittapi' 2023-03-31T07:15:05.0811533Z Entering 'third_party/kineto' 2023-03-31T07:15:05.0848424Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T07:15:05.0886281Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T07:15:05.0925364Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T07:15:05.0961708Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T07:15:05.0998680Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T07:15:05.1035208Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T07:15:05.1073986Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T07:15:05.1111284Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T07:15:05.1148832Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T07:15:05.1186587Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T07:15:05.1225679Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T07:15:05.1263052Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T07:15:05.1302193Z Entering 'third_party/nccl/nccl' 2023-03-31T07:15:05.1352365Z Entering 'third_party/neon2sse' 2023-03-31T07:15:05.1399563Z Entering 'third_party/nlohmann' 2023-03-31T07:15:05.1438815Z Entering 'third_party/onnx' 2023-03-31T07:15:05.1592051Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T07:15:05.1655730Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T07:15:05.1706207Z Entering 'third_party/onnx-tensorrt' 2023-03-31T07:15:05.1775882Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T07:15:05.1887696Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T07:15:05.1951329Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T07:15:05.2027214Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T07:15:05.2079077Z Entering 'third_party/pocketfft' 2023-03-31T07:15:05.2117221Z Entering 'third_party/protobuf' 2023-03-31T07:15:05.2212692Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T07:15:05.2271097Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T07:15:05.2322003Z Entering 'third_party/psimd' 2023-03-31T07:15:05.2370506Z Entering 'third_party/pthreadpool' 2023-03-31T07:15:05.2421313Z Entering 'third_party/pybind11' 2023-03-31T07:15:05.2469156Z Entering 'third_party/python-enum' 2023-03-31T07:15:05.2515833Z Entering 'third_party/python-peachpy' 2023-03-31T07:15:05.2592516Z Entering 'third_party/python-six' 2023-03-31T07:15:05.2638716Z Entering 'third_party/sleef' 2023-03-31T07:15:05.2691015Z Entering 'third_party/tbb' 2023-03-31T07:15:05.2730834Z Entering 'third_party/tensorpipe' 2023-03-31T07:15:05.2793311Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T07:15:05.2859260Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T07:15:05.2925055Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T07:15:05.2998169Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T07:15:05.3085316Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T07:15:05.3140017Z Entering 'third_party/zstd' 2023-03-31T07:15:05.3220392Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2023-03-31T07:15:05.3251613Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3260211Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2023-03-31T07:15:05.3296020Z [command]/usr/bin/git submodule foreach --recursive git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || : 2023-03-31T07:15:05.3575807Z Entering 'android/libs/fbjni' 2023-03-31T07:15:05.3596681Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3626962Z Entering 'third_party/FP16' 2023-03-31T07:15:05.3649404Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3678901Z Entering 'third_party/FXdiv' 2023-03-31T07:15:05.3701062Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3731016Z Entering 'third_party/NNPACK' 2023-03-31T07:15:05.3753151Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3783251Z Entering 'third_party/QNNPACK' 2023-03-31T07:15:05.3805314Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3835075Z Entering 'third_party/VulkanMemoryAllocator' 2023-03-31T07:15:05.3856241Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3885497Z Entering 'third_party/XNNPACK' 2023-03-31T07:15:05.3907290Z http.https://github.com/.extraheader 2023-03-31T07:15:05.3950778Z Entering 'third_party/benchmark' 2023-03-31T07:15:05.3973273Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4002481Z Entering 'third_party/cpuinfo' 2023-03-31T07:15:05.4024255Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4054093Z Entering 'third_party/cub' 2023-03-31T07:15:05.4076009Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4105818Z Entering 'third_party/cudnn_frontend' 2023-03-31T07:15:05.4128232Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4162986Z Entering 'third_party/cutlass' 2023-03-31T07:15:05.4185173Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4223679Z Entering 'third_party/eigen' 2023-03-31T07:15:05.4245876Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4278800Z Entering 'third_party/fbgemm' 2023-03-31T07:15:05.4301629Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4342896Z Entering 'third_party/fbgemm/third_party/asmjit' 2023-03-31T07:15:05.4364528Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4394801Z Entering 'third_party/fbgemm/third_party/cpuinfo' 2023-03-31T07:15:05.4416452Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4446763Z Entering 'third_party/fbgemm/third_party/cutlass' 2023-03-31T07:15:05.4467688Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4505831Z Entering 'third_party/fbgemm/third_party/googletest' 2023-03-31T07:15:05.4527594Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4556129Z Entering 'third_party/fbgemm/third_party/hipify_torch' 2023-03-31T07:15:05.4577739Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4608342Z Entering 'third_party/flatbuffers' 2023-03-31T07:15:05.4631566Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4663126Z Entering 'third_party/fmt' 2023-03-31T07:15:05.4685775Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4714947Z Entering 'third_party/foxi' 2023-03-31T07:15:05.4737464Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4767535Z Entering 'third_party/gemmlowp/gemmlowp' 2023-03-31T07:15:05.4789547Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4819328Z Entering 'third_party/gloo' 2023-03-31T07:15:05.4840870Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4870369Z Entering 'third_party/googletest' 2023-03-31T07:15:05.4892249Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4920813Z Entering 'third_party/ideep' 2023-03-31T07:15:05.4942559Z http.https://github.com/.extraheader 2023-03-31T07:15:05.4971897Z Entering 'third_party/ideep/mkl-dnn' 2023-03-31T07:15:05.4992948Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5031113Z Entering 'third_party/ios-cmake' 2023-03-31T07:15:05.5053405Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5081712Z Entering 'third_party/ittapi' 2023-03-31T07:15:05.5102559Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5131207Z Entering 'third_party/kineto' 2023-03-31T07:15:05.5151828Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5180385Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2023-03-31T07:15:05.5202060Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5231345Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2023-03-31T07:15:05.5252788Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5283337Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2023-03-31T07:15:05.5304316Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5333098Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2023-03-31T07:15:05.5354122Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5383138Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2023-03-31T07:15:05.5405915Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5434401Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2023-03-31T07:15:05.5456113Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5488263Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2023-03-31T07:15:05.5510145Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5539618Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2023-03-31T07:15:05.5562180Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5591587Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2023-03-31T07:15:05.5613468Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5644059Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2023-03-31T07:15:05.5666328Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5697835Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2023-03-31T07:15:05.5719170Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5748435Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2023-03-31T07:15:05.5769748Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5799652Z Entering 'third_party/nccl/nccl' 2023-03-31T07:15:05.5821340Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5850877Z Entering 'third_party/neon2sse' 2023-03-31T07:15:05.5871952Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5899485Z Entering 'third_party/nlohmann' 2023-03-31T07:15:05.5921079Z http.https://github.com/.extraheader 2023-03-31T07:15:05.5951378Z Entering 'third_party/onnx' 2023-03-31T07:15:05.5973842Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6018084Z Entering 'third_party/onnx/third_party/benchmark' 2023-03-31T07:15:05.6040244Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6069122Z Entering 'third_party/onnx/third_party/pybind11' 2023-03-31T07:15:05.6091338Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6122313Z Entering 'third_party/onnx-tensorrt' 2023-03-31T07:15:05.6144381Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6174637Z Entering 'third_party/onnx-tensorrt/third_party/onnx' 2023-03-31T07:15:05.6195903Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6230267Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/benchmark' 2023-03-31T07:15:05.6253163Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6282137Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11' 2023-03-31T07:15:05.6303196Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6331312Z Entering 'third_party/onnx-tensorrt/third_party/onnx/third_party/pybind11/tools/clang' 2023-03-31T07:15:05.6352843Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6386874Z Entering 'third_party/pocketfft' 2023-03-31T07:15:05.6409126Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6437225Z Entering 'third_party/protobuf' 2023-03-31T07:15:05.6459127Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6493691Z Entering 'third_party/protobuf/third_party/benchmark' 2023-03-31T07:15:05.6514435Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6544028Z Entering 'third_party/protobuf/third_party/googletest' 2023-03-31T07:15:05.6566427Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6597498Z Entering 'third_party/psimd' 2023-03-31T07:15:05.6619453Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6647840Z Entering 'third_party/pthreadpool' 2023-03-31T07:15:05.6669261Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6697991Z Entering 'third_party/pybind11' 2023-03-31T07:15:05.6719424Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6750069Z Entering 'third_party/python-enum' 2023-03-31T07:15:05.6772181Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6800831Z Entering 'third_party/python-peachpy' 2023-03-31T07:15:05.6823249Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6851837Z Entering 'third_party/python-six' 2023-03-31T07:15:05.6873846Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6902291Z Entering 'third_party/sleef' 2023-03-31T07:15:05.6925582Z http.https://github.com/.extraheader 2023-03-31T07:15:05.6953693Z Entering 'third_party/tbb' 2023-03-31T07:15:05.6976680Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7007513Z Entering 'third_party/tensorpipe' 2023-03-31T07:15:05.7028946Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7058237Z Entering 'third_party/tensorpipe/third_party/googletest' 2023-03-31T07:15:05.7079777Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7109739Z Entering 'third_party/tensorpipe/third_party/libnop' 2023-03-31T07:15:05.7132084Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7159919Z Entering 'third_party/tensorpipe/third_party/libuv' 2023-03-31T07:15:05.7181648Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7210221Z Entering 'third_party/tensorpipe/third_party/pybind11' 2023-03-31T07:15:05.7231503Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7259002Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2023-03-31T07:15:05.7280542Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7311777Z Entering 'third_party/zstd' 2023-03-31T07:15:05.7333885Z http.https://github.com/.extraheader 2023-03-31T07:15:05.7583170Z Cleaning up orphan processes